var/home/core/zuul-output/0000755000175000017500000000000015150245554014533 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015150260710015466 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log.gz0000644000175000017500000212310315150260565020255 0ustar corecoreuaikubelet.log_o[;r)Br'o-n(!9%CMc;b[>Ǧ( k%r83pg}6᳔giζ߷;U/;?Dެxf^Wox[o8W5hm% W̗8rTY\]f}uo>87oVo>on3kvX nM "^\,V۳z7Ż3ڞ*L~FjHedz:}yy;#)a "b BLc?^\w4[ztl\9sFk 8DwΆgS^KrmQ_W#~[oa0vs68/Jʢ ܚʂ9ss3+aô٥J}s=7FEbп1 FKX1QRQlrTvEf"mcQۋd@FNsdxό?2$&ug"Y%\ߘfDP'F%Ab*d@e˛H,7љ:72 2> ƴ>70tr>P,mv'L;: ԣ$aɾ7l7;̵3](uX|&kΆ2fb4NvS)f$UX dcю)""û5h< #чOɁ^˺b}aq3j }ɌDSd1d9nTwF%\bi/ Ff/Bp 4YH~BŊ6EZ_^_39L[EC ódvYOļU[ Z.ߖtHp(-J C?<:zR{܃ lL6_oފƧer(^0"$5ڪҾη*t:%?vEm5tqa Ⱥw^eH6'Ύ >Kdg?z7| &#)3+민,2s9R>!9*XC~ S[qq7,!yq%a:z<\tunL h%$Ǥ]6f y[W` \roƐ%aޗ' B.-^ mQYd'xP2awEڊL|^ͣrZg7n͐AG%ʷr<>9 2W>h?z| (G>ClsXT(VIx$(J:*~CQpKۗgVKx*lJ3믫|'x7{VkϏSݰFNw`i7LdW3Rʕ"uZ0E`,u{C'F\ъ.x3M2ֻx<ć;_ʧNs9[]zC.&Xz$AX0-B-lNv*_]d3N^[-הp|A*Z*}QJ0SqAYE0i5P-$̿=_d^"]}Z|)5rC jof'(%*݆^J"~>AMMQQ؏*NL ߁NPi?$;'#&立q\ >hl%}Р`sMCכAztԝp ,}Nptt%q6& ND lM=ָPZGa(X(2*91n,5/0KN_Ď6?Bߔ)bQ) <4G0 AF' ]J)˨bqENjʵbu'b߇ٜK;tf*H7(?PЃkLM(]֟-Xصp&NI%`t3Vq=Mb㸵2*3d*mQ%"h+ "f "D(~}moH|E3*46$A'>7aX)󇛠ƾ9U^}KmJ?t 5@հ1hr}=5t;J|dͤ߯R> kH&Y``zG,z҄R K&Nh c{A`O'd1*-B[aL"T 1dȂ0TJ#r)٧4!)'qOϛrXMqHe1[7c(+!C[KԹҤ 0q;;x+G'ʐƭ5J; 6M^ CL3EQXy0Hy[``Xm635ӯ,j*X}6$=}0vJ{*.Jw4?؃ E"#1?|ђP? -8%JNIt"`HP!]!V 尛a;i`qCNG?UPԠ"ƎoC!0[r_G{j 4Cr+O\])fǶy{0$S7:z4efb#hQ #_ފH&z!HAd |}p TRi*KsmM+1 P0W YW ].PK%Mj˫-Kp`zbbq$7&&{Ldrǒ*!;[9@M:C{Sۈٟ%Lԯ6Ӿ?.Z\ܥM<lu Y> XH\z:dӀHElL(uHR0i#nq%]!=t_]ヒ+-, f~*^g/5nYFhU7um_׵dP΃|̓nE}xA|Au;yҷ}=Ww<*e'&Ж0(ݕ`{azZsu/x)W>OK(BSsǽҰ%>kh5nIYk'LVc(a<1mE ء\G=~j{Mܚ: hDX>uޗPU{G7C]Ch',ެJG~Lc{xt zܳG鮱iX%xQݸ}S^vv^2M!.xR0I(P 'fΑQ)ۢWP Pe~z|ءPQgOJӚ:ƞŵ׉5'{#ޢ1c qw zǽ0 2mK:ȔsGdur[MF*XCVOy4%-Lq6d@CYm*H@:FUф(vcD%F"i ' VVdmcOTKpwq.M?m12N[=tu7}opY.W]2u<ΰ+a1tHayɒ aY(P*aaʨ@ΰ<pX X{k[%Sr>Ӽ]\ hSQƗL rH_HI\:U}UE$J @ٚeZE0(8ŋ ϓ{+'h=TԫeVިO? )-1 8/%\hC(:=4< ,RmDRWfRoUJy ŗ-ܲ(4k%הrΒ]rύW -e]hx&gs7,6BxzxօoFMA['҉F=NGD4sTq1HPld=Q,DQ IJipqc2*;d+, Z`ͲH-nမ^WbPFtOfD]c9\w+ea~~{;Vm >|WAޭi`HbIãE{%&4]Iw Wjoru ݜmKnZ<X; ۢ( nx K8.|DXb +*598;w)zp:̊~;͞)6vnM!N5Cu!8Wq/`FUwWAֻ,Qu W@ Fi:K [Av*_958]a:pmQ&'ᚡmi@ zF(n&P;)_]µ!doR0`pl`~9Fk[ٺ+4Hhao-jϸ??R<lb#P-^39T|L /~p│x@Bq"M/lja\b݋af LnU*P(8W[U6WX ZoѶ^SH:K:%Qvl\b FqQI.ȨHWo;Nw$͹O$oEE-eq=.*Dp,V;(bgJ!gF)892sw*+{[or@x,))[o新#r9ϋ]n` *3UP0Sp8:>m(Zx ,c|!0=0{ P*27ެT|A_mnZ7sDbyT'77J6:ѩ> EKud^5+mn(fnc.^xt4gD638L"!}LpInTeD_1ZrbkI%8zPU:LNTPlI&N:o&2BVb+uxZ`v?7"I8hp A&?a(8E-DHa%LMg2:-ŷX(ǒ>,ݵ𴛾é5Zٵ]z"]òƓVgzEY9[Nj_vZ :jJ2^b_ F w#X6Sho禮<u8.H#',c@V8 iRX &4ڻ8zݽ.7jhvQ:H0Np: qfՋ40oW&&ף \9ys8;ӷL:@۬˨vvn/sc}2N1DDa(kx.L(f"-Da +iP^]OrwY~fwA#ٔ!:*땽Zp!{g4څZtu\1!ѨW(7qZcpL)ύ-G~^rFD+"?_h)yh=x>5ܙQ~O_e琇HBzI7*-Oi* VšPȰһ8hBőa^mX%SHR Fp)$J7A3&ojp/68uK͌iΙINmq&} O L-\ n4f/uc:7k]4p8wWLeUc.)#/udoz$} _3V6UݎvxyRC%ƚq5Щ/ۅw* CVo-1딆~ZYfJ"ou1ϵ5E bQ2mOΏ+w_eaxxOq:ym\q!<'J[FJ,4N:=6. +;$v6"I7%#CLTLyi{+ɠ^^fRa6ܮIN ޖ:DMz'rx#~w7U6=S0+ň+[Miw(W6 ]6ȧyԋ4ԙ./_A9B_-Z\PM `iĸ&^Ut (6{\٢K 5XGU/m >6JXa5FA@ q}4BooRe&#c5t'B6Ni/~?aX9QR5'%9hb,dsPn2Y??N M<0YaXJ)?ѧ| ;&kEYhjo?BOy)O˧?GϧmI C6HJ{jc kkA ~u?u7<?gd iAe1YB siҷ,vm}S|z(N%Wг5=08`S*՟݃*־%NǸ*kb05 V8[l?W]^@G:{N-i bɵFWǙ*+Ss*iނL8G9gJ8@o2k'Hr~4Z(I8!H G8HNW%1Tќ^?z%lOONRѦmDVmxюݏX}K6"Qi32\-V_kR(I-wtSJR^m{d a|y,F9$^@mdH֙toN1 < ҷBq/ ۓ,j|z6OSu;BKŨʐPqO K\{jDiy@}b|Z79ߜih(+PKO;!o\戔-QB EM;oH$$]?4~YrXY%Ο@oHwlXiW\ΡbN}l4VX|"0]! YcVi)ʫ09(M +(PWՇfR6#ю3Ȋt ݪbh]MTw䀩S]'qf&)-_G;"1qz퇛0,#yiq$ՁɄ)KٮޓJ|̖D?:3mhW=rOf'/wѹ8BS8]`;=?,ڼ"ϴq*(A7? /W= #^ub"6q f+=^OI@߱^F[n4A#bYѤwd)J^Z{*ǥzw73LuaVad=$6)iI gC~.1%YmҪ+2gSt!8iIۛ*JgE7LGoş\bC}O i ycK1YhO6 /g:KT sPv6l+uN|!"VS^΄t*3b\N7dYܞLcn3rnNd8"is"1- ޑܧd[]~:'#;N(NknfV('I rcj2J1G<5 Nj̒Qh]ꍾZBn&Un' CyUM0nCj.&Oڣg\q0^Ϻ%4i" ZZG>Xr'XKc$2iσֹH<6N8HSg>uMik{Fm(W F@@{W+ߑ?X2hS4-=^YgpUHެbZ!y!ul@ڼ63" ۩:6=TZõ$E,ϓRV|G&$rr;J TtIHFE=RȬ]P pLm|?$%>Eü%mWO[>Xmw,*9.[G n >X8Ī;xW%dT:`ٓ~:QO,}j6j!yڦʲT:Pqҋh] H+&=>g| Z;D8ܶb:! Å{2:+au 6:!fF+0#+̬NY"!6a7#񕪰%:r|o5Znڧs?si/W qEU馥˟^_޶oڷOj'?nc]Rn\t3^邳塨Lɏ"8k8M~?M}OAH$˧Byd}gs9QNʟ. /ӦxbHHAni5(~p>/O0vEWZ nY3 cU $O,iLacoW1/W=-kqb>&IL6i}^^XpCŋ݃k-$pxbڲ&6*9mg>{rtD)wQ`pkKyt1?[ˋZ5NhfӛŮ Qu8Y4?W֫/&W˸~%pqq{% ?K~,#/0'NZ׽Kq^ėSJ6#j8GO[ PCbʍN^XS&}E9OZ]'t$=tnn&nu [}Ab4 +OLuU{0fIb { O݂9x 8ퟍe\0zE|!@E " ;9Ώf3kZc7BI:ZYUv`Ƌ-v|u>r,8.7uO`c Nc0%Ն R C%_ EV a"҅4 |T!DdǍ- .™5,V:;[g./0 +v䤗dWF >:֓[@ QPltsHtQ$J==O!;*>ohǖVa[|E7e0ϕ9Uyzg%pg/cc6RS`HFLЩ LkJu\!`0);Sak$Vfp~C%YdE6c>1ƕ (0W4Q>@>lWN"^ X5G-nm.8B>NOI[31,j2 Ce |M>8l WIf|\q4|UkC.gr`˱Lϰ} xr.~l-ɩu_Drd31V_ѺUib0/ %IYhq ҕ  O UA!wY~ -`%Űb`\mS38W1`vOF7/.C!Pu&Jm l?Q>}O+D7 P=x@`0ʿ26a>d Bqε^a'NԋsI`Yu.7v$Rt)Ag:ݙyX|HkX cU82IP qgzkX=>׻K߉J%E92' ]qҙ%rXgs+"sc9| ]>T]"JرWBΌ-zJS-~y30G@U#=h7) ^EUB Q:>9W΀çM{?`c`uRljצXr:l`T~IQg\Ѝpgu#QH! ,/3`~eB|C1Yg~ؼ/5I7w9I}qww}U~7뭱ԏ,}e7]ukDn`jSlQ7DžHa/EU^IpYWW兹Q7WyTz|nˇ _qˍ[!;n ^b k[);ng]ȶM_u)O_xV hx h[K2kـ`b duhq[..cS'5YO@˒ӓdcY'HAKq^$8`b $1r Qz?ۧ1ZM/G+qYcYl YhD$kt_TId E$dS:֢̆ ?GЅ'JƖ'ZXO݇'kJՂU086\h%1GK(Yn% ']Q; Gd:!gI-XEmkF}:~0}4t3Qf5xd\hEB-} |q*ȃThLj'sQ %؇Gk`F;Sl\h)5؈x2Ld="KԦ:EVewN ًS9d#$*u>>I#lX9vW !&H2kVyKZt<cm^] bCD6b&>9VE7e4p +{&g߷2KY,`Wf1_ܑMYٚ'`ySc4ΔV`nI+ƳC6;җ2ct"*5S}t)eNqǪP@o`co ˎ<عLۀG\ 7۶+q|YRiĹ zm/bcK3;=,7}RqT vvFI O0]&5uKMf#pDTk6yi*cem:y0W|1u CWL;oG^\ X5.aRߦ[_Vs? Ž^A12JQ̛XL:OEUپOY>WK-uP0\8"M: /P4Qz~j3 .-8NJ|!N9/|a|>lX9T ҇t~T1=UF"t; 8-1I|2L+)WȱL˿ˍ-038D*0-)ZyT13`tTnm|Yhi+lQ&Z!֨řoҒ"HKX 6„=z{Ҍ5+P1;ڇ6UNE@Uo/>8.fgW]kY0Cgcu6/!_Ɩ} ' Ў3)X<seWfSv!ؒRKfs%(1Lhrٵ L.] s?I,HCԢ[b C-lLG+@_$c%* _jR|\:dc5u= A@kUc\ǔz;M>dUN/aFRĦ@x؂ǀ$6%}N^ \mQ!%8j0dUo=rh>*YȴU3Q,̸*E%59sTzɟڮ2kg ۱wEUD3uKrr&"B:p`\E)j<).R&#ÃecE,dp"nPS 44 Q8ZƈKnnJei+^z '3JDbSK;*uБ:hF ѹ @˿ޗ~7g9| hLXULi7.1-Qk%ƩJ4^=ple;u.6vQe UZAl *^Vif]>HUd6ƕ̽=T/se+ϙK$S`hnOcE(Tcr!:8UL | 8 !t Q7jk=nn7J0ܽ0{GGL'_So^ʮL_'s%eU+U+ȳlX6}i@djӃfb -u-w~ r}plK;ֽ=nlmuo[`wdй d:[mS%uTڪ?>={2])|Ը>U{s]^l`+ ja^9c5~nZjA|ЩJs Va[~ۗ#rri# zLdMl?6o AMҪ1Ez&I2Wwߎ|7.sW\zk﯊溺^TW^T\*6eqr/^T77WNZ7F_}-򲺺VWQ77V\_v?9?"Th $LqQjiXMlk1=VzpO֠24hf 1hi D{q:v%̈#v^nBi~MefZF >:/?Ac 1M'I`22؆DT!/j璓P åiw@wgRCsT~$U>ceއE)BI>UljO|Ty$ŋrwOtZ7$ "i 8U 7bSem'k?I+/ShݴFػ޶%WK(l^Xd'љ hM$ErV,9VqbJ$U.U1HcK.OzJP&co$8Tš*X|AN2a8T(XK aE^mJ;4!CU%\> =t{P{fy5joq:yD14,,MڗZ ^ EZHC!Crxзܟxl}u;I1 {Dh&믎â:hAV,PZG\ƀ 7+tL:.q ~ir x/cC`y8~`MvmH/+ro,i!=@%q.Ln{P!dnp?Rq7ĂU3%+<&NlLG&=Y "2hKT"$IXFsZ|1.Ex?’bch; ^TzzZ^Q*+~f#\סyff#B?fȰ V07|3 #d&3 ܮK4|=)QP %je&g Ȭ.h CлX%a}-0o 4m͚Vi,rA l^=(MAQ9uc#)jF]rA kT,lKb|GSFד;T*'Sg)Ϸ$UʛT'XňgZO醔*Ѣh~d0sZD#V)j$3G$,X* &PB: wކ?ZүiZ|zS:c -#-xt;ydg ŒfQ nF- %mTB.J5[MvU4 h&TƁws޶]UQX+kc) ێ\Q2̷5JeԂA/?ٻm.bh<&rQ F`/d}3bETS y>~g~Y!2߹_( $ E@]>u} >ѤYxua:#d68ok4,?n[qt럴+BH`PM,}n˾(0=.e/0Y-P5|:kFNơ@QtSS(5}۴]#Əd<&ޞ0쮬ZNZDaﮦ-t!쮩s;M?.pxDV]4bN۞6 *"E=:gwUnU}!|Ѽ T㸻gU~~)dD`s:)|pߠ0nwRKq"J3y4 ]Oe&e*RWBARTqIg1 ޴P^\Б=TD%X]`.QL'Cw5Z1#3KzCH=|Nkl#]ofķ_}K펦!X8BTGJN&"N.ybj(eF=#Xn2~D<*|whp&#S-jOG ӋhOG_oѾrp0L8B/t5WWQQ[0݇VXj&Ӣ 5ݓ㰬*8s$(\i(n{7\67\쾀48T%;~jXR ƒ^Ob7gkkdURaN(O\}e&.V4+y^;:3EH_e,{;GR) 4 &jb[ j9Hɦ,[Ke0,tL# +-AXU%55)[nLT٦z$*JZ(jq,>5*ϪE]+rqC 6W*kNyDjk)/ZG3VAoZ趆gnS3pUqy}O[! *M!d6krBTDjZk?UL"@+1W_A ,o5fyڪ)V :oMS\WV*L[.QhET,Ϯ`6}<4RHv-TbnkyJfgڪ%pJuv>>Ze[+EӺ>`?NJ9 |aI8.ȷXvifLg!ӝǹLq쮽JCe![+:-7n7+!Jů U;(Pm8ݦs96䨊l_E9TQݛG4%ˍKo]־d7!k \qp΢:iz}Wm?i U V#unsHzNOzmG4TVӼf]IRdGe w]i^yC8eQIOO^]j .lZ~/i*Ld!՜_g'0q a)6;/D] da(|X "wU dPy$C$\U壽.*$ V5 $᧱.f5 מ9(n< @NZS"I?RPxww[%$u){xIʇ,K!2g svU!a&͚U橼r[94EՉn0u\ؽLb[?:7yæ[`Y\N-*˞MzVּ񌐷g=n"CyEbAPT"-QZ>2:-#rg")(wl㭋TDí +}𮹶mDνEbzak f7~׉薽'}h;:d&/# [7 MA# d螱_#^=#:b?5辵 Vg G ܃ً0<_DNb~Ll$t%bPV00bX?mlb 7?"N@/Q6x ycoЁM:٭o` \WpYX*}???)w}c1{UO G pl  'M9G A@ #X%mf*"V#%LhRqI]_v `֏PMқ'Ü^D`y; Bd} C0&ȰKB h;&c1zW56 '\Y'HjB4<#Etw.X({@t2;͟,^BPЏML`v_LЏ?,>ؓw4ޝv@(o:} @e= @3{. C ;CPzQ*v/r^Q>$zE{ׯ|W# ۬)* =Y.@٬WQ)XBUz} 8S{/}ַN`ځ_ AK%ҟnM4bqM )r9YWO[XAOE{v ,ޏʀwC0?*ݾ6Ү]3S"Qgn #iwq|x@CEhMujK疆CkݣlnixO_* @2nidGD<4v(vg-.cI^4¥"_/[իO$S:)rQc5`y}!IԱ_?PHR}݀@#%b NO)]{i`%$JX6g:6me^>ɗ~ZΘW L#W')9]g.Iq$380ctqcf~Gr>$rLѹNv9?p+tV) P=6mֳxnk`^dp[.^*$k:X uzi3Q<ײxh|5') 'I^}jCxhtScaXc~ y[Ei,YTI1&딤_@dz],ior8FúF*/ fwHVo)&Ӗ), Csާʫ|\ Gd`2 H͠nA)t4({JgxOz.-j["w8Bke, b2IDŚ)Hh"7#Py<\Һ .3Çl<2M[7.}+>f&;8HsA ],ۯ>!hB|@PVCKcQ&/V|м؁]=<="72B}% 35-_@]J}ؾab̌$Bkʧ4CCI;W~IjFu4AT Dǣ7p~5|/"3f^d:^̑y"b/efxDl㛖o:PӲUi7F2A.sӲ׍}>n_n5 xa54^q"Ex/Zy`6{P<_~suRܢ/=Uz{C-']w[-Aroqqjn:q3!P^tbY6s{m#m3vl#׎tt(2yxVƎ"f7)Z ֦x\OfoFo` NzO~ Ft=eOv'wBne7eNOs/?E4ϓ2+0A)aIq3c"[z~ݎQ3rHY?,Ѭ)K_haVr8K}0h`_ ِbG(x"ë oƞQF8$U(a 3d@!-J4!0+-p|  #R LJ3ChEd?>%] p)o27g@Z΁yrIbc\w ^H>KDN0q2m*Y3D+܆@q0Ƭ^cqf8(f J%X ς51DHZ$ߠEst^} Jil)Y2|zK16.&gb".$' ѰfM)No9JZ1KIG^m|/"Wяp[_L0Eh3KyLg"vȤ뢚IdhCǕ%CHLƥ.xUqrB'm~ hF5(Xd4A]J@wVhAYQ3`K?vzK SQڢդ}std+o]JكcJ{dZ{kO:n-v/kȅb:%0?1Rڬ%i' l1A?w8D 8dH2-mɥJ#'i:^/'TY[4hcNT ³tax@(Әs'!u XEMOOy64׽+'P~489˓-({mP|  O%`tLCI4ݨL;5sкD+`~XLgEpwB7=^Ogj6I0϶}q~:hOB'RhY8 :;qV!ށ'Q,k)Zjo;Wˌ^#9|ΩN :ǩA1j>`fW'4yW֑dE9}癎ZI H)' N7"ԋdU47u$DGn<ԁ\ӑwל6c:[H&T9~"_w~a|yXe0eރM ; UTL٨F P}$: wtI.8Rf'!) ؍{WmR$CQ,*=AouT:TpZc d:go|8}HrcbN. ]VMu)!.}p)b`wؑF\:30,B朂ǥ@REB&Ǫ>$8'vH Nʮ41QB0q&-jW$xt w_Xps!7ĎiZ`SZ&LFy8LX΍1i#`hc'&WcHtv v[Xp4sEYIKq#AGk2h1jC0DDDL E<6:8;rjD]h(#)F"VUx^fMc=o">j 5YTgG"HmYB)ZcHZTR$ghUyDA/` d[| 9Exk=8&Ok_~[~ub3VU2w߲`m`# ESV6GA"ڦlA+ wHjTe2y*TYmH\kBsRx6l༯xhw#DUDJ"_Q](. \ZOhUlڨcJ)U$A b 8?nb{>~#0risRG] a33˯, .o&n!!/YR+)jZ" cxdI,๹>L*9=Ad5{WbXBtv[ܰ8w\jx1ګeG3;ꉈ l1JB6BB*.3E̍>GOZM4wx_5o&Lv1A` F<)JSE5[<-ɉAoFWC跙$f*X ȝ.I[,?;U9'Gr'4 <$_ GB-qQмL]:h>d@2UCzp9drggLx绕Z* E\,U@Ĥ $NniQ$3ש4lM%bj<$hy<7@&-eף1Q?[]MD?q#ɠ%hD]0.kcdĺhdtV]Ig},8{UykM1qewYpMtFةOY@0'xZIm&H: GU 14RPـ OQRd65l(A J\UtI<(``?8X\țlg^G_Uax$wxeIanS7~J"AcpN\N<~Ip_P)~߾e)܁n툌4tztAnځuu/S("i1^[G[Yr1Do 0,@$K3"Ik#w[B٘قɵq{\v3X OK, A*/&&C,b<,%Rzi=g;9\3D3!Ztx' ` n kEo_Y{ReXes]AF-1r1LH zbsm]LgOM\f=ED<EoGpUt,cR Z\G |S},8Nn@Rh 73bnkQMcKÑ"{wO,8Nn@IRT,EU.g'ڧ|/b{p`, N.]ʏ69A!GF"d3ݺ5 ALtKV ܠjVb:AQ\նٯ?ܳhgn蚐ڙw)8@bf-}H>g<|Pb tr$#4Lef;B;'w$] O6Iw$ݽX᥉78Jfٌw~@g'b) &ŎLQF몕VRXy/[ǒn!v9{r`SBwNg+R )bDǕNvFziuAR 8y %^FVcQ!E F,M^>i;ʂ~ {J\P K (m XAjY}vrSPC˽{_'xHĘ@pjQڡz`vӧ/99A ͱ|){uO,#+t,f<:xz%Pasv[Yp\'t%H#[N2[_;OSS+>{&3]ļ ť01{ų8y';&v`ⰷ2?<#O:8O莲Ѳ\YpArAI27 p ~LSfwOO,8U(&Af /'lB+ӈ A;+]8Bi v10h)?s"7Gejm$2@k=׈Xa)og**zdD#E/Ā"鸫\Jijܑ[w.^XL@Q@AWEɵ`lPržfrҥOҧttυ rfrEF5mհU:) .Hmv*-WPajGs|( /UR%T^).ze'*A}S`Gg | GX>K~ZLeD!󞥋Z7{$4iPus€Bem!$>s{\Ym"dgrA`T/9Gd[}͉ xJppgfƊ8ܻ]n9pt*RxX"2X$y;OƗ΂{8bxpw2k}[3j@X\SԨ ɢr+#d^ P8 m5oQ޹lG78+*ڥ@ dOu謢v1m&O~F¨`Uq,QH- ~QPQp8 Ȣuj߈dzIg8B(,8N^xiGGBG9)85)u 8 ɔFt-w1ܙW Wm=3[J*qr[𢵐ſX*Nޗpi{b,uBfLY1snDjSǢ!|٢, I1ث]/PxEϺ ݰPY}ЬΖbGm[8'+^:p`CcNk1;_sakI I>m~-&7_?aR~ߑbr[o++)[m_q؄M铷β5] nuin^Vc#eSgfpccbՒD "I;"6QNCO|}JۯU>sBs1LB1*|lǔRHQ3=xq= J^Ռ[c2e;l&O ӷzUxhO ~@NpFBsK_mVV/{NպmǓ': ~_rG`ܧMy}V>/ǯOQK<vFH|n=*/xJ75tۤn3>qTn6#pʶ{Wƍj͵%0%$'ClQikZg!dYq+xD?~h~73}wB|?a?"{c4m}/'9=boFլ2?CF3縝(p,iQnR]"ML%1v<<9ZHıdSve*ZВ`k>\ cFImp .#0C8AϯpgcϿWdWcp98 p~ZM;YnN&uoj7qo#_Nt-v+D{U~!Bx~n?SkcHrXlv9螑4ka`ŏǃ2@jX$N5_@7gU4>[)ޤ4C}9Kt>Y`?Cڛ~)(nr>^J3ƴG3gQ`&z3s/3g mvWi9֢WJ =RZN3 7N\ΪaJ+PJ:)$$PGf'u_/ 3&i1+WZx6# J{![VȦpw!^Č}%?+W) /Di@ߎWXն?]eEUAQ ,nЏmN4W >i]LtH8=phZjqC퇿@ `+u" 8[$| 3\s͛Hփ@^`u#Nzd~ ny z O .Նu.ԇ+h )L}j8,&>M (8¬rd c}}^Da/u{/Nr|9;hڮ2}2ãCN ]s> ]cMČ3%;Fkk h&.ޱCh&\ Yx.aw\_׷EZNT*lƊO\M֜t2KJ4 : v|~"E ?%I^PǏ9L3zԲ9ͼ={:}6'[J1{ H^yFriPgÐ1[(~kx&> 7nZ7DC$N$aS4vRFS}zNmbwoEX="_N'@ 㕄VE4*mv:X%PB #H<Bu 琪F}n=Gu4~Qڍȸ%p:4u ~j3̽R]LtKTb-Lc A8+'coZ3P"sfrih&ߖ{yfty;A3/*5G 0q҂ ѝ/"f;)l+~ۚ~M\Dw q^Q[e')l_X#6=/lBwV|YDMʵiEWvꌈs.{szK- \F8y-s44Ve(KXV9 ̐:{Pm7(/ܙt͛0] |"P pwUvza@Zp 0vݢ|B ve4Wvf;h5>NJ>ŝ)]fTb5$|kQwo'uy|k}mҶ A3`W~˸?Fd[WhRHUFK뛆nW)I63#qiR?02dɬ73@X r;~aKDhq1l|5~>_/+L1>ƺL"L7uЌaWܼâ@ur_@yA][2%1Ë՜.i_dWP_XoVF/x%1K!!X4/;ź4)I--Ĝd*j9<ܲ5cQABib(ཉ%viUnN6WZmH(*gʈ 0I%$()QV'ES!#][aXֈ61I1zm=qSlV.j-s1Сn /`#I_GYAoJ) ?B!F6atr[{^1D23B᧶mP%vG`na0#_Y grH$f`nB@ @^R `4J-JXBcJYkA0fe@BG`$B Hc92`gQ0X$@^Zj aDzz㰓z[)GDSCII~Ѯ~f2i%!bź( e <%Gc"yn_OY_@)9>f`}X/A|m^LƥnPH NEGM^/Uri^NT݀x+{9QÖ!#=O iy;g!RxjcMB1)K eMggjx8_؅ le7$klÙa̋DHʉ$%9F,uUwWUw+t"LlMdD2%>e!l( Y0*#TEa\GNS )fyJă8Qp("2bsBmaia5%r`x:s-@(H!Z9ϭFKtHͮ2o؇Iic-qL'=1Mj6*Y5Vcd1per ,l eL)I,/r3R(ULaScjq%FTX= %"scqpj+fyҖ6613I LIXnㅃ"Z ^ 1aڤ]h :Q٦9g= . Gk!V ɨ-)ۚI p9[$EIZJziC1S\"}R+"-:L-itH$` F^t4-[ۗ_xֲ H)#q`rTh ֬,a+ךjK֦& = rҚ|S8;Fb0`:mZ,C k24lKwreN답к!Jn*-j(VHѢK[8"A Bz.{i: +iT{/,:qWbLo./m9i]r#rss4q3η׋DƺE&~S$`;"cZhߣQڸUR4JgwIV]\*g٭jӚubzOMhN7~򶲳ٽ39W7rgjY*hz)RV?}M,7IgV9g/$޻ |×w=>8yi'*}wx/J4L+[{ .O;/^>ڠkGP%T]c}SLo.fV|m(a|k VELSK=X8:= 1:"mn57.foǂ5^$uWp$#|i77cE:DgR='s)6Q36&Yd$r/e8LZ o:*}~`gLF׆?O`O(GO`z9}ONN5_=-蜃ѭC}wLNcqBk;ug? ʬIpq-}+|v;1RQ\:;,w%NKwt >8zrhƬfvv9uܓ|pJSxwe3FF|ǫV%z}niZ)wx HGP&) O^L #w@tr '8%LOIvAkBuqg{jG&.в]M) &E+OrX8R@*"ik4T.>:j@-Bm6NAkjLG*ˉ!0+ 6iN ɸ9̟\d4Wgg LΔ=s72'51 #gvp o}~$ •)ף |D7I8G}L/F9f2qvLN~w'ϤC7)5D!z[W 5Biѡbˆ/З>Xw^E_nYydZ IYbݻ;jQ</gN}:}4!~y[Zۏ~sw{z mUPz. Fy^O&e?_o/nJ2(S] o:#x֤HW{( =F 3.:^h3i_yWiM]`YL D0^فKs,DNLF0*{vނMnkDӋ>2¦y1Lab_taUv'7Jj!`Fg$I=0'hƓ/$_w/.ez˯|>ncW_ۯzS ?aGUqd}v H7zZ,¿͵^u[jܼ@w?TSP?g`<:̎gt?a+vfhN'#= y}C^^Fz6j28._cFŴ;q{S\_z9? cwcګ$˘as7(w1Dnlc޸Z=B"*{< p־K>)^ei26W|#5\]&"ÀԂwpqOwx_Wϯ?yP\eB.4񀧹ezz rꉃ9ʮ*S9'מ:%^noߴxĒ>֖]8Ӈ.6Ok-;f()`;ߍKC s 0㜒ȦA>Rf*:;U3#0p%`3#V*mL\yF=a|wJ\cLa.kwtr^B$tp~J$~pxr5&j'>t.뿢,óMKJ $Gcp]Vx]eX\kȴe 8b1=b`JqdJ8&r[6\I#c*E>lQ^8X23#0p;Gs# y='^|yU1<+FP:-h+Rw]]E J1*BԖDbmWֺ(DCkacge/m\hnkU~Հ^kOui)RW3X\(9Fr#w9Bµq!Z<(Qs*eg[`ç"3v,̞ϋ(s>r -m:ҘpSp۲ ;>KnrO2&,z0`AHdomFB+g[FZ rTj.v˥XLL8Bճ]3s\@$!z 6 `R~;8*Sk"|Zof򺜶˜5r EJSc255&FԜ//o\IuzӝoX=Km\Lμk^zߧpDYq/4.cf{td+ЮiTnFN\vi4|]SRm";`rNW[#G<ߏ@˙]~k]| $q>P`0'ĝH&U"zg`cow82.Tx+!Rޔ3х<[F)9hETCIka(KuH[/lP_5S5 vhY]3iSG spp7q"s/w[32*+dcؖ zEIjy'yQQRo%B"`#cRyG}ZP֒nkƁ˺5]p L2e8ზv1>1C1t3 L+t8J>!@:ng{ZiQ@J#nz.JَҨKGiQ4RF 001&WGgsATW?LJU?OI몟WJވzȒmi"ڳSc4``2k*ak$.q Jaܠ@ZTv7(-A Z+;!ѣ@FFAzF6b[ՂP0Ș(* =>ZX$C ZMP>AdjM AFLEd_ #,l,lc ;[Q\|C燪[xa괻==&ձ >8y1'?-벟ƷW7ӣTw[w.|`Ζ~\\(Igq5 ׭8c4! |5}tw0qEr(^ZlUl7O |E珈ܓϗ$~ L %^ʲآ:mLHZ(|8FnL&Zɽ1PVVmAq LG6 XCDph# LuP2/])lPP{aA}<4!IUުW޽.IHI-oD7b3Y7n_s}=!$im!etnɫ+]IH/diIRt-=(]r*SU -S/y ; kCvg%,otOӑtPd LlƑ+g(J.4b*#l6 rAp" SYfc zt"Wdic3nêhuoT4ͫ:;0o .Ƙo g^*H O=~vϿCes{ߺ7cfɻ|g Nm~{wɜ6#S'oT:vJ؃Yzc @JsubL4'Y&}^i. XXa/9?,3H' pi]]n.*#V>-ua*{&" n{S|gϽnI#5 4I CAwxrڒ$ rU[kܤʔ\b ɑmO?czBbTGiܔR )ƙ 1c30#ꚻT)㟯tcw^5z jLkTKBB !`3FFg[Y/jXOפ&/J  l8u0TRjM6(L2%LY$`ZD(dZIBHӦ4%RXH PcdAԉTa<dj{="Kֺ`iJsOCW[m^'E@A"t c `--UCb,UWT?\moh{u,S2e=ʑiڨGloPwE*z$2>\moGg{E&vV*Rbouz-YF;%SLA2%CuCuzz Z뭮X]/#"聇5 PeQ'ej@A2l/;~pR($RqhmX I!m^й(TCj[moh{Em<%SordM I@%k=^ۋ`&j/k*q8]^S*M~KR)cnSyoMz3s=p]ݼ7mӝ_\鼻:*ض'?]OT={bR,o&2=G/! %T^nuqOw"c3/yv[7Z?^M\oxǏ?\uʜxbC n Ag:X0Fp FR1gzC}3C?pUmV!Smj u6f 'UW{MWwVg޹"l>Z De0K&`Azi Gt6%Av/܃ꥋ[ z4=0nh^8=O7l՘1SbQ9Q zl$U{Sԏ ~Ŕ_L(I4cYApu~1El CciW,Bk #qkҭ9P!Cgm.! 9mQݿTVA?@-+ c"E$A!jg_KV qjEXİ\b8WpDmX9zt9>=9kufqW0w AZ'`>FfC)ZΓ홚[8b-gZ+C 6@-y>EkXUKwУIiTY% H nɓ<;!-IXh烔#dWLsDĢDdDd~(KDr4mcE(%Aa He,q8*KʒyrXm? 9Oxd)%b*KXD4k'Zλxsw~I/UR=q=b?8 s0CPAe2l;Blm˚ȴ7iǡio!4+ňPSS*Aʡ\l*^.>?^i%B>XaD,џs^cB2j8jXr(ztIMjR*@N'#H1|v<˗`>mJﱯ]rqJ`?uU0BfC;O.{vh0-im"Q JhaPGy #H1NEX@7U7VR*t7mo2Û= {Q'wǯ" FIߘԿ-\A!b w4b5dog&?67u7 >x$#<91s1AL[m'9NRGT ik;ݜЗ7,q-\bHg1b 볘90FcS~f(93#$s atdmMr92~[]Ų#sդh2)"PCR9x%KS uN?\M$qjBڄ$4!LhX ŀؐWBʙ ;ڦdv) k67Ш +VH`Bdɓlbdg }iv] __H\l!Nب l^08N!.x-n32C'i8:g:bW, }԰H=sG]ہg>g|kB@f+6,%șczZz"5&s@0f h, w{|2;|! ltT5n`IEqH"[)-REy=QkPdaj6H9 \J&g<83LdCLθzW6AaXŒ7*fv] l8Z$-p;ɀFFrB͋6QƐ\X͋>/u}& AJ7FiWaM/MKUU:/84\j 42@R4WW ko_Ф2J 8Bs5jYjM[ٹAZj<=tfԴ"$#nY끜uAgi'!>7z&2p@1z d!d6H9aA_}ȴ=S}2>d C֞!=,#H1qD`6!X}2>u3M!ΧVe#E`$ !AAPs!!mRup;KI;P ŠU U7T\h (|޴qf"#H9:3\T=@s&tPvdBrAwe(lc+(v_BR2HS( UAAO,777>^ywun":K0d<ɛ2Øi` ׋tm}Niovh,5lяSR (@(I^- O5ŌM5@~>F9PD.rY2r}67_Iu0ҧA̯A Rȕg+(}$TgEKgz (@KH%G+ , g)G^/t%<hM"/`\#a4 K8ۦJo^tO0 3!2Y` RX%4nm*X 49b8r֛s~Fr$-&=esv1V3,)iO`ŹA ڭ[򞮠tuymʹo`hP]hΊRuv v1mD_?bGgMp䲰2Oʕg^(a4Aݖw!=D96a>F1r&kK5rvM䜽r)9R2#=4YbVrY ZՁT|j<\F @NЦŮu$TKW#G7읝߳#eoŸӕm:_ 2RvH͠'8# HY-w9y7<1=\#߭iYg2#U DF7wbo#qgqNjO ;Z/Y/{yPB0;]3ɬ_d4ͮlnwu+@J`ǭD_*[\C&ǏiaF1q[=9@p9<k: O>.'kx>5\qbӇmHkmȲ ’`IL`]zڄ%QCRv }OUEfDzd`If[=M6B|Oxc@-DgRb`۟Jh~0qLA- ?/:/>H)!ab\+ToYVPz1㦣+pRSx>@0:6J( N!|>IW{ʳ8Yw?Y:[;eַeH`o+%l{ʓ|7e?<23^0O\0BYnz*Gڵ>)U~(P#_ee/+J*{}=)FAVR 7gLv4e,m?s͘>Kr1|_ !,秭TT3OP 7B$+ߴւn=Se clUu_ 291Z,{Ogŵc+2Qbq߶US$ƮM0)B |SJ!xb{TlÕP~SÕS%, >ܧ_"4%R$Ґ/xt^ U)fJJjM D$vlSf6Fh.ז[.g*q''Vֳ=f F'(Jf^rrm4DJrL"ixb,Rd+E Fs>Xdo փ -[zEOf ܓ -U0YoQ/ NH9r}Rڲ A$0a'D>Z꽐PD$֓_|9_#e8|\IugtGc i?HIE3H ak.rlB5H$IA~tɕVJBnJc$+ cuI{^2AXYAsd}RV\?dm=TBry1lYY8]Z9}Vݪ\TFugA:sf8Mn6sQYSӕ`j&ƓX6Ag6aes\Suӿ_< m5'qyM?6- J0ݫ]٠|Z58n 7wz:P^Aӭi$GFa4/OɘKzTݧ(pn:bg&1+dP,A).O ฃ:#q f'*U87F><mڴn۫CTkr_?x}O Lk)߻X=knp{;Dl7֩ݙ!ڣ~} M Tx!Mm)mLA`at~׽lhY7Z׏m%q[@M2 18 Ů/U5/]Z_Q }Sd6c=hoeV-- fI.3SShfVfa\mf;]I[^- M3[S'>koe³㺥t!Lъ|=[ٖLF͎k" p}^e)^7v߳#;5dz Zg*ha߭pUڻ6si5,8ufl*!W#b;ʢ ]`BDtŞH:|!Z*}!K__CmC\٥}>N[.be}1mt֥d Zn0nЖs)?TXY"]`*Mͺ{afroKP^C<N&{~|iDCʈlsyTjE,Y^u0Vc3Y㚺̛`ƕ'NoR+dKORY*6e93,nVk?o._ѻ1Od|:hlz4Gn|4OG~GEDLNӤh@;'m9YeeōPkj$# ~ӆ<.N~}w|)0?Lv0GY+5MAèZ!^8t6M:)URt:{gM3\롽j 9lΦedžK7 &t TG`٠aAjNb@+nœYGPvcl cbC?sz*P6iVev,,e##%QzOc4`JfS7\yZCGZtRt-JKt㓛(? N>/c!w0kX2+CS5.^v\{Zv2QFCI:]o/;[fy1[.kO.mKyҬ |_xCQ\PϯX :h@hg7?|W,G.悵l*QKwFG0HPK5%kvB%8W!e^L5h47i{)oϽ>IpMPC57_xVނP'~*l-|-(?zhB!RoOQ;9)n->%[(AW_]>rņ3ˍ1So~"Ƭj֠/qCS]znys[<{钶Mi#qQOM[^~}N77VSlyS> "ZE=|ˆoa a:v^/67ϞM4\E%LJ) ʑu t\N :xǣq8?(`s1M Mp3ir*F'id*o'깢Aɬ*P.˛|ѯOgsH ȹmC{ݬo~8@1PQ,_pS,;4hE2kx$gQW^RG$TZFJu&zmw^P8k R8'." L2Y1R5ʅ۾uG&cS8 ))طs E*(t$!E3Km΢쭤c.ORPf'`f%ZK  1PRD-q*R!e(F:S#u;˝^iF0rՑ j/,eq$-"22CIɤ ֢Mµ8—AJ13c›!1 6sPtPA;GZaGS8uz3"ӨzFj]6dI M,;szQU`ߦ[ w*Jrs6:ZJgĤ\B ^"w4"-▤}&ątheAr4E5zxKZ2 T hvI)۞77BUIY(]p)k))x$hDm̭!Y)\;Fh0Z6KF%KeuUR q%.IVg8:?Cw5chܡBr nZR2m.X !Ed*M ʢol\&VS|ђ(6E{CF6d2n.sm 8k(7%.0N-K !!:xK H]qۂa8xqFH+75R(++Gl!xDRTv *&(:-Z+vj9(C)6+#띺N%T|U*b1%^f@2( =! h[p J֝AQ\0@kHRpjB]G-X>T/d#+!acDjB41{ODaLfL0Rnj]PPQ*BXw 6zd!weT*X.I*@&5`68 H\J:A.Geo@>hJ21GHlcɂ31`,*zl\Ri͐jPoȻL >"~Ⱥ i ¸1J DbHw CH 53 RUYd%  y3 ƒmM!*hpGhy>oIyfI@B¾ 4 4l]u"шk0,3VdR9if!<(-7Y.$[z̲NdM"М(E"Nhޕ6r$ٿBavaIyD!`= Lcw2M$ $vc/>(Zkh:Xz*_0V3n0ӻFlwjV=Ve]ӐwT$Mu^mHe Ձ/{Æ dnO%ӞKꂆCKe405n<&#q &`"E a҃c?@ I&bZ&D^ʐ>XM.**";h{  Ueu2hnuj@q+"qz ,:{P& :U(ϓSVXR ܂l2DJ"dX~ʃ?E^ns$] -تr6F80`= R=-A$&ow!> jԕ+P Pv3I#`aJɃF v)ѸcY 8A%H/3Kl`R#mA9#ƌZF+my 0)@s" xd&.3vmEQ0: H$ڃh  e G쐻#1ȈȰJV<;( s.*@aHY$QB!H>kJ fc͉MG#,a5VY xHmJ)g۫, ^"{ZzB;h[@ ^ f\M;-[ `\2Lb1״$ D-qef=0΍Q> D0Y]sڻT5$A"d`09 7|(MַZ p4נ|=AExSA94cF4G> ZyI]r)r AR`&TQted}z-Mbx y_qE!I}' WD#! V1d ZZcx 1򨖌D, U=$Xx Q:7j#J0^#r pNm(pcm +7 HFuШVF0*I2 т#L5%m\4ܵ#kD?3(WsUΫ5o (R18gȴ1  rF9pˤ3R g@|ePޟFpJ8n Ȓa8^4T5[%Z9lPb.⌙$w@7aA8\dk96ِCL*V: b] 1 7&YŔp2~ׅK:(`r X-( _cy.-k+3!zvW.b 8 ba.y~ֶ{_Hz6gK8pyٶRͶv$|PܮߝL}#T;9.f-Vkz}8?ʞ y;矛5,n_Kjgs@\^Ɂ/[ PV2ӄP쥚j*jnr5TC@P 5Vt䮆jjjjjjjjjjjjjjjjjjjjjjjgVCF1%5i&95ಞ htj(Q] P] P] P] P] P] P] P] P] P] P] P] P] P] P] P] P] P] P] P] P] P] P] P] GVCGjtP !'22H'j(@j1;mm8H CVTk*%6` .?)AQC\փb̆ ^ojv9o˰ ^Hi==J磂#3ڃ XHE ~!` LVM طkȩS g &7qGVY+G*GLf 2kj* x*`Oeyն!i!U ȶ>#뺸 =.DC>lX6ƃ|*# NTZ)q&I饚X G6 ΄z"`l]&VI&'VK#TFY`,5k^<X'KFb#!-C7Ur|KԷ۫2> uHsi_^wG/o,-㖷|Z6*-nJ']-ޖoJI U"7߾'hd02 ZdB oY:֡Zy9y䋫$_ŧf"Զ`lK^J=CXNmqW՛ ~Ǐ'߷ S;T{ JKzsZb޶~_ϾYZ,n=7ztQD0Tj@6]/рt2tT)* \\!ENsF9YǓ'Nq冿MSIJy\#PoJ[:R}O,>^nvomvds۳[0P| \u)h"9p|)MKk*LqAVnoH>xtw/z\~5 F ^V=Wn"Ze봡h" XoW/zC$G+1g! mb6ĘUʭ̒ل׿ɽݏ%MMʙl" DkdZ%LDid*9>vITR\ö-XrE'})cN /Tmfj ^Ѵ{bh^s 9-nR)G,\:IqH E=$LKC.#l)03˰fPʭwh :&UkC7cJDq06[~OX1'Q/Ks2:5E%DkVfqnkkH&aM*: JX[_+V[~5T0XUCٚ|=Nӯ+sBz ؾ.zg:|S7!oY8B*ƝGXy1q1Oh|m=Rl"1~ AU%V2*J~NdlֲJTc䪗1(!O:>+yK7-2M)5Z)|\kK*6a,YYabZɂ_,'֭KuQZv-H@"^XOH}иZO0G"dU"lJ;#kf AcF 65#+ !D!: X{(!H\]rqÔ:XˈOhlb1Om#Y'XbkV`Qn`ZXC_Up;~X (%3XlM4 %.dp/CL@sicmUI{b!ޠ1. -a5nfWΖTϘcrTLz APR1]`VEP:DÜk1 2XJ\s( -w@PAmq^8Eڙbb+ݛ*g-t*X$xP J@F;x6IH \ ikmȲ0"ma X YL0;jkTHJv2s(Q"eB&ݷn{N=Ut8!oe[@,3 _4YD@2i,"zY|wΘtGgj@o]?t?1h aޙoC1<ʒJ`gEht,Ԋx;T|s:M;pT$UMSciT66% jEf@8"+%po]MEwF%R92rNcj G_,"=H_u TER+:4$"[ХWGE]% AK\6"I̼X\jEQ_U0VrvBD ǡ4z\k,@й )! SYe9$VcZWuD8M d3iJ뗝uKk׽,3O+fHNc@JA;L';\u@0- |g?֝4:+[YolSǻ xf`yD-$ >:bvf@MLIK2Jȇ dPg rm _ƂBG1tl1J# I2HeV**C6VYq EduX:{T$ 8U+ɉ᫘w%,ܶMxYI N;XwO˭(CLm]Lܵ!"b+pmpi,{%!>\3.HʣpC$*f7]_GbT@mnk)1m,C~,#m60Ze vp d`: (N6Gݝ9hs^31b-NmN`8(ZGO{d9jAq-F|d)ebzc"Õ,-ƂywT\- 1xXːB@::|*i5۝rVHڳB&0 ֨ `f)A(U9kzYA܋,v^!@̤A@ |fX2rO^S:(0qK*C3FR [0x @p4misT 9$bRDL0K1!;Fff%Ф1%\FCd@䱓:.klׅX~EoW$<Neg,"B- P GeAR8]m&ɫ#6иM/#8emWӷo7~J+\j[8GXb!}yHŶx)vId=y}h s,dahe]Ch/c_ymVx5a2N٫N?4= 'o9ѝ&`WT 8.i NK{>+t>t4Qt`Le:ҋ*e:Vr]e:USe:USe:USe:USe:USe:USe:USe:USe:USe:USe:USe:USe:USe:USe:USe:USe:USe:USe:USe:USe:USe:USe:U\e:$ӱU72X峗JSe:_Ly*ө2*ө2*ө2*ө2*ө2*ө2*ө2*ө2*ө2*ө2*ө2*ө2*ө2*ө2*ө2*ө2*ө2*ө2*ө2*ө2*ө2*ө2*ө2g*ɽ`#Ӂ?xZ{2XM|2Y]TNTNTNTNTNTNTNTNTNTNTNTNTNTNTNTNTNTNTNTNTNTNT3\{ YnOoT Jmjf)gx~>d.XǝvjE2ˠ|a^;FCZY.]}g c[ɔպ'*UOZzbaG{b,-c"SeyΜ|Z{+^2iy9KXeOXŕ'jK400ٗ fqYǝ'zKRV_lvMT,Լoo6OO pvMKWtA}_ kܳy؛'yˉLʞ+W/Jzzb,)Q|_B'krVX+':*_=1 ˞g՞ƪ'ra=1VH&Rr@fcNei)=1HK5kᴽq'b%zF\}}Xƒ擎J~|2MΩ*w/J킿nU h3䄸Lg'mJw=J#&_9 67V?2ܵ9ཬTL?=6usP9ݣp/M"sK"k|ኬ40\I= Wd W0ڗpErY)l W_aR챀ܣps5s} Wd=\Jp+WF ]?kWdk%\Z͞YQ](3գpeh]ћps JWڿpe(\`ێ`n-~v2p+iЯw-*BH3t9mRޣI^v}xwE'<ٲ٘CifP+:N9 /3 I/`q}DMN~h=}j~7o&&5aQTͻ6yލ'0ֆӇf X`hS%̧gw+S"ӭNnKNu'͚0*SOs#/eQ猬zY8Թ!xqpQqb! `~~Lƥo2 .M6VxaeTu ^M|3}_e6D7ltVcӊ"%Yn-)fkfCZ.)>;|/-6\^A4{{x>o#^n25.Mٯ7<xA xj ǃ6j?E3@gOE;z_.w{8K{z冊˽(\~&g^z܏/8{`E|rCz_9JӿOrdyjAaOFnfdžo7 o%|­7|>Z>g/ rp~p>P^ 2gx>iAOvny?77߼nuRzwtOwt9{񮿠#N;)^~V> ႋlv80amj|>8'\(w Pi\scp;~16oƓ'~aG6_:⶟m>k:l`WrAW.m0/۲`Wc+_LFM;~뛜\ ɬC RID(!U lWaG[h1_t6(~8#z9xöCu6j?49ˏQ{҇m1 ui16F+ߣlXe\RA֪g_.KV*F* : mhgf)m|pQ&h+`fW-ܬr;UG9ύmg̛{pvv}/Y!.pn}8揧xsrr'ffѠN#j% Z Rp&,V}Hm5CSz_vqZiRʖedE[xB[0fhkUrITv R[9yvm|q3z^Э| >7coR2qeK:~srؼ XCgb[~݀+AŜp!z{S~Td=?/˱uZDΖnUE6\P]:Yv#=SF=_|H qJ%4T2$vkc&ˬTsm [-wV,_/4ky@xg}Z3fRM|Gpqzc[WL;bI[E__VŖ#r㵫_TwnW䯬_Fi7mB7!f6Fp2Nh49= eXhQގ;+Wp-upYPC xګ6ݒDGZY#Dh4Q;HqE!ssXz>x/"5F[U-oT PXq,4v1b8%~IeZ@yj8\k4h;i /Vdñ 챇D37)pv&mec471l@,\[Fwjݣuu޽J[ reJnSIFVH[ԺSFwW&񻋋%(79DLՅڔLlvW7 ).c$"U]W{n]CZsǪV4dxT |(+ >Ea4S+9c{-ɎnqcuktZ/@PE%Hӥg(3E(=}Uͷ8YƄ˸0Q=~^5sU-h ~?w~mT&x{d3G<ǪU5ATb:FCT [8ãrXڼۧ.@<,$0D)/Պ-Nڑ̏z<3|-T|[Denoq44W܋eSBPI*H\Ù# <9 þ2>5v(uf:|ǟ}2.%JDR)uGQǩQ&? [;3\MpO7l`}IA `IYPIkWR9M:[wq2"]#"rG<Lo5%0#buLv X|f%b ,ToԐSrMET+*r98rXkZ{ܽCEfD6vzҬrׁԺd^co p3wے70.WϻfZ/Y-Rݎn%l6bDRRyOq0:+#,_ifh@Yl>w_AR%%3BFd< [8>yGSHɅFPcUnd1 ZRj/ILNT$ȨH i+b*պ窦QCԠf$Bv\7FT$Zd^/jX2p/vN/JUaQo{99U]^- 2~ɼ,tK2f D Zą:qqR9GW(*.]v2aj | \xIz_r46x0_MkPׯ%BK*d+esl3(vWOiݴ|(ƛF^(^*&Q{PvT\J4 b,MːLEzd < (*碤䨭J*zG(v8pHtAQ=YxM%J}OΈQP W4D*.IP2͢:9.b0J,{%P/dP¥7Rbra+=;5K=2J.^[>+j ;IU`$1 Eiˊ, &Y~OwcDX5sUL-Đ*`J䬊 7X;X W'j MeHUn/-%ֵ0۞4n|v4#h64!.\[^b=ʋ{DG0eHIAp)SW\eIWͫuJguKgPcUuQ1jI tG5s5WllE>V@կةi)&|Xl2C8Pɠ7NAƪ\006I qJKkgɋ>̣gyb HZJ'jU?y_q/y[6]4Ueq+zd^،eBp`Vf3#RQRюs>6Y/|lCԆC`p-`(n8D.[8E*]3 $xUU !Za@'KQ-+QP>\"*\SR XY'TQU+KD1OwFX53[RPBnڏ6$WSʼn;|ʶOV |n?8u!#xtQcUwbB̠ TOM=RT-޻BeM/PcUn&K4 ș^q/8mʩqJjbJzd\t)ԯaP UE ՚PUDR>Z#dL<E>VZ|OUB2Q0gۮhf5Y2b5MRLp[>Ẉj7cWDuY%[&W[JYnnuG20lhoPȂ{8A5VCƤ"zd YZzkc} ?f"h N1vP#UsV@+v۰ۂ/8ܸE`10m=2 n\r0ʛŹnJ>(A(>NUUQ=2 zN)kyu}``! ٪]vҶg~.ɯ'dTU9|- bpx;\/퐋F |q /<L6mVSn|n(n((1$t?}JQrm0R}-`dwqWb"2aZ-qP>(*9A*'jh0#`X<.tne0-L(٫^&*{d\ )(A 8š˘'i|amf~kt\e>@__ KRNjFzCU OCsՄ_P?=2 ܸ?K>uƤG\Vu4pͼ`֒o/luƻnR͐cs `-%[nh .j5.ECj4Է%!`>,"I7Oxi`ae2 Q;@DRn,]_`vK2rGԌQ(x.7w0kb@f|ƶha_^U6G ɷ i*+Њi[v>8]E>VV8^\qL' qj3ܦ)!]801B=2JV/?6aZr-E xUs"5koPA't z (Jg'7B5P\.u|y30yV\mb%G/nBZ.0=2JuX!W1IapٻK!j2v#dyYȂ!E)[l22NϺqCY=51)ecr犚GZ5)Uo;ʥ J3n_c= 1+(+ )L+/^O&jɠ 791cw?xEЂ8PW{˃4cyI: Ķo"4㜖WAZLuM]׭?Q8D"O=g=(efOm|ݐX5gFI-"f$HF$RѯfݚX5 ^ZⓤVw& oh?/i {nN!`i3 -@Ռ((VLp1!nķ!Up"N"xD cHp%J^GI]Ǫ|Bf9 GFxJ-}rW#d߄ۢPP%fMxܚ^bq(vMAۺJEƪEPM^o)KWv yK u%` j!2fPqjRm bC!CzqK,bӖ>>qWs-`d!ylV ^e [ϗm>bWlzT4Hmg:rx:cLg,|슨ћX͗DRʌ݌hGM׎UHNp641ĸ"=2*G>VۇUW|k]_/h;KUT :&. yGAd_0Pb:8ZK ~Y!2jMzd<(*mQL9BH< 1*.Qr,pn,fH~*Yk8P;dhHhk1YeH`Pc5t1>jԥW()3CT QnOWڕ"9خEר ߂$..iV4oym`\͜kC}A;n/Xw/*y\o\^~%72|'_~/ F-6tTfZ.8p3DReSL wz&o5/[wgS(n`ҝSp(⾅;՚N2E(9-]7!f>i@~vz8ߵyo~m~S}o-=@eˏEi|7w/鎋cQ1}z~Yɧra0p7ӻrgDG![3 ӧ=f@ sL`T)T 6r8PW!*bk(7 E ?iP}v3L%8ruBwGdZ+{jKo.ˈռ焺{-YU:fH@ = $c^T+͖VD<"\F%c#ɘB2 0*BC4ʩQ\7NqXOi'VyOAFq 5(Jj}~~m8#`Ai --S`lR8'#dhQoq+  l=iVqHTxɼPP3cǙ}OŠa]fȖV)Jxo򞵷qɿbXv7n0 쇻އ,HJ4`bUzx \e5*'ֱB07d] o j.D$ʼnZQBu ~C)#bps s#vhxX!61 &mcv(}LGT& 3%u!Jآ["/ɴ6Ec/o;_/&];D!bN׋[T-&y, g$kw'ʉ)hmlw$aCE7-t+?^ 8MNK ӻK}  ǵ3'$4jɔ7MO[uތS7\1Hh},S8WnΛΞw."5'D?!QG89ғC'kѣsD.>"/S2/>! d]o NcgOݛiP;~23r ]Eo|u*%*owषzNC2Hzd8jsByNä>ֹy L3M06'{@m[s hC 0xzcS mYrЗ d 3t2Y.cZf*ƖYMmWcJe^?`6d "$L5jBUxQw{4Bp*[*44?/yq3 K)yq&IQxDeiɐ*RVcbdݴiJwzq,kƁwVgؤ`S0U5 Ϊ\ᄹ8Ex^ !e3)[D|& V=+i;ANc}evVJW:JJq&PR cSRQRקdHN- ~ b':}gUIW{ %, iݗ\.x d%ߚ `%ސ1%.JpR++ SpAI ?=by3Ȉkwlw&̈́\jmqEXx/X `5d "vR+ǵnp^.uI Lvb[HoDac/sfP{qFRkSY'e 6Bmb{4ϑSMs٬*@yqJlmR.R1#%Owyn+L~KnB 5 ­Nu,]'?̞&_}脰ޏo0|s;&f?jE,Uk\Mpp g2zم_mu*P_S#ם9 ϽKg0L;sH u|W5h@ۍSozH蜠1Ս6gIy{j lrF!l/Tэ pzSɘ=?Qd %K&.6[Fyn3(E6JׇԖ V!. ٱ^T$7I#FJk3%BNj+$A2C1{wLJ/`=*!{.aLkb;ڤcuhPg5 /$)TV+FXKm~EB.#㶗ސYvyrgMpMovgɄ!Hݽ܃zF5AP!cc! oT&*AcCSؾ$\ }S$g %!9y'KxJVmSo+IXÁJ R~sV8&zKw5ɯds:K.Y`dlta! EV,|Vkqyh`owΞ=$Y3mhiOPc[]t(L'MN+mŽ>VEx*_X2D.Xq1Sxm#NpD g4d?ht< C qBN.t1ҫx03P҉Y3.T[:R.:NqZ q2q穮'j ^LB6)Ex[I)U(H=O^+O@ NCŴ!c1՘t* *dv_ Rk{:XЙ2gz ('cM:AXWC''c J9zHY`Y4wJ[,sOUv>51qMȱHoEV4|/ SY?[2Ui2~W0k]d"˽S0;U4OFo\sٟpd}ÛNQ75Uκf]Hj8hpv_L]/vFULu,nOӥVSn:u2 cQefV24S-V9~8.]K +?]PWeQnVGi8?݀0xP:⟚ɨξ\Y:۸V2 j2I` fV*il \ʤs\[.VAp* D# W.dLm;0ɳXȓo울uFAT TD ḛt2L1SI0;]’ck:ĀNРN+k jAUd ;I;d6> |Mډ ;m *Oi/就*"AJf9SE68ԩ,s*y G\[s GBLj b#{c0GHeDp`fm vX,(;/zmv2s5+܃CE!cQX]kS#Z"+k!,AL} GPzٕ2a;QeN^y!dWd. APazX8 q)p9C1D݂i 83mA,\Qd "|, TW]+9|%׉P8*?|w-]akRjYi}֖AnqMxH;KpسaHX`60-8#5MJxxRr.0wPMŠAPG׊R؁-:yq~2Q)CG/+24`qwJX2'8d-w15=L<:JfXe OfgcT&T v(# *7=/^?N6l7Azu3|MiRզӸ|2wGDrcL2VsdA*LfEL Uq1E9sq9:M|zi>ԺQbw77MwqTMƶu736sq ݏc _ΎL&7m͊ӫl8$M&kqF0m/< "Xc<& Gss SlI#nS-u?zŮ"UdWD2I I Uț;$ uRwn6_V"#<%S{WE=X;Ӱfp~i<ݿ >TȾ?tUu>6B6Ùs'?kq5znt:{-33?~Vwp?hiҸ!D(xaSj&E.!u ,BkÛoC;s66ƞ6={+$6^z_('߀gvo[uR@s.BB|}_菋Ǘ@R |DisdREN# YkcC0?n|](_G?GLзpa]= L_fϟPd0%/h.QH_ ^|sIK>"3%F0N R6Xl}$J,w.K4RrjbV$(,m3DsCf4Ri4‘Q~ŬzfpT1B T=^R9xq A|} _$[1U\IIgGTIz⿨B}غ䪕AljK굃q08q9DͬQTpo׀蒺:iQ)!ξ<9Sֿ؛B͕;YS2͞F5DρCLǁ1Y<LKdI|q`i#v$J W|.]uN^lbPLgNԃ_uB e7˫_KH2+-F?r|C~܆VQ>x\_?fMeEBZl5_1x˦k 3lGr+~!83Ű\>Zrt^3~6߭Rmd"d(y\C\+ҷ9_i9-_ U,߭sw\Y{(߳ d7_?j6rWYֿ,+ϳ2_3|^_ݔb?!/VJaR¾O?|V/ϳש2I ޙ>2q4%)xOd("XJ ILۻA5(C|,Fx|RZyf m\rJ3y6G6Ye/߮B\6!(bR}Y)mg8F']8q,of_J*Q !Z3*QTQT1 _R2 ?6ow^K5<\CCJXU1"(#3RVp2'Ґ'ڐ;!@H~[m9ǁu!ieMł} CaH8*sdq;h&1|6%uhIQs_bOdQ$4YCaj:KR) <S"ZJMI< [oɀC=k,SڂJAB̊xX5‰(Sxo`O䨏 :LmrPOI,[A2mLLe"+$~N't"^G2nHC|>4cC20Żl&о>urӨ5Wh|,M16Q6vz5ÜyP?i9=;ĚeU:d|#dB !xYܥ;Q'l|ym6D*.IP2}Ѝ>ʔ0| 2%ÔWj`}uǦ^76}qUi SR,69l^:a]]OˊފpH*H#9g3 pxcnLAa?ЕU%:Co!z+i(UǮjyC,  q 2gF'!͓4F(fKs Rǝऍ;7jӷT+aU޲oWNJ{{ jI]=4؅jc4aW?$΢ִ*⮼O<"̢"5絪PnV'`M[4]mյe,VbMxɊJBǥArHsCi{M8*d{uhW΂;F.E+{O*OYMa hQʍ\չ-z9YSUw%߲CsCMs~Uᑛr,Ka {"f٩DEnj ')?,&/eلt?`zYkN ~fȈo$tWws ="wsKTWv49рmlv x4.RVG,``hMk8R^d4&, 0#M cZ 3wWtԃ{Y/*K+afkkx$ % 1b+xiCGǛT.ڠ"xl *3vs1eD4leRvm~s>B/nHFy&Z= $ mߠ#Gl'OΉك>ۭ32N0)O; /b{ i|g?x' rJ)/|U=lPnJGX7=P 'msO>?@tKC/g~o;\dm&iIB aJ7Y'޿Z]V8jKѹYFIcR8.l"i˻s!3aQ&Gtxyyy¯  g nxJnV80oT=br1r4es<'>Ȳ9iҴ9P1pC.zz͎r7Gjv,X ;exYSm<@ݚ0[Z#Ƶ`sِP X1pr+á>DlyP?5.wydu,:/ e͈M$pPiH%J(wV<{FϿv&sT?hǏ4.eoX\蠭)zM?P߬ȯ7 *?chosgOw0Ԑ)"VbP?duvjc`H3Y(Uܼ͊+p1S;$B>a`  :cpMoxC&{ 8ʸ닃[Y5~eJ+ ^~>7bs[0oX=E❵n = s6 raߕGܓ]~y>8\?WG˅ݧCf؁CnH`[I4 $ڱ C1TKUxכt!IdpʼnWc~u:<#<FҢ -.nC r & 8Ib;jIae,Ik,eBf7!aBB~! =j$'Kv%3({,BĦK2D SJ0ZÅ$ïx ,"&<>. `sVW˥ئዹ=t!n>ux U;)i@ʧ G E)bYx0|\XAnjy ,)$1'|F XkmmnXMˠDZPά$QB8#+h@K7Q9+Gq aq{_QE1˧c` ]d= ch_Dw;EfhV֯onX8øK0`M.˨7o a=Sb0XF*5VD`ym @u];#`IcrFC;Qwᣯ࣬ 7:칝LSwTGۜ-⛜׍PmL)_4Y5lZ85Z)P>o{aX԰YXZ|ۦ({l-<5~yw(,ֵ2QZ !&cGmm1p ~{2ʮɷ=@e-<,tyn;'lk xKLNɲmFqZAth e pX3auƫ2ach+`?_+"=4g>c$>8 w\/gM\*]Vyqsk.@$00i5K(g,0nk KyVARA RumR("$h{rĖu+]?C[5́i<+eE 垯sHA :(FnSZ{t(ui--brڠ dm2X,T5z#1'1~gV^Hyyf>c[.JAd\>L؃s`=dQy$!5Mȥ!CĸR8&q&.{ 3c`߽SR{ 3@EEV" Sl b :=NJfY_g=>I5_zl}[x ,) w<= QZW9+o v.M4Vi$=XK9|&U>O Qs,PwLrƸ@'v@eUb7H޸ܥtuV]л,x2R?5U lڜ-Q$;cCa}aE.A|`FgԄ;%bl7"s)s0urXvd2UcG9i'`sw Ƃɭz\^xRx8R5^9-8vK2‰YdA;,]YA c`5EV4ET~ 1rfa"r0HyMjz"z1%nIFzś\m0lhݫ{rai/zld-</H=zqy}V,mvfGBAͳzhΤy0oYfTiԻ\7c`du`t-k旡&L"-E2' Nq32_4=O焤o1pxy>UV3ܽ-<x̞:c\ FmQ3{TȲEnXƱe=S-K>=:f4[x +%z&7q2x44;e۬t*7Y<𩸶 `2GJ%^q~yJ7ANPH7Ǚ_n698~VnՊ_' II6il^:?[SV"2TF[ob;o%vʳY]O|򥿏'ŇL|/L~X/qno /fnd12hÞK@]_-cLzR&؉p`IRS~;h>ƃ՘>X7boL(Eۼ/SL%QrT꫅-[Z J,_% %`\T М'6Żk\Ջ[iՊNY3OcMs-10Z,g+^+ {tlr Vca~*a@ 8lJ5ig󗤻k nA[lv;^x+WSY )\8g˷A\TRl>.:_|Gf#Ṽ9qje<MSH ?6i.L7Q+`i~baCݞL6hWhqnVyr %rge; xkY$R7v2~_ш&9.h(9^G >JmD jfULh^Vۤ$Aյޗq=*Z#X[ƿ5-Ϫy7^L@Lw_㊐oD€)OƈLxȍ4ZFqň|H1S\t"3_.`oώ.O3Nԣt fj=Η+I'߀)ACDe|*ȠHDu,rQk;nrIjuՎ#2HL=s ,aהQVsӼVDIKTi6 MOeʔuENr5#SK-NZi,>jm #6#E}]30#S(OY4|ٜU>$5gzT D|X;[ay3)h܁&ݙ럫\J෕[($Gkm=uIKrxV(jgPU`I:2;)-X̕ݙC)eqi$Sy'Ma]Z9LZ-4%Ekp46b,ur~[qkKo p~!ysHZTb\>Q-܃`w}lc\9@g c5Jid'S6Fk~iK|>RS?+fO') TyI򱫫4ʍD_磮R#7OH]0FLJ%O&a/=TiK種ʭoi Bpߑ/ւ])O~s>B L S|FJ9MH:gP(^RYzĻ|S9 #'6áM i| +LX񙧰1ET*i[ǀeTCP( Yڼ"(qq>#vׄP:|ҥga":Vv AºŎ(w? % xi$h,r ŘJ;gP(­[IROf)v9Nc %kF0(¼1רhN|5fWӛ~9J'vq}s)Pri(%`XFձs(2.}tAxpIqqdRӼY ,skMxaheob R Mq5rO9cQFeMxMT10{BpN4Z%_Y>-q83LFHspEQ2Q2(F:^>BiNgJgUxb*~`jj Ax!{i]+M!&cGY}CxM-OEmaΐQ +%ayBUD5եs|(%l)a8{N@k[duϠPp!W[9 1Sscy %+Z P]ü9,(JkkkFZ!X۠IcFc "w*-yu!-2d 5c`aKHHpUZ=+ZG!.sPic )1竴PC)޵y@$AL'nSf!&4S1!ˎWhR$6(a҆bjbwDuUZD5&&{e!L!Z6z8]TikL0m+'pj 9qEc^vwҞ:S񇯴[~$Փ9|%+^~ tfS\}^OsOJ}ap%=\zwG,j \9qj \9itrRJ 0!ϊ?0tK?~0߼?u9-n4Wy0Y *bLC3[ӢpXj;*(FqP^6 L \4brĥ2ՄDXH˕ArNRJhbe4IcDJcƔ#?nx0ɗ?~ԦQP!654(1HjTR$cicfDp&D<6q@I55 ĸ+N}7p?h׷gg@G73o%yoIF ~u46tKঋ'l z7}Ǭk;[fV=,}<|GՋ{w)8xNn~攧 ~y`)7hGg>z͐z)q9KO+cQ4fFҽ+w{07x>6 AEKR' 6E N\ܶ["%%\ _Gd21%aĤO&뗅82'O+Їgs0TfMw}Vnm0g+qA<~0#9)W"X6 rtݟ};U}S/7%X!t2v;lOR%Q~~40WNl 0s*i(ޝCуoe[Vp. W|W)y=rDX>rnƧ?~rNF3 qT~@m%LS LofӋN{Y?)ʷl*jUT2C浿p}xHozH0AM:E"،}hkm`AkeD'x3Od8`"-u{ւ]C޸XL78] 3G-k% طYťR3D0i (%nVEqiR"YoϠIz q .Cw㉏,m'G@Bp_(yKޙ isL.$8;'|ơQz}~[%J$7,"t̠(LDBH(bSlD]_'MX`cB)71&(C_{Ϟ' >{Ϟ'< IDA~ }0r%{,~όi:c!3)!%Q(hCjCXh P$}m'Mvn񸛙Cw4vsHm1HeGgSSU*26Gp!D(X[Ie, 2^TY:^CL&`.+`Qo{6'?7x3Gfw՚Λv F];+OȫW2٨lT0 ?w/ӲM$dǔ *X@5>FWSjcڊRh8] ]PhZȕMn$ 7:;:ZH -bJ[u!B@ݫ,%ju ,tmY"teӳ x\;ͯN!FM-qRJqQsӅЌ\N[g[zNK]nڇqÅB^̢dfeA3ka“ 7\tIfZ]i˲6VGT*MD(aDupFvVLj.MYYd"S'ӢJ5M &a^uӂĩD"oiqKRLgH=;C 3-\?-\kZ;ġcw|giX Aj'hΫCN\r0I )ª?p-$LXs ĥWNZ/L1]#\)Y@`tcʉXS \:\$++N{{ =~ >'W m~&G`@`зg;d8EW-c0*Cn=O'>kUSxsl5/W[,?LKke[r-$q{fֱ2^Ω#43}jL9lXD4N\՘'4fTSh5Sq X#,3<"KH+(zi l1LKe+R pNbE)V :n%5[#΍`0 {y ?Sf;nێn y魱 L^Z?^sӏipC`m 21{\iIYD (&(Wه֊߭j~TESmj]^>O\*ie?)Ņ-Z=\.Pc +'֍+ rrp夤!B! ?+c?ĩWd5evuUq]"%Ժ%&yN'C=WxN'd99s2 +arrRqpd8Z4hĥ1ʝIYrZW6Z^:\$841 dHxBR UpsؕWHoRBk+͈VApe5@\sH<%Uۈ zYaQzfO\~0Iδ3ޡ{pu15 R7Z4@ZFإÕyF".m\ +V:>B{Jrwro|:S&ژB(90t84= ŽBۙقi%kh D$D)D"Ar*^x4WJZAe@&'V$M&"Q)2ZK‡v@R.(nMkR4\i RSq qQJ Vj\V!wǓ(4! !,IlU(F*UbGٟ:|V ʐE*^#$*-x ps>$ RL\ntY*Q}[I0HKFLJ$.LT59i9t.s+b"Z.FZ~UvrCyU%vUi:]D tJͭIy~VQgMv0+'] + ™BI&p,|Я D  Dk+Ÿ: B%Щ̭stqtW*rZeիtW.KSRXmE !ʴ,rNPW:Еttngީ"1'_kc>>C3r+|?wv Yb^QPx5-}?5d3pиoOXAJ80I^,%度gm;u R䳶$4ҩTE:Ou^R-̋ßb^Tiy!!k!>ͼ:\u^>ƒ)RS"|}8_;dE7[iRx8@J4i\ɦV\<\9)J .Y^:Us^:quc U^:)Kg<\]\)Npi~e(S{06$|.^bCϰa3AߞךձSKΐRJCi ;.5aQ\r1ZVkõ\qF?vQDCm+KiؓXƍVڹ:x"I ]@`Je)VKhگ_cD%Ipqes側\pWOzHRϾ6#"/L/nb{gæ!KrXJ=GJv"۲MYl9!O"'N\O2M;i8MR*D9]Utu 3DDWJ` CWZ?SRZLϝyztj30B\BBWJZŹӕKeЕ\C.-7 ];]))tJ 9 ])qɉK~2ttdDzztEHt8+-.LJK+~3 9b2 mѕW JIɠ GAѕX 'Ԯŕ +i'2h)Odxtj:+-0N];JI8wRBb BY- eIቼvR3 t=] =T ѕlѕЕs+-@!]! !wx0tE ;]i)w. CِJ L ])q/ᴒ᳧+%%t.@ ])aYlѕ f3=UU+)CWr"؀J ,p6JsgNW "NWJJDw>JAŮѕ `J3{U?t%0;meQ~yfjT2.eSzov=S@[k#?^.BQj"OX?a (~/eizO?MgJ?yc(ŏzmMe%*⇛L.O!a"V3r` rqZ `@=jЭ@FSrG" k$u?G"V X|lVKJAv~=xA׻s2ʁ$(dx C2"aHHm[ߧӅ9v&pfnw 0l_|WT( W̞컢֕)%-UȲ"xl{6GxusWʦwhZrc5w1ԧ z`_W߆asg^7&mh@HN 0  (SZ2oyoO+MHhE^S(do-OŔOvZ:Zr$5~Σ#&QUh=95;l];ړjZkVԆGj?Md?MAȨd>ǻj膽l /Q\]K\˺F(ޱ`pL9^ɤ K̄ 6SHGTؔWi35<_%}IX Z;l+ I'ǞBbϠ\6R%3nk:gy*ۍN]p)yב!BYS'(  Aba,}`r ׌S>T$HIY6ǫbWj#UJwmL I9"U9]߮>wSח vAj>vRẬvy]D;yk{s Q?U'{\I+oȼ:ӡ7\gj0Êj -n6~Dc6 uXJt5x̐:aFϠPyokANWJ]O) MCeHNJ_:M*.zZ@ qNT9yx/= \-,y辜JL=l4izwbX^S }MI:K^ɿq1 4{I$s2ⰜᝤGWJ+Zih]3ϳ@r2q,9X"'.y󤋭GBqx1~)bj[QB 1@&l< k\YfaD.W9Hu צ[{ʔQ_FTb̈́K2E+oͮexJI Wʮ۪L4²wVut#x"{WW{m:\>Ww .iu0/ƍ%Mgj@UVJ46HdoAu[Z!%vw;swusg;5꿼W;Y[~нۄnw+6kx^G*fR+(&֨Xv[[D#h-u#2P: 2T5W;6 :nNY =ItC:{D9yN'NrcF:)+Ϋ+<]䮇;=*$KB@o%~?dM^,B/wb$|MJgbUW]2!5ncH"&`H(FDXqYJÍWW\V94`BYߘQd0DIB8 !)%H!p*AbAx D(y e=r?G m \M. dg/CY)`aEZb zD1N詪c[kR=7!H.F#rGcX1+j OI >Az55.aSs? MfdM (eBq[3#bP,y8JJ 0(D#RAIC|0&](;GEPϞrj[T7=hc=h6}?ӹyi+#hv1h<9ns>#c>]a ym O6I.kYD笤־V@YF\}\L9z䏗9zFQ}\hԆLhLJ{ Z>h{И>JGFܶjRL;%nN26Wz|jTq8\U *􋊶PEknq5z!nn^H@/bkŅb?\ ~*O^tj9 BY<͖E;ZLӸ|.bˆf~1᳦^Sޘ"V6sSNͥX{LE:M}Mj}и ϷK~;4:VC9s*K35n&%ױ-E:+ܜ7t3T oڐǗGN#Hv1<`77u bDrEbDˆR@ پcGë 5XLzh NpMP=quOj{3rE3 ?]؆[{0pV)pzW4lޚ5;)0s,VdU^e|}.9:PjF <?_ B:EGq𠼋Kv*~K|tVA0WJT_Ab @[saYš+_B~T\ӑf"hT4xQ>bj]'z,uxUװZS7f溚 pi;|\h/iHJ_P 8 gx 3P=.-}5FS!Sw :,|Aqj4Ͼ,~sa>Edx]6|߶}ߖ`5ɾ\SWn  *I5AJU_B!kɗxYϵT3qqnཟjm=:1;ח`=}UN墷lSY&/  $HzyK:K[ڽ3*\nmn\ۗީ|,ӌ+0ߧZƉսOe/ߠqjLz f¶+|"uB5n,v UO㜃(Zl^$sWjڸISe0k7ӓ94fCQ7rv?1NxXViSdw }p/͊F֯CgBۙإINJSTN))![& ip lr [WkJp%6S :k*G3^r2N$9e(Ihn,ۯt"ߏ| (T5ZB l4Mw"TWgP_}EA}A-,cHN"4Phc.aHI8&J!Ʃ$ƄS!_z4ʠ-bOf3j'; >+PcR|[<)KMMϠxHkL`t,;pʒ%)^D&1=ڊzZ7AueCGgwG`c,oA'IH'bdkRH9{A"rED5Lp.7J'۩JEq;w{mSמ89_Y/2;ͺc$;R~~AimEc9^*WBG{kj]\wzręo/4 AreΟ|>]ӠWKI̲BQv&PSl&{/\Ӈq4S<.(<r۠:]Nԭ L|&+7:?D:+ʃoxU}'ҏϣ+|1E2#dPB>"cD Ki沏2ڴ{l շP?Э}pRaDJF Jh:}e5%)EO/zJ|Oz@@bq1JIpLQ!MG"( ƣ(aDdD'8}kGWv~Lrb1XOY$b9=n}Ȳ"io OPFH3$\8جg83Uuݻ_va(31jym,dGh}r PZRKcU 3V]v4bΖ|y[=!FKïoNtGm[vO`=@Qh}Cwg^"+{vޯG&ҳ <{^ }e=.(0˿S@i t@i tm7*d.!$g+;|"Pɱ-.EtZn r|{6΢Xf&l?o-w:2ɻLm>?L+.q.{T蓨Xv3&"Bp&Q:WMl`']I/@3jA+$pu2ZŖ'Y .;H߉At4"Z34˯@!p]:xu6ݩJB)5KIFAV=hlQT16%QT>8V=N, C]^N)NU2 _:F A~pƹe1Їv .wOYx+@xH- Vhe Ak а h"ؕrylC, %S[]**0T /)6ڰg_DV a9,l1--N?K.faBXIEɄ(ȇwy"c^ۘ FeNR.AuAQi3d#kx얳knR-9YW,t@PA5#+gQ-+q:t`2fYv}\D A$zǘ|1?j$շ:/sJ(лS<:VKuu8(gJ&dkW?onMV]4mBok1)|Ӑ%^d~^⏨I;vjQ_UEᖑXy^NjtR?״Y>T' ' @Q۰h&뷓O~/:$ N~nOWMHOߜ3io'/jomin7#7dU}pe?]Iӛ+_5[Q(߇z\@=Yа[މi֯k;i!^_>.MWP=1#]])_Do]]Ϸ EW?pͷݺUI["$ dt=ygZYcCtzW1Qc$BsO6tcQE+ cf(c~Mع`UÄ%D_cgZ=T>)Om,7EB67(0JJDvZDCq#BhgcTŢJ +K2GLCTMFKG1stZ!}]&n;]w"NIYNj4 +Z_iz[?<'yUWyw-˻=&8"h޳4T~= c&`(',& /v͏1P]HUNS]ȩg%;p 4x旕W* $v3jwe!OՊu**՝`{ސ,6L/[onfVJԟ ._nN @Rg;$ #hD =պtʭC.eN6"B"Q%U9!O9Ec9(c[rdkBOoEDbxʔ9PiUVvZUi[_؇C Z_ O!ŤRY.)q(A篇|{2=h(ߌVns887 4, Y`8i Qlo*QU3ս6gwc~{v՘wp[{1[^$ث:iB[~BVV߿dQKibAJJJ6HHL-XAdY.xg$  Yѓ^Ao;[C_4ȢJZJp($8{- ֯7{+ Dk/$#pCX`|ңi?RPDEn!SZgZX:.>_0Dc?g1S3"R*y.^|]ó_8@ecB Pʒ,k& 7 ܬA: jShB8>/+$C2|aP3fZ(OG"Eȩ;s{pGY ZW Cͬ@\)Rf>'(% |= 1I`>̧›TOp* ` GdXk~rc6i.W.E4]ҐaHb. ,PВ=#c8@KKt&"M"Z߇ uP5t,F:$:3Sd,=ֶRQ*dD5O)arЌ X oQJRч xQX43E.b)Q`w,a}I߰\ `$Pv_[0C TZE;x6lm5xuoX[ZoKϬs`ӌj{̾RWnl UwwW1.g >*}geܞf`-t2_Syy֐[ÞyyX( ix9RI]ds 1bYZSV%[j읱㻝K#J/0@9V uNF&Hc`hb+P)A8ް/Es.< B'B_Bbk/;*&CQRhMvl4:!ڈ& &"ՙ$rz!d82g&Y7`e39SQ"[NX;s*d!ZAًpaҞ\Ɯ 11,fMI$$dS\vZc"f@dh*[@9d]DT3"xPYH%y,L@G΋ja&gV!/=sEwgňcmג\k+F$lZ/Jy];i#c1,ؽQ Z7HcU2>6|gG9 ^TX7L(ku/{QY qU fK8A4NJ@Fh-L$^2K=C65*+u289 +t 클N+ &!& ^usӭ}ũi+B zH1@HrӷԜøL{H`Э$;в; .,7%9$zR@hyXT ` %J7)hȝ^+& </IJ>}@@a K e 2`.X$E"QRjzSvmo_gp*(o%5)z'G\5gW7'٤7g%VEmykgo)9I =VLzHD+V0y[K7M͕e%7;=&=ugOi)o!j]`msp~?Z/=F?ugޱy~4DNAwxIVӄ.Oæ XOvxA_%1ݙdik3n߭ʵf*UǬK/=a x}U?p΋Q ;˳p *$q?;$:M8jbbE2#@T5)cttJymE U8# X>gqNS-z{d\)Ր[댽 \Wݺy1c.5o_ {/xF]x+y:!\3lkc1F?`϶˃;G$OG}@k+aN[>͊j1]}ުNrM)?V>/kb&Z>Zs}ǝz9d^~ joF6SVG>ݓNXV<Ug{|WֹTjK Zz-1稽W΀3Ic}%GλY-J`&7k#?)- 'Hκ맆m6Z(Q( iДR,:ziB+K|UN?VHx¡@T .L.u|tngM؊-xN `l۬0[RyH8xFt kl v%Ԉ~/1Srr@E1eTd.A#g%8), O em  fZHX܆tݴ* ^_j[V;gaA2JL)>l03'Y$%`F0A,ChysH]|r1iIB4R*Qʐ”L& Pn90>;燓WBKvy\*XwUm̈rC ݏyGXY(iS)N V<'ĕfYt ӎ}:6"1H,WQh59UעvD[p-XIX Dz?IZ׷M߆(nLM5Ѐ!2Ǖg"g͗d |~+0{7h_16߱)bS'U;{}1CdtLMe%^zh ~G7 Ռ6$W^ANs#W,`3=,ܽhݶlYR$N돔l%wlɖ%L"9*XU_$6K' T~ V8ٻ"m,{C^޲=Oֱ_h_Zl8¨ T+  ;H~ }?l5 AnIYt烏#JOuӑ\'ub—&&v'uXqiLxdD]gJ;_o]sJॆ]nRt:\x>8"N'#t9fT`<5 U{ LWf:H//{:|]*%wmH B?:N:4ܿ8_n4~<-Jc%`?Wp_' v_~WJ( _w@yOE&5+րzwחa׉Y~bV |/:f4{S_8θcqyɍwzx/@ַWB7*@8Mo-= |e%붘0~܏rk_7sb_IݬVLXwM0k\ iҸYRdz; fe n+@.2BAk_eC7~ʏҿdr8'PUǑt;\-|q3Ƽxkgѹ&zymx 5ggy7V)+dVth)؆~ 4VqpFy_ML.``s7 P5ۡ\Vp.~rZoѬQrҸG&]mlLSټWZΎڞ;nˣi1d{v};bٔX0Wu;bSIP/d],ڶ}90B@oԍo3=;OI$J *!4ВO9ŐO fQM:0 f=l Q64}q]"IߓFdvBri;t#*w#Bda<67$("G|bTf.PV1;a/iU1I,2Nm}|y5(%A8v ̄N[oǂg3KQ ƈg;Qn!a U]MJFDi.ŒB9 s7}^7cq bDb)g`^k_L5HqrX[(\b$Yaߏ:D9]Tq<{C'|*B"}$qPzWg? dx"I8K"{Eb۲*˚j17(cSKaH| =,||y?+_{cq*?39Wgb`IX2H0N:~h8'l !ʜ궫_w3Q)<^iv%r~P_%dM=y%¡j$pqצ_:kmME+8C7}V˜F ,~,C^I}^hQ4aR7]`<M(iIUʙ)D3rCUyLM0, D_ÜW`^^ϫ:ʡ,*a$qT2Y 9b=Ϡ߭iʨ`\ 1fz6#砼"8Ui);TƇ]t)j$`QFB!Wʿ^8awJ)z`JR1#~3:S* M]M_ yŦz^.W/e9EɲR%YF!/P.P!Qj$,TIu:^010r(*U'ا) cTͽ%n+iC䉺?"x> |.Xk}:ǀB3ϖܙŠe֊ܭU7y0.BǷwQlGCl;r⢺RLw⁥U*S"@+s,([M`gb.ds]VUd#n女'hWQF )}88A Q(P(-Ҭa ߎ{6=Ӎ%|<D_N0[@NJZiե\o>*q*Ⱦ>XY<8F1*NTt}^s[6>=\iKJZ@Bh 5]|WqV *@>1ZcU{cLv}o%˛0EX[4U^=^K.]1A6;3r}'aeUt^m]UQXW!jj{lOrƪ;,plq66J?swbE"%(+ixζz?8^(E"2dH>zN%A\dHۤ$Zzuivnޱ{)hė=5= fi=r# bu֛w3զڀllԕ^$WƜooeT[xe$@#DIb  ok |[ 03͑O|;ŲHju9x,FB;,|:y0\WJkVh"a+ȅd h$" Y@ƀv*M]͆@nQ+ܠt=h0 Ն"L3 ƀPHjTrܯkCphd`Jet .M7P5M˔Ҹ$*]qkUUE!+ΝQ kޒs!"̘ۀtfx>Z1dFnmY;CQ!=42oBẎbM X+Hh+mQZںimǙZJMLXjvEEF_JCi9+; 0 w?KQBU2a; ?hAqw8!}<ÚѠL疵), , A|a$:kQ(O۬eܴ4ShRNarb?g˗ y{Iy HA6]wJ@L/Yt<ь; ręmKr֩jSFcnTt8 Lfi!+8b\_sxU$0#v]q<8|rv!HZh8HVC#.!9-0N9eVd)vS.m8 Auw6(YBaZnh1;}]-cb`{aaCujD)Q9 h,S+VH+Lm\U:GN)pk'ת޻(d648$vȃ^"m~; 5*l;4L:K*Q[Pk@-k s:vTJ>irZ 2i%% *"jVTGĨ۩|I uy礟YΜ2Y,/ G7(EK$)Akqv\P|n!q}SpR3>ܭ/܄ \&*y[!AavksvΠemNob:}oqh\m8W[x!h'9 T 153z5y"qL |/]ԭA?yZ0K|HRI9T)kdμC;7N>|5i^N#Z2۠NӀ:Jr Q_r~$|PMI6<-moyKTj8.e!w\ ӼT׿5!]R':N~k=sEpU1o+iq$ezaݛӔ7{Yw}h hЍVzJWDUH 'Yp|(~}D/|zeaWJ BVO~d/fm%cU "x"8vjʹFNRzBq>`M(zVԾq$"4Mj}46ۃ|I&Av& dՒ%Yv*A=43ȲYzQ2$(pQz.G?` 9t#DTY psf(?5`|oy|A('aPYI+Wqj̺dmgM:w# ݏk 's{YHױm LP\,8oK՞ yQyŤΤdTmOGX8xdžK_y[u [T^R,[d&O׊2O&yv12S|o> οf+_~jNFEh#$1$&YRKͩJp }QC9=4<,iΞWorN47* :0Hi0 (>ycVUΞxA,`(ZK)~*2pQ/.@@Je {z@6tLX%息4 *wŎKpF^P{`5pԻ }Za\ =5c(Nf[5/8a~ I = gVAEFQ!CNd9 6@ ~fEAt9~I,T L9!2xACK[09/j~X)LRq@JF6 RXD--6OGl|Rf`HHFE+ yuɂÂp_raw,JEB˪vccM7˼at3Oe? ckАU4~o&Y ^ D XIQ@`=o7/Pd{) x . A|h7"oP'{S\ sXjǓXa@1]?'K4#CNw2 8jkQD?bVq\A ֔~xH\ (I4۾5)ī{ZEZ;FH˜lxwZn|PJbe(gGai`]"O|n @)"n?Y\S߸є.TP4^+ے򶉝zn7/B:^EP<Uı?$py3l0(F["WhVcZGP(V(Zy!3-5gv#SR`n.4@yS{E"ml8cmB}m`ӮߵYg>&텉dԾ6,򡱛_^a#D7*MYnPEɌ$)7oC6?E*9/zrrH8ѝ? "ъ'F<ه"dS~r5g||ƒrtK A@ZXa$$@(zߌ$ʋaRf޽R0;W}`# pa散H86CRTE-_ B>S;ñr^6+ivpGcۀQe: 9U3wդ.9S}zjoM?/rhSsF(ĉ^0Q ֛v}٤u%^sF'3M0ަꡋnN[*T[#|p1keݱ,ia (4 @4H+N"&34X_Ej,_2V;Ѽh[cN;^ {:%/֜12ZֈQau2w3l_\QfGvhD#siV*JOi2YUfD~k$5z"Xu1:uG2qL$ަUn0|zNӘs~.S`'5~HRa fX1}, |Vq)1zvz|/ xN&vhV|~DK$SAV*XP?(sJFOL,|V+Q[Z/F 4it) B_>\>Їp25#~1~x+k؃ 'Dd ߗtp7#eY*γGRI; !6اzB씋 i!u/(no췯{ Pë!zG{ɥ8b}+Dj0738 X>wR _}ԗT6$tdT6nig$QiW39cC Q#Zq*xOؑZ _~hJJ_Hz Q οkmã:)%F5/E; ysT7iFf4 41]H.[@,"Z)R IUTU5XndL%H r4YUCϳ2fyE:||f?uS Ll_[wFLp6NH>@(x`у5MR>d .vm;HHP3H *~J!"[L&q`NU?2esbd 2Qcf'gu0h&DX>Vv{k?e^aQ!~Rn|y{[ Pcqg3=%RCNסwf75FC]܍a\Nܙ>xF)'zlUׁ_2{BJ,[zJX:tQQ-<>c#.1~x^e l:a!;D/ o{Rsu6ό3O|ơr#$ibn-ȖI5Uqaɵ1> X#W\4r NUcRJ358rBF&Y7+x$I0^*N8 V}U;T>2! {zG Tb^uz*2pQDPGq 7'+7iq/^.LV3 G-UnbKXx$R: +Y>ݶ`+T7z1c`H;NU??RR̼L,l)HjEq+F^`yL æVzܶ}v scr_'F)zEQF^{RX!f(\5N|sac Y) h%KAv8eo*Wb6 %Ym?S75t9zeխi͚E؉8tRy2IŨ yyn?4pA|;dMYܘ2,2ދ-eRU5υ>Q5CB!6#kH}#Wd4 ̛JXm#FG%sPBPImB1cR}QactyH9N1,F W= fsl6wOHI\^"(9S-pyƒ{@]߮zK(f8ntY1 C0 3Ҿ_a z0˞Y%NJϽ̮GC oX/8tZ~,o@S,|c VQ8"pbIiuٓpN}cn v*$4R"]-bm|cą {mΜFX !TzDBJJ*Utශ3 Y}8CBulƕPXxRG,ePA;h DQ ٛz;~[%Œ"~=Iu}Id"WrQtF 㗂c͐*T"`{S3(` ?< I׷7ʨ+G8f\"]7MYg=fd0KQ0x\kVSff9S-j@bERVcG/.ՃB h=0O9mc&zǜ쑥gd\Xʍ['E#$8E )b=Lߗ@hQRVl-d_Eb̴uޟ5};<%Ұld%F2-in) vNN/W O+[UI*S@Yhj\)va%Ve qK+ WutOy̰bYwr-۩͋@_ $B#3 !R!{XcϹA?O?~c N׆#)OǬv?%n'sSߚw_]P).;yz:Asp̪ )N0]#^.fe91ɰR9/FŨ1HPa_/_nrrc@4lHe1)%Gt35\/o.4Z?|5_;_23 ?ۇv2h~WÐ0p8A䴄qt6?F`ş@qJϫJϫbs5AR#ekq(G6oJ OWb4^yzQJ(uAdkdUYj1aa׊7:4 Όc+^iՇb~U*\hɍq6 rh=2ZshB"]?!ny S cjԉ)c , a`)`x<+ ](6C?H~آ!\U7eSo9^va nЪ-&D]WzMc ;>q~B)6T9΀" 9̾ɲeNP8<AY0܅;QҳYEl=ȑk~Z@s*Oj{a46i*1q-G<4!TʱO1ր3X޵ڶyHN ?Uz"AֱS:+o!Oh,)A{PRf !*e(%zh@q ʉh9 E1AE,  x(,Ԭve5vAdXT>3$]ü\OI= Ba ;@St\&:i3^I/V4vfǘJ8O?qz/gڀQ"ud/fuLO &e({MVcbZjC:{bol 3Tq$!g*П,N鐋ȪY/+TB_ppU*?8ds?wAW\, $}58'ѾPyI3u(+$jTZW[ DA[PAcXLIJbVKx31Vn%BzZQ`TQvmz3vyQxMR' ̗iGDiQH”N7%)6pץ X[6N-Yp6LJftS+dG]jZ#RJL~tz7gյIб|7½Zɹ _峟5" D}('2\Pu/!BH<.`ti4G^^k~WùiCtS6vލ@OѬ.W5 Ul>KQ7,ag#PQ*(D蝠7ǻ1q%Ğu]UJuUWXWCd2Աu6u pܦbCN\A5e-p``H9Nt#4$ !ՌBJ(X?dgeхdyBߚ6-1ˣr>3ÝST[EKBZh!*1-8Ph  ʷ{m ц)BQ0 KLALŚsMwumdЍ9C)EsBܪDv?#MQ)3xQCM*\2RsP /a|)+_?lcc>Ou/io'`ҷ̿75Y`$ũfؓȭ\y:`Vl1vy~%~E͸56}oWBj`3 =*ǩJ*)(܇hm J/)A#X}bXCBa5flR?6ӹTO5M$@P#%DA N}LM!`͘:J憪bAuEڍȷr ;]ͧ:q&_>:Cw x$Tp !PF`ͿS^.ĿlGUEP4(rN{kB+`uKI@1ShGjܹ,۱0ۋRhhPc_% Ehϖ`6Þ _henG}|osaQ'oow'+=y:6KS^A,vu|[L+ȄßLT\K [+u^ rx6KŸϔWǰ!=21^SEk1ю,Hd.+=[C)DzLH>ߥ3Y^5$(z\LgU=K!3J*,]~gZ+ 5D}Ṇ zyO(;QUVbVx,Q[QTVƙDPc?5+9I74,F>o~Wsgr s28' ' ay+^xxyzDWm6fr'V-d@|\ygmYZE{9^}&Pдm:?5,8\N۹Rźjv9`UۜWcm+՗{Iq7&vӠfYCl'gvOYuÂx7쬛% ŁdTinr 8Fb-PTA5)ƢCv:xi3X۱ُ4NTs/s4*5H<NʎUUg?Z=02ÕS'wsrڌkWR|Iqr4PU$ıA] h䵙P*^fץ"f]a"W aGy/ABmQ;{Y3~k-h޵Xzw&D3$` ZA"<4*4SvdB >p[̔!DQ0f=Xy4{`c(RMk <\ o:*gƝv18X@2n,^S 8q I!.nT7[fh7|N#d܏jBkGp%Z`=)|ãBc_V)髺 ) 4J}LRv0)hHgsCޫ[ΗrKف 3~4,R-(u4F2@`(4aIp3S8CZ'֟M3su HrxL=Q7 bB` ~˔ĥ&޻`}2! PT_9TJyE`"f̔xh""H]x*c,-|&m)L5qOeQ9@<š;?_EvȝX]QP+Po6 qGrI [0|=wؔ}IT>?[]u`&l|v:`Z~;P=(YJ($[lTe)QIiZj`neQdZŔO\g`@{Ύ]2*h`g1wQ}G;!HZ0 E,C"_CvZKBWP%8/H搨;DÈ:+93 dt:kKq ;;*p&c#-.P=2,qF\DVdcu7ZbO9> - <:2zUA {=aweiCm&6b&"4h"LpKqdan,aFϳ mgt`yp[=M=dr.B9IPT9F!9T4~*J N,<$8n5Y)F >?h&wVd32{НΥzcT8o z:xN)+ƛH)`ٻ֧F$t5A?xg.~uԫH%`f/KHj4=̪|rV~qMo|}|bj%#b>A:?]y2+M<F`R&{(*.)[C/QԲ_IΟ|kcџxƦtYY Va6GW KL64aWf` 2:Ul} |Ip |oW+ڋ2F@5ޫ`6+`Vcn]qôXhD;OliKBwf5IBmV&7I&o:,e į8CaE( %0e)<;1LSo?spjvdΜF,fσ0a(8=`Kg#W"+`iOﲋX2͕xZ/8Be9)s?S܌UWvwʗ=iOl;N4l,{lzll ȞziT-Ib(PhNr`t_yw~B$=Wjlz>ex6,:Fzƒ#ƗepQJM oQQ^OZA4zzV8IEw{\+ rx?)X״ԓ..Fn\ug4ň4+J±fa@Č?ch4j`':jhȉ柦m,%cL  5cO{[; CJCzB uq?skEe#%%7;o=-]5{2>}c}j67\V4]~:Υ6pRZE9Fv2̑A+iJѻʦWV5D"ظ󵖄0o|jrZӳ3]V9 t Wo1IJ ) "Ծ 'u8jIDvD4m) Y[`i H'aN~+CiAzk=oRIֹ>Yl+ LxRI[A]?5^k(zC%eAj;@s-f}r}=7Gmnز4`KW~=UJ?(JUk.ҕIyh:9G˟,E /\OJ:OMDɮ2gk*zg|~$%#hIm`O%%m[+'稜זsR!.I:"jʀw>>=i־Oh<3s$F]9 X.C'ѸocW~Y/Dx?A%{g.tk~L+e*Kz\kh5Jv5¹\- bÕ4[$=.LJd )IJ;d^T4bfO[Dw=yVM$a.(:Ãt KA t_WKZ v`4 hgrm~NFɰhɲ[p/T]gDe& seឣ{FUƟ)(]s7:1Yij}[[꽤{/}}d/Isx72wwA2n`cw{Fn ĆEygDQ[9I,86:'JrIȅ&en)$WOM:OB 3 &}5_t,6X~)%8R| n djllhQ>onC+x.ћgg[;g|4khrݦoK҅|t}ͪ1[u=QNc A{T|ΗLfqiHws(+N:2ˊsnݗ &8W54ކZok <خqۥCz-.ՐsJ"15|KD98 pQ UP+qd=kcMG96C]U&Z ff%jO~~| RPCu&Hy/ R jh nn3ͧptT/>7 %%kh8']:-]rIf@ \З,4owPaB{:qr𤊄=.N G|mP [Y-6JA']ESlmbRZ]M4Ɓj [۫q<../gl\eMYg&2;[BJMw we.\\p46 fʵ!, C((g* v#0dn\>8ϡP$Ap+(qA=SwR'U1pvXj.k;"<< JfL2B Z8l ɫo,፱ g? 'ӻ0ևcUξWK fهYuϿ~|VNY],S2=V<֞83YpVnbl{5Ywnz<9qec`Gԫׯ>;gE+<B }yt{e'$xUF[3vCjLfeާ_GVq1Q\>vg18(Cm!C+X뮱ǖbN[0`9kq%#hdm-x"j; _xCUTH<{Sפ`mjm7 |h^jE:(FM$E~0C#:(+;,wB Q,Ǐg`):SAxgͫڋ ]Pv5;=+a eMɗ9;3lXԔfEʼ)3^%թ5̅oYgbJFۿe nZ4}Yd ̲rFAy9)XP`>ШzcsϘG'3_ ЪhܣvG)_IB(kBn%5(XIV1~$u#CU_xc ?(TI50d,R"&1|"O'7WVȤ_)\r+wiV)KK!~ KU=c<&Ƶ D7PPo8jGTDv~;\MJd11d\a6sT/-iܪ5dKQHiTEn,.>MhF[בccſrfKx }:vL._OtgU<- %SΔ9_Oϕ'\; c\; (IDd||zQ6spz^pPAzCf21ooB4Y@v+O'9ܹ[: d9-eFg klrFƷ8g1xMt~d1aNiޕqqffWӌ!/+86> sjΌc69DoNg'٩ߙ!Zo{\='h`J<"5+]s1 ʃaPy TYBI]tOT~*FNi+%mrm5-DsSpRB}b|ͽfZ ,];RTc ri>#TU? v{1ʙkrMZ)Esɬ51w%(,7zXҸ;dH,c3Vܞ1]ФjyGrr5TvTAUC}z9.lee8nZpos+}2l^߿}E8̆k m%]_-B1bݩG27X< 8 LXXv>dB^DT˔h>YDtQYnHx̹+TM ѐ5Jw6=q+bZh:<StаC41Gnh[$o-@t ؞39S \ # )rM 0FȂfy/:q\zSsq;^3kq}.*BK9*YC()v䭗^U;^ժ@ Bt,ux؄d1l\8sRN0:u#goēSIGVqc3'A.itF.9s'limKJIkɜVwx4^3n~8h/N cm0`Aq usns| |" :@0\)XGy/͸'H$rkk1+;1i_9#wz M%^N)!QUبe -x.TT`wf䆤cZk+X3!wz `H\PŞb4O0R) K j9-U8"~~ Fwpe__9L40NWRHaA F}#>x9ޱ8NJs4#fm*P6X^f! گD˔B*Pb [*AB#[ :T;A XXii{#4%#t[or": KBp sEzd.BaRX2Ǝb._]/rc 4bCs\͸n|~h& v$ؔҲ||fGdE BT6gv#BҀRn@&Œ8lB `)'4k3#--NM$F^^-C8&C`3&r+W jRImb b~jp|toRp-8Y"=bVB9"cP*@# ;"TP4Kp]R #ZFEs#y9#2yj^!{MZhHRiA "Mk5?υ.$cfُGz-xLE0(D7D(BR8]8")[91&c02m06HMQH"Tgzȑ_r8[I~: 2N{Ȏ%'/?EY[-8qF[fꞌH}Jbƙ`e}H8)2Jn87D-ݝ7](,e̿'=nU1~^ Q~"hy/tXMЩw=LiF:t䨹ا(#q*!.\b(#md2@'=wpm(jIHE&QhD Fc2ʒ#֢c-4mv69vr3c-(J: ߧ+B4dX5t؝sb]FtֲxRQT&QJtK.+mgf۽LTy ̓iͷ#хO- 1̸2L!bRbY?PRj\0WFL't}&JrLm, ?ot)!@* })8U{qEjx984r-Z#d3,MWۮkח_YJ-GUk<N|\ߖl:9f7=kRmvm[Zn`+;g ,3|`[j#a}梻H ϱ Qkw6 5~̚`JHx=5䑠sq&GE`}MX׈<Đz0^MiZ{ k|m*Èl2@BE ȬH)KQ^f>Ɔl%!U3R|"/~s{3>"rS jߙhCb%[R 0JPMqgWɷI.V-,Xq9cmgytˢ["?|pE~'Y9EJBJMD"ļ^>ƪ~HʓdEm< JxMtX >tE<;MZYف>년z}׊#*>!4&Qއz^ME@b; I62gaG|D!g@e:q&lA1ßsbٴL A8š]X>-l1HzF%j2X򵜅i2Xi.G~tI61 5תmUDk*Y)dA9`nHnԏ.qJh#  27lf Ò ey!\]~\&k3 ehfkED[4Px>Jمh1$E2nFuOEC '~Lm6akO(1J?Q8{DxѡvQʭ:lă~Dׇox~K8͛j3gO:Tʉt8TPssm_JwTZ)8a1_M0dc%n}:$N(@Ifȭ"huY_/>]N>X|bǼx ;Nv>)|5jW+n0+a-%ӛ{Sf f^=f,Xd:Zp+Ԩ#N3TjCJEڱtOGYƙlݢ#חPi|uD+-mSUK]_ltDr,4RK*I0!b9̤ pZK-B:&[&jm2FU(+D> fA B^xHC >MȽBT4 ޷V閉Z Y@SGc6D-H(@oU)MT2W-Wdj<:H'TBT'y_Lqt]95EpAGXonV7mFy4 'C,crjb 2(!=cp%&Tdtit7IkON_! %%&9BHYƺn5q i?Ё(8_{[LD1+}8b.V77ITO~,}:JU:^"`!.mG(P40ϡf2&#b%I rL(dA"Z]uFi8m½4$ҤpvjG۝s)K(Ly!I @ƒʖ6u;K0BDU4&Ix#`Nn)wtXcQ U"pZC1ɺ]Ssu>hݾ㧱n#3]&]ں}O KҦ%]a}d˗s2s >T9ڮ_]G`XmO4 ?!@TqG=g~&Ĵ*~7w]0Jxt U-wCD\zW$0K^ZHgk=dR!uҙV&C)c R TS5*KB (дt^$@ӇGy1:3kQUU(hAbcaB`wK}TL!>R*',q\jЂAɻX~tG҉W娔4c+GtS5wTʇY6{v(_+.9fKX@$CV9jRŽv6mv&xE6&:AgJN0H8lҙ{[A Kn#F\BXƈj wE W@"8g V҇pY(#;ćK4es:$J8O&Ep}&0{6mtA␴E_ P|ɴqLChW ȕ /Iz̬Hqql 5&']$/QjKXnz7Wzb>w k8X sb!s?L`wTq-@v"rI|{8jp <BgVJn漪U}XjU֬JX+fGf\քp`-`1|+D5y. hGOeT+Ģy`su.Dwmq$,%օb8O! F]mEw =_:gGWyf#٤cxwI= |Hn<1Oo A݉ڭ &N0or6dD Dy^AkgN6~66/nׯloW'= :7z'׮~(>ڲ a4}=}^iu]\ (iQc͏n~WXS^V|Y.4BА XS-Tgo5{ٙ:5WLI׍zؖU/EmZw>EWNcHΊ$9'aSASܨ⪫/fÊV pģ^ോKNŇShJ.GVo>铺rL}6|&?bGW_q+iEVed)UaZ wzCͧ0]?v@Gm8pltVLw0VNy&0Ez.zraL#u?~͛iB|εtʞxiȃvs0E! aBC r}0oCs4Z}yT)֐ ó KЏR%d nb)R$ M7+/2< ܩJjR=|R sg=7Yayޜl-ڹ LL!5Y<poe?%yJzߵ,*}.tm&IvpflY36҃2cۺP\seO3ZcNΣS:l0ϯߞ3[~7nq-gfO4MF2_b4\ǽ:Jqw)#W.oG?HyY}ywqǜlW/1crcTjL*0?cZbLM3;VGih}ߝ1ց8G['=jBS #Sx/`wu(*ccоOoZ]ѯ[4T LE6<^Wg<GB㞣!v]qSx=8mnS 5nrv=TQ,M.vo5A"9ZWJ? ź8u9>#a7yG/nYOKp70‹ӷgkʧmo`S0_)WS0/P>}?L5zmFRp5m˜XM)џԳbC3POX{^,gC.<} dt8{޽.#n?;> Kgk[o3k tSYr+ش@#cAjrjh=rg"?\=}ק< YBDօFt`@9,o&ቍMrl_8k ѧeit9MwY{{?SNY=r.V;ᔜ}l$vKnQs;1lTSnqhr;7rc /m!E&ڌVG!hT&v]޽ u t'ß˭]_qVQRrf7-4N;e_l+# k{wWJ %,$W JڕjudS3Aw+[gK)/O R+ tF $E!,gl:‘*t5ݗ^p6gNϭZBqĀL5*:pflHIC*'_wۼIyr"z 1afoC݁~*#+6A!I cYtN߯tDR=#{*)pN5z}0L1.okauGyq0zWQ{`ڭ;阓5$.ƅu+[Nޗph/oEMYwF1Ñ\?Ǹ| 7Uw4݋7P=ln_)Kd(~Yre`a6M.h-R\tm `%s f_ޓd3GV8}q|U+AT`vJqȚp; V4PUrP6RYUW{p*NX҆/Gk)QIY@dfJTPhg_?ݚ^密J$V..Q;uxP~ YjŃSu,@9e1f,~;GcŪ9m\޴LJ)/1~ZK>wRJZ524dz7{|"k KTPl$۩p] 6ߘw[K>TS\J6Un,3?,"lC/D68+pq+%83༃a3j_W ZrWMiiI\O,JPr Ap$e#w{bv &HlҬOE_ R )U(`i` l9䲯AEԯ[B]EۋZgpZ!wq@;*9v|Ӏ`%b 9@D{ɺ`&9Y4rFݥc}!vSUQAPqM\ش긴!+Mmݺf)F ]=r] Dis(# 5Gk@*z*hmƊ+TRRu)jn \O@(U'q/cfΙtv^Grec} -װf.*fU*cf' J%R@ڗn:ւH$AB&ԫ7Y {͝SZrR'r;aʖ;o-y{DҖw}C,Ao> Hs IE7$Rwr| [l+ҴhSm-ؘB I Bu3kxa&{O9ȿEOwV`y&ߥU>uNVMַ?$\eroIf!gcf֕y43 A<\*Yctٴk_yΌ1 \ 7R#z[xyK1)֘p7(yˤ ܭ$zE#9K4&l=s(jh{#,F-&h_x{@ѳD\jl>bflEx6e?}_׼N ^\p N{^˵2^*Q SR.4oEHd g1uI%tZ})_ t09XVu']\,gjsy8#@J`rbܕ.EBs;UŻ19Ыyd-ݘLܻ2 QqeNYfL3<+yOho듞~B40|7f'h]@Hs噱ֳ|~[-l/=ٔh&ECf;AN]Q,7k]_R}RVOjU7νfTX>Q\0US5.0RL) Ъ [kGxn4╳ܔ7.QM7J5QэRnjې8d棾1qm=rn`}W_fMvl-$oVW.Z@UVTִIHeLIj{cIAR*./hh7Z?Qhu~ٮC> 5}ut\靠@6 gSc냓p!f.`\[WTnNUTɫZ6?OUjl =Q^ꖪL՚y&9HN^2PH(rŨ?phSNȜZŴ g#>w߻u,ۦRmȅGJs=rWQQIŞ)l٩b,>]aM ! pX vBp@Tz)q&j3j1+4Fb$)&lMcS5 ;UV9qMa o5wҬog`'5 @<g4TT4;v+ E'!Q_s?I"*82 DMN6PwpFxQ1&U[LjI% @-1{c-*1t5k׊M\UTwq_!5iU8N/gII\JO.Egg.Es t=U]Um#ggTi;I,Ұ&L1JP Buv!k%T$crӁ(Q*+'80l38`$B% CZÕT!0`Y U9Kbɚ41֤\TY3$0-3ZN6w8Fs1# F1ZT-:Y*86iTILR.ш!̕ma!n6EM5C͖=/=yL~)d`R$$@&fU3jfҨdYB0ҢIYeEdfLNXDhM2ւܗ@^GiL푴\sM1T-՘#i$mȾ6ۏ3oyr?]iW^8 onU.'VtvTXB˒G k}/#pNw#'w;F98 `vE`FAslu;#u.7_ɆM˶r4IعauzZʁmҺn^Sq` PN0έU~h߹o+@7nfvHV4QZ"{W4\r+׎tsW^'P@έl\:Wn|!s.2}42oZ9`Cn4-ꡙsO(iG7{$3{ΎK6,@ln qtŗ`vOڍ7}Jnn cvfhw!SŵcP8i_3ݐ}Qd5x1R]_HmRnhQWR)TNog:ͦv@QE2s1e9rȓO(uF ڑY_;F<%h# <&onv>KN]-gIk p)r,Yo>~jA2~,e;mhLkyZu?bywr1q!1g[#.wR?;嗋p(r?ѝ ,r֒>rz]cY°",/NYZtGxo?ud, M?eY׌+cDO_ƫ9R:;/ﰊID w{Y~>Ϥi>V뽑v9; i9ZWw%Ƒyo{)R9b]o>1Sݕ`r~\?] -AJ6k"&-EUdLȣD 4xʌ<)6:xOw2N{)qKpc(fnoU!vS$KG=%x݊y;nX*J\ZWScMR=MPl6Kr-Rd%GZggIJV|'tNS\I㏄[7ju/$(AorۋHo1hKܶFqӃލ_MWީ!P=1lÍ8*0CmiajsWWuFjOl5x侔lNuUV˭yeڿX./Pe@ϝ'U] *STdEih[DKp:{} jP_')^2ZÑT/-<sFN_CdH*Iḱxj ڙs0_8*>LѻHiڢ=gsٰ}]|R>t&!J"se٠[,'/u:i3&x0hYjP}aV8]Q"{ ʓ ֹi6v$V$5!aN;mX.)|K_o_f-BPQ\{*,Y-(y"!b*\!d՞q> y2dBU5.INq'µ|&gr @_;PA#(,%Nk]Nf :XFS 3:[˾$یqdJ~P>nWg\l8ës7rWZgˌ$2J /!3Y:"at\,44y8K=5ЋFԒ EG1YUQ^JDT݇40e&*WJF S!&ܒSTC/7jŕZY)3"Be*11Dv=Lr 8&d)zY $Ky$׬k.if~Z}pEҫR6 +^ҵE '!UᜍBaXj2B$}%۬-Bږ,qUNו]* z {_%JT 8NN!:+{ڲVHHY;Q[R4S\g8d^9“ZҧtR:^t"BhlN/>[2W)Q=͉|U5UL(_r!'E!A])vGM^ Tds RKf0r؉RBɌ\ȟr){h_ou挎iüm:{[27ϟ)w{Ofgϔ<ҊO|̳ˎYoEz{пi1 K4祾_f$zS9J=]w.m&>'7$,+F~ ˫[ek(ѭnk|I7ayMgZ9!"Y]ю/U[>dKooX7ڧ֏K i8-NYV ۺGhVUHk}EƁgXD&+=2tDgA83 G?Wv<٥֨ RR봷, dD]rd 8i z~LEݎ ,[n"S! =G Y&&)-8rBq7Yt)C*A6\β*V;s"3hu=Q;,9BP oǎn;f9H|03 H7AcYgͯf}|5+6Jy&5OfGZ><[|$_W-^4Q9oNG ]+ >R!esAkN;Hr>0uB<)K1jkDDۿQݝ=)|ij26[͒S!霳,CR3!!0Aw"M"k!"rovDwV/{!ן.* rN)ps]^p$lh0s]MLd溄T0޾)&v+)_{,#~zh(dTZ}h)bUhMViT֞,\ je-X"ΕB4_Ku݃_ R"-o(_Oƴj{kBlft!I|uUfW %c`EBz(.qPR&SpR,)T6JR5J`iw%L*/\}:b:R^YDƖPɎ!:0c JI͡QSLu2R1ZFFM2UkTDyVEz"YE2D}3døGE Ti`vVvQ3ق  l<Oǣq NTyt*.QZ*wɢ љѩTϺӼ7uҠr/haގ!]+I=IƴoImZwY7*cAEiպM}sޑJ2MlԵ{^$):wSޚbχ(4O_'1S;c̥gKW-pRj%82P +dQjV1&wuKIEU2C?]T>2>rL_KcHZ뵎f؃{m9c|@Ǘ;E"/S=nVɡv]mo9r+~ ߪH/ C{,i%3#5#Ϩg=nubŪv4\V4ۈv@ p4hoQ` V ==-@_#5t5f{ic Uq?=! )1 u"^1k]U0~acL6R/.l"ck ֮KY MZb v#XrQ J°,xU,ͣ;;Гlkk Gj;&( j[DMr>;S+k;h}T]8R!.26Q~ƽuYnm]i-fqZkX3P9b[Q"c!26:fUvcJްBB!nY$)MZnĸL5P1hM*]5xxx{ҿ5"HVYܚ3jT^4XGE@y޷^?} 2|vuR.mȵcϛ*qw2W }}yB7sӲ7˼ ~Ny_@?1Z8gf3x7~~=wcBnrnv!/\ES>LM~O$¯遒S]8p-yho+*oG+Y{Ϳ|2><*%5+3IiMH]3t~x~;eDg8s'ȏ_;CޙpG=-fFCgK_m%ů~t|iiU^Ω[&׀ð<47}6(CvOFc8j.(AV̜ho}6Aoʘ Bڃ!Zᒚ3},c.-PbOb8qf TܡlO5rGa olM֡/nNӒ6%tq Hžнc!Y,ir%uZD2_}Վ]$GNKqMя_Чܲ^7dot5#h[uT, jrΈTq*퀔Xhe#jӿV`CyzDF6F>&jNT XPQwcVRѡM?Qщ5}Z`!ORo×>idfkHEUHQ )L&:P]PWvaA]f7Yy?^Wy%f#<f/On~{}ʤ`pu|oJGj/D VS5`k)k-LgHm9usIuS௭7?+a{?# 7c_4 ArCKOx6K1e';^٩7ǽG x3W}6.>7a,n<( !^k>ŋtxQWGӦSAvJ62!f!i5Zᒚ[+yavBR3/)- +?oji/ :[=;N<1꽖jk-;s7h¢ +,›UV`,^iU},_y[#kIdq*cHӦkȕ8 ؽQ26^}H|DHRl|`V-(3n~u4?vo\::RP+w(}0GǓ`YB.Ϗ p\K*o.L&ܪJ$P[MqeTMUQ`R*թR2e1(R@E@Zh! T)1J? )zJS0[ՒT1W'kl;K1D1Rd C]\lDvv^|h _;R~}ǧKB >C:ѰhQX¥[&JݬdG"W;C@]+kloH9Ǚ48Uؓ )Be1OUu0P#)n*X|Rm-ZDCvbN!6wZCNMW5y*/ݲlSV^Lj_)sENHR]^F%ZvJt #b+o. m8PC ,(c-7!)-!6k dhýAN ]gR֞2 qAXQ`? HrgN2x)q REuE;o@IbuGʮƨ}Qa!h,9?zmȹ4%/\=>aic'># @ 숉wu/6(m(DV!VŨƖ31hnΰiԈOp6`QJ.bq;*c($~d?//DH]H8I$@eQ8ԶZf$9wRRb*@9J\ 9GgpsYc4h q);M ʘnHb $uJՉs̺]0͆ɘvw =ck Ɉx\z=v> _[6ܥv"tQ:1dϞHa`.YUȧUځ1(c}j^f`֕!/%Ybҵ 9 [I6l-_FN V]N1+Ve-SY@>V1-E{4.V1n(O) Ed_ů-o7}HQ1:L&"J/T3%6OS e4+e.8yF &]Wp }~Y1Di '%+r %p#z`@lxh-| g=%fcRթ :B KMid+ٙiȁeYǒI:;U,B5};$Ĺg A 7%wl΋R6-F>` J]>9b5q18dvlC5eN]GbGNU0)x~<9l /D|`z26|T'|ɻw"emhɎ@E20LRŘj0Qzv}?E ڵP欫ۻm3f8yAzvsVt ]iDJrA=YK(}.]ltX'1z '*&~`\uFDpI%6^ W׷<{Jr@"镅>lc7x&.W\D#DS;SJHJfu xְ$y暢 {^C[㝂b7<J?mˁ5jjexTr7ʰ|ABR zQxgRЍgF3Jl/Tp 9 "á(]!fŇm>tjk[?WDz:+l% 6.Z1vkw7-ZlWQ_j{-a c޺ y;=<+fw?ur~pypݕ 9G^n-OuŨ8{T 3xG5y ~{}~|6xo@N i]H dzgp3fG54HRb38=DgҚDǥڭ 9!Og&=~p5ܡIy=wpZt5-w3K\3[`fXB`.iIgxzТțvY5SئǓX ~=lo(7FO o odfTd PS$=4tvbx/me|e7\߼_Us}us{Zu~I̵գ1auD\_qqA=q0# UY[kjgBUgqHmå-} %.ɇ;S-IJ`>lj-=ꦵ'vw:HvvuR %=Q?{W8J_ TLM~D)ĝddz藴ɎlىfWWwb ZV K&΁8FkL5_2Xӱ8sqBu۳}kZbA2"nqxWcؾ iqEK45nL\ hm`ۖHTv~A5i4_C&vG]{,|$T,x.MZ,C5[VWjp7yԨj嵿z߼;y3 k# fM37B^#89>i5b2UTh٦5+c;]" `U[@6hmw# xlsejU:ѱ nr=U1jpNGͩ09ǒsM]טmkD! 05BZvZ~>DaVWfeONTI=+D<؄"GOŜP.|{*Utv GA<݁C;81A;krxLd;>h:>w*Y ڹ"l)Ab!8j`Ah$;"h|s'[C.":&'6jm^t52=T6&ݚBƦn|)L6CtL  XmL0]!WPL;Ӧ  ,|OQ)fx1}auGo>XK-k7)Ml׿A.׆=8 < ˢǛS@aK ̟/z0s\iƃUʆd]¼0!^۪ryP*TK/1F5,7]:wgT0DH%+dH`XE~Mc*+XG{ҳ=S5+4DY<E U.(/Ce+]$=_%˖{˭1 $<-{iav@-.yZIk-lZNX u$OKPy< 55jҰXt)s_Bd);#<{5,QBƿ*"|pCAbM:J)_Ǚr֊l}8spZBnɠX'VH %*uzbRSsiZ%=)&߈WUKm1g\0v{xv9(5!%30ɬChk9YdO{r;ӢZt?",s,= xU侸;yv<AQ a"JV3 _%=yȤ >`Tbi)Dg\z)crG CqVװv6Xɧ.\ؠl*aS xF)t+twYTؼ*Yyh3Ș艼eKޱ=hs#.T{}:: ԛwoJ'ukoYt`hwL:̗8e4%I)Xwturji5kߥ?=vܻ#Ҁ ~հޏ0\lDsMb*B2pKŹlɽ6ssr(+A%lPf" "z>r jr˳7:^C(- HCF6|;ʍ{;%Jv|2Xv]ސݟct2:ҡZV5 $=&W[RڌWԣ/jW5hj;JQZ1w}?Ax`j31Rݎ(=\o2*]v`wwV%z sU U^mIZ+| (g"|=_9!)% ^A`U+o}eJ]AXWKI0Nb^+-hg"5\PzVPU9V:/Z\D%( =hk%kҶmm-.y(l"sl/MȔ,K^OV]tEP .VAՍLRh_=C%IB<Z[%so.|U?3:`{&X cZj>)/Vϟq~\M|_.Ge\_羸fL󯞳F,W(t?O s4*ޏfH_*֤|4}΁w5g JW})Y^E20" Qksu\==hhuTbr"Ya_ )Z@v?UϤ e笓J+Y RKc*KmbdFT֨ 9zA0X.TQF:.W"lRYGWL[@e/dIEHK{;VU({KoZihS/xVOr 5YzɢK-UńcĦ}FKmAz{uQKPS;Vבg1J}Nlr{^d$c-[[G-"Re#QQUYŖFJ7kH=@%ŭ` g7ll]D Lh HXdCZLţ_՝U}l')$'!1f {L ܩdX|B˚ a?}yS|p?9tt?S}X]y;Sz'tta uAYo[F,H ̦O8g' 'Wݿs;'O3/J`clsZ!_M)?rAƖgB0='Xiɭ@|bb{:<8Y :rF:M/QHŗT rG^*W3rp^*JæccCsDF()#s6Av\"cfA 2Ox._Jڭ!:y]>Ls.oBD'z /1;% a^4˓"Rk0ʻo~|6gs'}%Da 6ڑNhvhԖd-*؂eIXG &,*@kcYf1GeY4R<$MXv"ҸM'a(e:a)^ߣżZ6,7C϶޿&OߗxMo?Ktԃ>&> ewcsg,m2o :`>+i11j"%xX)ERRydkMb*l:Dh  dm7̎ic,_篣1^ST_qeqIgLW!x"`,BYdOP궦h? _֐ِfz~JQ9tl) _mپ nߓ^{}}rP ?-Y0R/>߸Qq-'7'k+m2?-rP|URMp%kOןsv< YWҶ/_ڃՏQ#[~7R2 C 7@Zn\yxU oWK sr[2l9!TY*{'Y<8;䌝/3'ĉ_Y!{ۣy{!∃7Yۘ7SApʟ=i4F,h-(g,QcxN+< ,-о:5 =%TRb)r =Jޭx:M<]Ot\5-¤(`](RXSaBF!ѕ[SF"93h`7ܩ}sAj"o9 `8Zl8~V@c s720{< /bCU鸍ys9̸$VJG㽥i[E߹4m5%Z[֧@qblmD۟Wg?O_^?^ .mIoߏ*}{8vY½6h2dS]9CGoZ9~wc0[5)w' zp7)AB@K"%O ƗΎɓ>G衲-G)x6\m$q:1vsh[{y`eT}>yv?W\>hjon`ZfUyl;]O˙㩔<|_p$B {bn%wBvZgI.W(ъruDf<wE[=Xʳ I@2 "7B clÚl2W!8?gHB<hח^<:H L>q!6VNßM̳Xf\]Qև4Աw+?}KZNBwb@U?+UdMysA]O77BNo[߹Ac'nnHΔ r :$$Gj;`\y'jGz{ճw+>ލ"5`}'u2^2B}?K)siI]ס:j !(g,-0~0qk-nDK-ђwJlF-avJ{ӀYp:g:|DX:0AwҘx48!X/k]|]Ԁ,'mʚKZ}{jV=g,_/x묬< &*cSrKLN>&rĩ6GD@%$EcV:P'V:֬l(1X_ A)wQa)4axPKxlPFFJ"Z 6*Dt!J3&g&ipPNN؎r6HE O4+Cvqzox/٨r$gU<&Kq1 |%)^h! hXJ"D 6)$[̐(DyLJRhF62pOE+VO=mEFjCEh"+{FB !,R$ND:m"ɡAx Q&uayd\)wAQ0$ԏDAzX ,2R#hEȓ!\xQ!0 WD-nOq6AʼQJfU* EMkﮢAq@X9;*@j{bl@v øHS/;DCCBIcPOVq`G׻.yE'z+>]a'b'pIi[gu qm,o[- O]7K\4dAI-(&][7 jf9WLQidGӝrE(lXKEui Cȝ6BРJĜڈ`Gܠ6y)&,C&# uP"#*0D ǂI2Rg4R/P]Lg㾗}}%;ځ812d0\]h"⊁6RtΘ&F)(!R[~؉ûEZ⡮Pv!+$=8(Mb,Q$%lDfSJ Oz<` xfW3  yww櫠.vj'VLS9%|Wd `|Llhr?+N[T 6h m\A+Hvre cbuɞF|t(Sl 𧽭!1C׏GlJH2zd$_/JصЯ׮֐Nzhvy.F7C1|~hޒҡaݐZG|9r; nlJ; ˃;l<öK~9.EbԩZr"Hx92c'-4k5x*g9ct'-lk`5K՟.QbxZWpEN>,DNKj ml66;Gve?q%_0;ſƄcwŎ\ eThrg0f =6HVzAҋfv5lɪA{{706 Մq*͢p#Bxcshd:çOɰh%otVk |k=)\D ץ>yFBzXGJKiipc6}/@׏WB1qNG¬NTqV<жZ:7!s1rhiqJX*^D &sAV ={7rGy4G2/0W) 3;Ae6f\; TT Gp/2HOmbNL:^ĩMp1av]w=۰PCWA.fÎK00Z ̪S=[ uPA=< PgE.t=qHr!ۦAg^Fl ǫ'`T(YVؕI5˞T ٯY2YYOނ'68*^0X1q9nwT qK\ ̟]oY>|pGEhFTx= &\n<[I4|3,«v >7>ջ7,z|kîPd&7@ B)">iPBZMVE;5#t KԀ(vjGP5EJ"Uɴj b2lb[15H uF}gHرNg(ˡBQb,ޱ&+4P0sb5߽zmU,( r:׎Y$sܷ6?p-.ͥGC1ߵgUB7S 0Yvݸ=eP)c_~I0y)C.>@^ g˳tլCJ+(?%rFNhȕyz R[.=1nws+mD󎿣lUɶc%$=[T%2;e#r-h{9$Y#V%Ri;=%*;bFR}AN{!$2[9u/o['[j^@ JhK <\!Sum E#|Z&;9)*.+YT6&oB`-5&3krq$rYkLgl#(El~T"b2x.M+,܌F:&AuAY/!,Pc3h_jU(@M9"&72*iR+: ÈJs%cTEZIGyУ0jc&e(=)Sߋ٦V*0˨Xk\m|]/VZL.SfJ:e&B 웍8򗗿vcWZ0\%<@¨gCG&MŴ`/>51YޝGCor黷%cvw᧔j9pg6o{rw`K\W&#+9 _I!I[8ち=]]@`Nyhӂ-*"H'tPa<"T ӈ>hh@cc |Ogルind${ɓ~Z0X^@^N_ý?̌z> _Mf +L̅/p՝Wp&4,?? ga}1.^r>A۹loK py#@Xn+ύ{9}DheZ oRzw6qx0w/?Oס7sM%5Hq]zȢl_ӈkmT,XJۻb2Η)8֗^,=pv;oL*c0Gў&et7.c [>svo_08ŋ?f{P0)m|1}~95'?!l6,&דbXK̥ Ng<5SUs%x3] S4.hoR|g'W Nv2^vfIp%o/'P>0CF% zr';c>+{=Xiei4M0-m>wO>Wɝ}ӑR=b' k0&f }3$ '0+LuG8KBg9I1yk:/&I>ޟ˿϶a}&] z{~,a%&n[0 @޿U Ƭa Vl8Yv(vA4Wj՛T(2|t&Qԛ%i,NrzŠG} IJq@@Eҧ<|AfnA:Ar}]>wxESAaD! A:ʥa )2 Sܹ!{ γ%mK)D&M u"_ ј^eO{]rcOs '|߼x|הj%|)N`z Tv e"NeѬwoA7IW/h_'Chf=rNbD!JPXw_Xמptq3W0. $O( y+d*' IBE#O:,. 0<>! Yd&T. /jy2G,KH;[*UgKvdgK%?~:[%;[-џ9ǖ^Fusq]I*RfK*TgKZڒqYے6l'Q\-I8UOj紴%;[lI7-ْ-yvVbSya\7PP0cMjWP@GfJE,۝$mKH$ OgCv6dQRk1lΆl?^RH@::0ǫ_9.`P!J1ؖf3uC-d-:K()SJd6v-YџYY([>nf:ϐ܈#$B=0b(2J`qؒ)&rˣ^M*rT1=2824xVg[1h\ǖM:C $=Lb[41ۈ|U]c㹙@{+sqʐLR9lfg4sOZxsƈՍkU#NX{ PIndviCS¹{!IOLHJ9W\;[N #B}j"\i*h[b5hUִϕFdvi#Ry'))ݺWF-r1G'L5]/8gXj4{_/?&M/棡772WX+])/B vHbNq*RS"$0B50. <>܀s7y.{$6H8H>/{ 1pwK$a[܄?{Oq| rs[}cF1`;kic~O5&WRM6^N%-// &Eӵq/r1mZ.b$K% e jyNS)4҅[_]+oNU7 "5mOYBCTmux;*I3e5UjͼPX́hR,1BvqL O "1Hhb(~ uCM)*D־ ָ=ɯw Q>U2y0 1FLnF@`8,(얎[=zBeUx)4y, y>Xڔct:9F&$H"fNa&tj9r$IIM&8He2  ,L}hU<_N7vsQۻU4%x?/^Ѹ'Xzq3-4&L&*<'pAu:HӦMutSqQsf ':DI1)#BPd*L3f-73L|j44<@*Jg$>F:%͉mzGkRn=E֚wYV/VZq~\wtt J7S#wj*'YoP|ϣI|? |~ &ǧb5.k g>Y -d!-,pҰ;"2p6,w?ৃt2܇cªqLɸ T3FqѐQǀE)rV›kr !y{d{l4EKvSx4YM-']%y>6&; W4q&cq|D]A!xZj;X~ߋj{7=zxfV찲 Q8EsL|" ci;3mT\ڄڄ~59+^Q23kp(1JuQs* Χ!uq0~; >pOd3AF|YFHTMtǹ3!y͉SI**|p+KE@;D;*鏏 'Ǒ^c! Qk7@2 a1Q B.gD"SᶊhHlsYGZ!,Ra! a(͐ў;bpUȾB !+br )Ly˰U07Rh,*m^ aj}4KبI7cfw7o޳aǣ1;fxS㒢UYSzAe*X9×OalJ0Ə =N۴QպmEblPY-a6L#JA{(ko4%&)dMlPWAE'RJP``TPl4Z&s$Fީ0-T(lBE[8]Kڕ^Xr2> dQ 4AOsr+]3EǮcCs4\Lހ҇;1 I =m?pƓ1f egZ @կ.)>STk0wd(ڽ\7XM3JP0tPW3"ЛrJgǯ9'mٚqk` Dt)-sӊex$2cN?.?ߔ ߞċQFJā2!-X&eE,+]XQ&D8``Je@1q"$ѪݜJ%)UEhEƒ5_"$Q8kvYxHT.Fpo StϘ]]]-ߪ\E>zHT1}vǣ$-Ԍ7x*yc 2.7pO +1 wUw-1&5D7=;h]{Yyw?RE2L!DҝP᜸Qzs0|cƍ&:MlT0\^Ib[5h{qh3KsD! .:%Uu\0TڤS-]uk@xA;*$(UL?iЊq%^jm*˱d3RLU,no5TjtX.C8'衾ڄZ(KLBrp;ӿӹEtQw^2?7Ë|YJ3Baq'G;MG$m̨]Bv킱6rc5_z:N/lF"77Mf08^X9;$!_֖)B6TA&(hlJ\!F:u*_mOQ9MnZzM:[lq4hQRwSTOrZ2͚tB8)Tf5A B wZ6!YHJTٯ[|5Ê!ע}xQ=»n2DoGGJ0-z 38VmE%mJ37WK&Ou;ӻ h2ޫt6[ۘ4K0ڻ>mi!5l?@&ejQX+ F&)IMljv&XJLfyjeYv4 1*rc8X?n2<(~xW[zϊi^c!Snh¬ Yx4b4ey.- l9\8%PŬW2ŎeQƲe_ԘncQZ]5Gmw۾6|j@ן:teS^)Ԟ "ur[Gß{dz!ʖڢ|M9c䒷U1y{ў歜.:eֈ;ȗ`0hݥc|x;wJcN;VSļ%5ѷqύYa[ t+DQU#H(Va ͑0K(fRnoKe9xpr3 J 0N'Ri!XngaB6xq@ƒ5WI 1 ()qFae2%UuZq{M88(J֗)Qqx}ңF(V#J7 %>LIXHV^c~Y!==byA|&KfDBGCĶT Ozc?7jwĜ*e`FbeFa2D޸QC N;n$i٩yl#~q$F c4ioE4l?OV_sҰ!`X'g iDPWG?} QwGtfjSTkdzV?_zFAa~<!ƤSUDQ$-B؍FYnpfЇx(T9٭lWmDlTy'\O79`0ۣMv4tuP)f:;WFl"Zs8?)rbꝖ 4M.*2;YLߏX4 += `=++req}N(vˊZ+8~:2TqO^Npek滃0w8r6E({mSz)uKY'&6P#`3u@ǠqO~u9 P}Jn1Pm= Z/iQ12 ƈ' 41JZ9ASoFK2L_/)f uT+),5&9aS FI H#8N3[ X*eΔwʈu$޺rF?^8$q Af,Ce u4ZrVDUX:Lx1$ 2|Erq }T"f;'uKr4QiX$ 0^pGr6tusϣ8HtJ>Lh8HoU)??<[0T(_6Æދm:l+a[uq[)Vz'ۤq6齷`EEt U^$$M9@?usy$" #A,n;381>JA"|d֨32MG{V$pac&-*nb"a;?Fy;䲘Q&~]Yݙzb}wgG=+~ϻ-)nVO̟`TIQjML66㝼|w.nпF$ @zwT$ ҋ)4i,_3XrC7ytɷU4UgaMh(s/ LkŸZ1JC5?PʗOBpp)&!b69b {r*\)0H( yi%Q.3߬۵x!LxSˑ,% 4VgfTjqӔh3PHe Ln0#w9Tz2-h1ynp3k5[Xg3 ^X$Han1:5 ҧ^6 rb/h= +knH}:yDnᙇN2-$k}<@7Qnmeudfeefe}(G ZM__!x>]5~jYM1:e-|~Q7poگ=<հrqo<^1D~ݜӿwxwYm6'=~o;\_HBpSݧwn4(>h#^NlenV=R6|"Z%SS8wn4(>hczDŴ[DKLB+-d$n_bTf%P#8qLS@cylb8WC@q䫉1pf(DO`sHhR$e2 ҶAb4l1^~Ӈ:bG%FE+FԬ#&"ݼW'\5Ynk[SicH(PΟB߭cdZX5#:F/Rp&4 g-_QHؾ"|%L+¸*"B+"a+vNVR)={mì`@ re\2_,tvsjiظ%5Gaٳ|-@{+; C۫\7a5ooӣ}3Zvmk5ԗPzXҩr,_~}߬a c)DQy2,yLsb c2eIt|(n(qzjcЛP}~Dk|?~sw8rЊnM?' ` (m+  3툔r#uT71^6hhD >*f[Aw0s jP#N2/r o䍂Oj栘^ @2Δd=1OM9EZ$"JH&'[G %fAmc˟rۧY}{4oS@mutyΒt.էۏnro3q}Hi͟f(d"r3)yHM%+5>J sl - xaSeW1M|͵X>O0'DB!Fc]χ=c㖆E F])!j2ӾJio[x .Q:9 er:IF3~At5bp|1%ʔCYq?мA7N |L6u7gWaUB8 n>H.uGETH;D۲D&TF: Œ%Ё2܏Y2Amtf 7ߏ^Y4}ܓnaˢH.Gryob{Ҍ̓󑶥X9]WՍPD~k$>ɔ%~yO H3 qZBQ#NI)]|,T's=4%&`d5>:A QΜ<̄jOP31%f/V VyTYxTCvFT}2/5dn6Kmx S[s{p)7cX^K\A#oqb˥ØHj}pʇ<a:S o${I9K'.)8gDдF_眾zjxsmсQr ;&kyjqAhTܶ-ݶۋGp۴|NۭQO Q- Jj1 !BBp=X ՂL<5"smʃ?j,,-rvjE[;/m" 8fR dZVQ:P/~7ݢaB3ǥnhEzkm9DSqe32Q)s؇NG b+b:J9: M$  Y`TrB457^&'Π*[RGCӁxOvNt Ps5;ڈDc3M$Q*Z*_"K鹦AxBzҌIYi0 =oK*\qUՊfDu25ᣖ\<72dN_+q$礃jSEg%cO< O^m#jiP1&p-K̙ ^Xa@a EtKxL!BJ}c&L 8E|o?ɠҩ@ +5C~#qjF`#anH  Hlŕ"'m5Ɋ1TQIZ\63T>##mԂf ŀ%Gږ H t#:v> |`vܢ\ɡ1~mKk&*aZ6x0I׍90foGk-- x2Ip6%ͫb9YNkuM͓PGTOd",:PwWLP'xƻ` >$8O5;(ˠH͔̚!wsk%;Ūx2k ڭcF6gyiQ3=ozS1?.(==SڨO*W옮n80ϒ)&]7dеjC z34Ѻ~bmZs|Qr UK8啳kYKM(͙^DEɐkL2\oZK ؁)魷1-d&'DObU#H`n:Kb8oEE+託B,Ǵ2ݔ'>-7;gȾ+fAXpm(Ѧ;𮟭%7Cr]rLǼ)b0g?fg#zJ8Iϕ)[bheUHu/6Y33]<ή71?#|#Vo?c9f˺at_y=ZĞxrfi /^l{M>(fr2=Ǹ!Ƿ_mHENg!d:1y&gpdg`xAA/mc'n?tbN/-ž :y%1^4TjZ<|V1:\u (uyr4dZiĨJySQ'#NzS&Ǡ4+"H,nѝgr|~Pٹ?T\FQ>-XZJ*Wwj6/{CF+#yŖaQ$z1$Ѡ{0q_V#!R]iԠ $#md+ƀq|D\K&]d/ ӓY>6CwJvlewgpY?޾ י\5.r,/Ezj%-072ކ/np{saʼwga{yv']|j~z0!!]ݡ{͛vyK>6*e.?<<<Ku;,HV 83rPX8.~ x/Yϳg> ]z8!ڐXpQhB Et3+X&ޡwB8jA=|IᘅUS-+)|w!fS$SM6$b !mkcP*7QH[dIF<}R ;{ **BL/$Yq)ta5PEb\ Gtv>2X+luf#)j+߃$@ DOE$ڣ)ȇFLX@KuD[A8h%ƭ;hd$9!#:W6.7_GZRUJ|tR{<N3#SDQH\p8 \k6IU@y ұ2#^~nԜYCsG(QI+1s& &K-mɨ2 L`YEX;t\d y4V§iMogoJC/ QTЋΡ|$ƆsӦYL6.&R:ߎyEFE}Bڠ0vS>0 1GJSK(YI49:1QJ*eVyN5#2f5)Q|W=m{(P(z9جVYEPc?PF!y3Q-X_1=NCž6swzL &:ՙR1_^Wg:#\ͯ΄`B*v|[QЙi1=3h|Hh =>11 rvMC(%=oGr+|9X 0%18%ALgg=:WRLćI#U*Ū"A0N8y|o_ŭ[oܷ Ś,a_b"eBvq^R=j^0?]m5[9ړjTK2޽u"%JDx)B kdg&9z^Akf.`zNT+AT@&,Jl]Q_c`zIEYoޔ)ƹ5+X [j%- ˝)9Y%.oZ>OPm`e Z+( JF{1݈!,Qb&oyN퓳k/}(X"DN 7%F>ð3w="! P%Dr.*~ nSSV#Uږ iJsVM???Tp1j}moAs M{k`03jg ~,h:Ӻڭ^o 6w ,m޼w 1f 9Wsf/|Eghi7G! 3?_ ~ܹ~o\Qdd=={\rC/=ެp-_{'%_oOZ_Z~!URV Wrן.d$gTbKTcOWsonE1Df ̢3zey\yyN6(@囅Y\S0ħ}.v@ 5:qGQ2rZ@ײ%)BBjC|0$Ŷv۝5n4ޯrTǞH!TS<PTC}1on]T 8[M )栛vL5.*jBxѪ,,@!ak$)XJK9-iU$Մ8Yc 8&f1' uh'qz4M}ߘ-u@$IIy$2wn=jHڽVjD(ERJsペ;KEZPB-8t)pC$s/q "*)(h4dAtb!H"HB&R> 6J$voálm5mCD=O6n8 ],}Cc4*6~-m9BS0]mvSng~,ڟߴnWͿ .qny=m knrIC>s ҩNҹg #D:=&nNCt۫F) / M~j~wq ҩN\7$YvOhC1Q16 SٴnZhNQTcIHٕV"^I-Aƥ5è ȁ EFXK4#YKa?@5k"VAJX0W[z"a#- Bb*ʊ3S"Humm 6M;D[lVkPoxrvNCl' Oi&w[{\!y {Cǚ4"GXf\kR`&y][UU5hvP @^OJVԄZ:'6Cvy , XVƧpmbs:ԜeqVwE!jg<0Ԭe 2[߬2-f-]X^(Ko 9g5궮rcm?)negKENӕDcS sé}ǡ_:x8I<#:J|2} Ul1G !|ybj|"[sxI$$}l~e ~~luy p>̩B^SՏL:_%Y!JAؿ7# /iѹPg>;V|v],P̕{@=%zo6NA%-TaB޿2i/ZEi ީ:S~,{l˽oF.s$fnICl{s0'eFs?LVp 1$(MY R#N@S$`!?)^ɜny%'Em)Iq*$Lld0-<)Z:*;%T*4}uRNIIѢ44Kr24wt4>Xq~nA <BXYh4* )AtQ㖈@w6EK˧@\fK mɪ"!WF:ŵ#s/䔲ajnJ̭}rD\p>nvbjS"T5J, 5͚>-%)jM1K4f7 D_QPm ^QW \C aq o[ /[vCcl3%ф "wbʼ5S)y &T9!D!aUT%4T:Xj1 )-pe q`}ƃu jep(4W9'k ekLU #Cׂ~~}N5arbsDC.> k 9!҂o@f1&췸]6 1Ne8vv4gMo|ݧ~:5~_p˦7b@ҋ=cru|bi;02uY#> tI<>q r;<R3CE-wRNFt{?Ͳ8J8bl죆Ra]R Ke)5*%Ie9R]?(~u)=.}M5Fґ1\0j[ā0^m3\*c [ZYlbf )A $%&s:2+9++JDAԂbD 8&RjvLڪF:=~i /3{F1/̼K>_a? YbѴKg ެ)ຖ4KiߝOA (iU 똪ʦ1#k*J``)Io0JImO(- "tM~@B Ǣ,oo P+""Zw18 5h. ]kZ*A5=3L9!pE+AuUR2qml +Xd,,"H0@L*XgR:#gxDswx0 6>C}lWShNh0h%Nfc62—3*QR&?>ub?Vq\jkD/њ!GY@ŃHd{=ɾ Ra"TLϮH!مvz iqu_T'dSԋ  I#F_OOgu2"v9Aⵎ=i:ISF@IfI_]l|^~jY v ^^1]0s`w t<3B1 o86,g EJ A+”CS 'RB7%z2k[)g!i4;! fՄA!B$SbdP" S4xwptv)^'#VY^ߨ(AսA[(2K<|)O=٧K)4M7v Nk4h(B*IMv E|e_*'b%U\J i!T_$it!FB( @|;yD"j%y46KBX!ļLo^ a^M's W|(ʑAsiꊰR-jfSĉE,1 ˬ崬)T)\J+!  ?Z1!(s9^8eī᜖Dzvs(];gY OS1T)i'ԓdʳd U4DG";O Lx dz|΀ԺsM\a?_T`QlJ(>36t!ʄSI!iizQ Qz.*143W:%)}~_}"auYwhD81z`9Л旓XȲvuA b1PT N_pvS:=Cdܴ/} ^*] 4,Jl]Qsz$JI烖>tHUٻ6n$LݦV~NwrljʎvS* —9CJQ8gV> +cM_ De7dc[6vЅ4n 95NS9²=~ l[ (dxWCNo,K8j+f`$RPD&-IM߉=R-}|H: CG{W/[8iUJpKGڈ㏵W,R7Kn0ң.Y Aڵ#Ӆf0SH܈n1dTrg].d^7e]-۹V]0\)BծR&"RYn.c^u""5AZF™ҭRJ Z3PDͺ$0;-%f8EFFc-Kt.mS4i mc%bHhE>[ ĈBگ#U&$W/A պPH.b$RbYupbr|'wJVž<q!Iaл虅KՆhY®~!YVoJ$pZF ŰH>uᾬ}WEmD0|qv/ DRh@Җ0({s []7Y/9}rJa(?)];Bn $](GYƅ80-෻_rB 'Tiɘ)aTzbxn{n @⌠²#- ]6tP~qEx`L[׆0JR 7~2l\j/z>+cjƱ\'.燻#L/sK'Fgl=W+7ﴝj Ѝ],TCց7 ],Z͍DÆwK 1'xmQ{IpF)wwxW*.n&ټ?5}ցvO'F!&y9NߓZc:cm8&,9ˉV.Py)0O&x^}~bU`N6FN$A1JzFk +1%yF%p ^>i1W.W+ a s hރ8 /s?fR̃W]cS =|N7<2YSoqЏ. !ؾåG!IKxF ˎIPLS(jj@HyMS2mZTëR@z%R&)OH' TԱm;Pl`邟/H9ۀ=93a5v_j`tD.qDNU:Q¨<21nMSq>nj>cj;Q N)&suNGq:T? Y!7s3''А%XdES"qnȭF!Y8YNRȢJߞu*--q3CņDoIYiɐtoi|FUaU a^վ_Ҿ_:DqyQco(qݜ+\+mbxg#K!]m>H,TgK3'S[,۝G}@@Yq'3Rmnޞ.>aӨmXY5e{@&>۝t-)yz']RQ%*TCȣKO[S. D5Uc8oyݺ)ϧ[A8Cnc0j;D#jg×;hzDKT-U? .$q`RrD'Cz0BY=5ր烞&2i*aǗW\7@Q.;g!TQ$f!3aYb%GH$pҥsd!⤴(guЧA2ZSo>LRAzy1_4Eye 9_l%yyx..sY:+y|>D&pUf+$74 Ni?faL`kɁ`D*ڑ\.䬝iD+zΎF,2n5#?šˏ~nw^˦PQm(z2\_t߫בϧ`{`Ȍwĝym?2MY0λTPoZźo}IdM=Z>M_ y;]n(A-̇,Wxo,T\#p9yNf@'T8KDJNˤ"YHaҌ[$@p{$f/3̯ˁY70Գ}%kx>X= 0 ]Df@% `SK3/31gd62óчOk?K}6Q>:WSW`gAO:1c zɝہ¼ A2'$Ԁ凇ve x"W[ czׁD\s'J0F۴s|֢rw7OɊ/깺KR|pCf>5eSYz;=y .[?Rd.^b}r P&DKR޹ڹznjykN+yh`Yމ p-ڭN&3uTИ9qfA'],e*8*&F8xdǩ'8ũv9XpTSb[gQb‡PӫUK<2P0|=bĥjoCċC m6΢puD0>p0@R"up,ULK0AL.Rk'fN{jCaE)2,NtbNx(R߇RE?*V;-Y_cLH> ,qHm$\ItWP{%TDyb06NPhS4D!yy=AK12(9+V(8)Lk ١ \ڇh[]8R^*z+CBbɯ^. ;svO˻/Sf> @$.l[Z2 aQxRˈ6YG &a~b}~Mll\cV$L0ؠ1Lw*GoݍF>tyua4w%4ޏxp_-]~kCۚԝ*ҙ0<&m'/I  'C _|p6n:0m CE(8йY @NjZ?2ԙX(Dkrl,FfHhFr-031CI$vI1)$aT,K z!Z6 NfZBy+h’Yh2XD3c[CeHӌD mx5%F)&|^(Vy4@"Ѐ`Zg[]sJ]`e \#z#s~8a}~8HxԳBk0{+#޳ B 754"L5TSmmS y3ͩ"+%J3f I(,ĩԘ( r8D YK PP%Q t؛.F.Pe`̱D'28Hmr';3b;aC%XͶN~]e~![tlg<1bbT*HZz~_+<{2A4 3k岼rnGCr ҩW_ʠ٥D|Bt e2!_}t\\ʠꄾϔ~fhUtkCCr )A/>vn`zgDPTQwA3ƣ ݂nmhW!:%8w B.BePEuBQV0hUtkCCr )q{!Ex2:糖ngRX{][0Uѭ U1N0d`6޹M/XZ}t<\[lȏ/_il3)d޿~$\D6Y%)?ڃD+ ;zL Tc^#&c&$M,,#J.U3NUXjNDƄqXhXԍָpS9sJyRaI>D'-0jy8c;*z, ,ЈmGFR_}thjMa*P L\wŤ7N>?^^NX#Ы$o LGkwn3+jSM`]5 gdzfgMru@zd[3w6xzY׳/mV"c,P(PUQ| '~ڗfvL nR%L 09Nr H̐{eFisdĕ&6OZb/Ӌǎ׋ᒽX])݇9 8Cby5/m{ѯdž?+*zQq%)[#@sxS'}NŝN0VhEVh\ DuU/X1xl74neT&ԅ~ؑ^~R Na bp0Yei/. \Ppa0qZq͗&7˱;kYfE 8RiDq*T0|ZH/w3SRFV]y_"A1_QJAߢJ )'r`Lm6l7ν`͹ȧ RڗsN Tqp'2ہ|#cC``v:pѵۤ1>d {EklN#L@C'픎̓>;HoUf㡭VQri<(t5}ʵ3Q?t@?_K+a?Nv[p~ШU mIpYlV-nkٜ&M4Wd]q;ruz1f7?~n:pjsz""ݏ ?>&1X!y? >xku~l@6t`LNj>Xp aZY00Ņ4vdERHR͛S!8@pF-w9L˰lSJک7zs :цi|c[Xp{i7soY~j{/PLL[']4hAj@=(Lj2Q6L9f)I:Y/D@\Ke3H|i}{=c܎Rxvu%ʈ4K]ǾXEuBTD?0tfq7Nk4ҧ(R)Bi6Y{JI$s! ʢ@1Dyq iNOcX_(ާӰ9KЉWIlegiQؐX P76g4kC"'-<&_SҜ/(_2iUј/ wFLPhN%:+!$uΘ)C$&`mwnFb4?yas2s^ VPb,:1P7ԭ̍_ s»MY: )Q$A:]EbEk r'ݬt;Y)&HM/g N_T$㝹 okfqauCcD2rdbTn\< ~bO19E^{4PSqynvMť^uP}VLs՟5I0Ҏ_d.}_J>ۣ\jyW {A'izXWt6TFc0/8[Ѯz,HUz5u(]wmK6@hb ا(ޞiz0xȒ"ɉӢ쒒Hijlp,&綻33ظNXr$}HW<,c@d."c~3=В+x"XxމLXCŽMAtˋ|,s $A '<-yuCv>gW ڙmda0^2XNb5uC pKJ͙aB#X/CqZPMOt+n5rN^7;xy .$Gu+U†,IGgBN)$,m:$#@ zkT~ʙ}_bB `JY}4z%hjsd'MJJ}[^w9p{*=(5",Hy`Ž3 Nn؅$計1yL[<;W6j0"hHihҜBBm3A~&1qq( “YGS6"JDg;\Ʌ_,5:RM %Xz߶q%S[K7EIjFv7Xzѭpg%KmueLVɇR[\dUy >D#E~cfwIEDrC/{uz Þ\=!c8:5ܽTؓwDٕ[5Rq IF&wseړ9??Jڕ񷿒 D9MI0o~Vvd>~ tŹ7w>2A>{ײO-$"ΏIlC >8Gsh{lb" ;fBɦs z!ĈOPb#LFŪlmM)/5zLWݖ*QjPF`m:Ӌ&y.n)8CJA6axkPB_>J@r=(ipα?d+f= e3,Fb)ɚ%۠=K t#[5'M5;Ԥk4Z~]f0fܰ|Zs 7o,R3qtgŽr j!'~K|;D2O'ξPpso l De?ې$48  zsQ1Z`:^)X{fn0z52aپUy~1T: kKᾨq5[c@PG3 :O1f/qӖ~twVW~u0ߗdOӯA={m%S3O8{5ߥb{]Zv_Y=u z*bPQ  \)jÅx3 ._.yM5M (}eXW 8׌IK};|$$hJiC!"G!iBh8|$Mwf|`zq~g`% ɔMX 4t  cIhDKAHLRDC*FÂ9hs +3ORO2_PIG|Ӣ)>1; i-fi 1Tmr!4C7al H=eVIkK/ٽRM8m0K^@3:o72UsR/j-!G\8R4kFmӞJ 8tl.nIB8UA*{fXM{2<سE=.y03*Z`|Əj95\]~!T78q{S n&S6isZp;:ΔaWޢWr\rH mqFh8R2)YW;kem(ASi` 0{md$FOP[ٲǞp 5uOU{9&$jgBBx/6uaMwVΗ ^{gF7+D#3ݠ n{)B* ee=Bpش2 Lϛ[Ƨe1ܜv1g vd͔ #Y?5(:L5L9e3TjLj@MB 5jBMnUj}5<sz^W}};^@'rV g]U8-x<-6|C}r2?tPKǴCڮdC硃!؇ (6)9HF_܆'+.^kh^P6]Սj]|DŽ+EEGVtaHo_~d2(gKsvǹ>.XyA#z$Ʃ>YŧlT3\]}묓srͤ)h@0 II! 2O'UޯW痢=),(Q m*]'xW6[PR ςI/ji/Q 4a@O@s%D3(q)F "{:fjSRGE3ka9JC'.*G c2ocL)#-9,gYL`oa _iCgmgi"ˁ^'VkI|g5YE"ǥdj}\\}whޛ$9'HʇaLO_$2Y䮺Ǧ|jis3^۔(wyw/Zt#nX^?sIc0a~Cn37̒G\~x{qYoAf1Axه8HHd32pv_;F.cwp[߼jNez{2Y! ?s-d%|))㍎c=))9t*>fXIԮ1tƌdJ"& _^d!cF$"v<Y(iǔYo"!vEqalRV2Ibku0 ԗ2ϯ&DTpr(`qǑUA`SZKyp(F^Hσa6ؿ*S Q̂z\l^s뵇^DK7g fka€8#5ձ9*l-\cIH;(Q.i*+ rhTdEaX 0#?{9RJ6JL9M+OXoJ1LB%oӻak)Oseyh`rUET7xOTp[HTpqTp՝{ٽ8!({*ѦGwdd1zɒoc?VS{Q tM,lޝ0"wrճ{4ML۵}m(yd-\RT9 N'^qܹyp@Mh%R0Vyr?g{Acz!rildBCOo ~%PFAV$6+̓ϪQ5^]9Gؐ׬JTåֹFyF#1'_TZw> v!")R6d](m:[0Br`SLo )oo~~Ka1EE5(h_6.n3, UC,;(.➨P>Pw;l6?@}|YPHuTGTRH/ϡ,6-جI@8]PY8ɕ^YM?~\x2CBڻK3W21M'_ iI t^M P-ɑ,6M'yw;:[Xq &-$Dh; 8L= hf86hc fr[W2Us+gj'x J6 ֚* Db1n6:^V*+X cӎpi]ٝ| cuBɖ)k K~{/9~;8(,67<> bljzXmE粬tQ c yL69>G7Yw`q,B t%`4 }s94m)h?g53Pk l/Hoes۹ky#>s$^;HB98;s\(fG9e(B^+9/|pLoZP+0畇|# /Xm ˖\BQE";P\)Ug^_:8X_{߯%y7/Ohuz'KPhh \LÑy!_[?miOic;Nhc؆$h ҄I3N" 2*␈P9,NZ]wy֪uEjC%Cl̏[2e&0.%ъ_fc+2Tdx/2`$$ 41WXxhP 8a2YQ2z!NUVRRDb(b# *1(ǠPK8A,z%98#9; 7^eif+RJqlN4y)HuMS9xV<=SAI[> !,F#/pɟBɟwFf6~Yƪw&K-]p඾DJo.% X+Fx?^~J[%KtӨxAc~(2,R˘HH/y F!jbk2R=UKnfQױBЬjr֡m&}Xq}gn8A CBP怣04"p <4ş& tzm`ktXO^㹛">^={dj${Mاpl7ph W[[ƝgiITgb~{ūWZwf4Ќzsߑ"3O`$|YAeMy$=Q~'WO?)٠ݡY3w֐m0bzH_$"f9'/;y4 ۀ 0ݍVh<wq*Z"Bu'Ip_ d^ӓOtJ.S0Ghdk%A'b")t$edjL01qe(U᯺r=%g(瀾>riP3McnZ9/cLT@wA3-UgD?#M"@`,$ySmIT΢TUA ir_F&4,(vq"Eˈ=R9+e(BL45 8HQIW\2jbѩfO ``D]q<5b> l@'.+(f߀-`~ZPÅ\S6&f;Qھb,iI~dgUՍHx$?f7SsbD91 vAg63s4uwN.Bȵf]ǷUІ79yٚcB2+~Sw7ͧ_ ˋU0pLQJJs0O^| 0vL'>[,`~BA B8:`GO~k23nHkVKfӨ=cx4J@= ғ  B# !-ѬqbZL l?18Di1Ķ5p TKYƠШ@YB4j0ٟi8l^Sߟ3蓛xc' #S dhB.I#Li0Z3/CPnn˷qI" k3:=ǍkI6~x O\7yRk >IM *H:GuDxS9;JU.U$rP#p[z{^ l*;7nҙsp0:S0qBi|W9  yi'1oϝxc^7I9 J髙l DR]IXO[)bELR8&FšQCAx:&;bZTlqՙ=;Y5xު8"V͘q), J9b8$`1%$ CaB CRegV]J3m 90q,qDQp@|88 `Q)D@e5('Ll'Pmt(5uI7{H)Fp Tϟ1&0:w5I$7?ݴO_[VFMY F!M]ڨ`ڠ`p QO!Si\&Kw)1!s߻msˇPny**5vctPD ݘv[EjISmU:TJl pޙ4iGgST`قAb738omz{3Z ;$K0Ōpg%طD,qiKwnxI47е!o )A?߶&rƚ{ K"5ҡNr7V0K0HӀ[oݭ/J ޝ&fƦ ŕ?j ٤rWٯ4hZ`~)tJ;$MD5%:ph,W.YfRFԅLߏPmcׄzhZVڗR5ObZYUP"UbG%9WZbG=}+ yRFs/Ѕx4q#fY"g*_#lճ;4WNյ}m+2m Q޹ytKM6Ere_M)޵,cC Cr"d?@VXnoe$ٔD }\d0V(H#ciԎ  }pQzWo>gfb.:8)&Y(g[pZÔHA$ƨ{޵,qc_QfVIፋ31̶c:4)S"MRn'dQdR$*Ssqq:wt- $ӂQ=o\S1NN͆v]fv0JMqI#%5.J~.%>ğI,o8 ,qG̣.qƅ#f}\#aYwH\Jf|r(OVQpaToH4J:Ii!(^z͓VX8<q'=kK$k6αheiD9lEXyBk@#\&RTt{\DBKX9#s7!I?սI5&CoLu<:}{˵wW˥}1ޕtn{Ua컭!eiM` Ɛň JƯG5j"UshDhdi¨`LrCMI$T &kIAEV2e k.]hrGa {:vHbۑtFpR ClX0@QA0!d!B ybld :;/p\+^iuV\ة8A)60N`L85'ɊXpOZdځ56`KbXcI ![13 (&K콐] &zA&etkk̲c" 9RLk8[)lM?ؚ~݆?/7ОcӐْ`fG-dKJI?$rw#Mڽ5Jv@-Q1-z:I*G1p#i?ʘ`wZ 95]h-SoUS͞TV]L508,_$_qؕfT^C0sQ̜w!fFM+츴7uo)8D!؟ BsF/0 4F)G?RA='AS$0O]%*K T s/G․ T9ƀ+J!i)S/(%kԚYf)VjB[>- EU-# ,§5,7ndt`JX=!iNS[/0ù̇?yNo=~DQZe  r{qHT1 Vv]ED0L X|b> [|XC+0(ov]ީ4KO-x+J Jx%y-WDkynVb z0'Pd:kA &-s^{϶P:@x#6j`|v〨^j1(lUV{d'?MƝPQ-:Ssi\&'1Uru:IkCe_K4:>宗!H"qƛ 6s8_&QƬmGL>=xGsx!BFQkbN^3%~+FYtj8K=Z8oc6Iq9`95"Xբđ49Y}z6w ZYv6b˷dgSaFv%1~G@?sc3 i\%"Bb4F qjp`5QaC7+A1੔Z#+cB-0fr`0N{K"\¨fߙQ\K&{]:b>.:).6,:RD s{ʞ^ #pZ;S1΄RjߨDQ 4V 1Syw9|`ƛ+D@+n|~Cf-u[HJϝ B.Q̳FS`Hz/V4p.O##d׵8)aumXɊvjObX,WjἸK"T 5>___?~nS O4I|]|dMVP- w)p79Mb L[?q*\/f'A>Sǖ!Wq ,"3Isz!a%~="e P0(M2Röa'f;7O-{˷ 2ElDO?>kpWwi@`` G :Oy( ܬȓa2I0ZYɅmv4x4o 2肩xʽb9ܫhysA:m9̒8|7p w TWh;,+ݝXQQ{ӞEV0.RQ*ZS\DopX;G- Jk̋u;1e*(Ae:AMaatڷi 0!!kw=pl gԧYh)8'p2i LB^L9t  F \PN@5\`(iu,gҚ[6Y` z_*i5L{7(KTp6;;JTOxp%ʏ9Gz̑ 39xbBN.X>_Ӥ~N퓕^$k{7TˠGH <0ؠQG^DȐv6dRGĘ. ,E,$u62hN;="Ac,XKYSઑ󧫸{[(2Fh&}x6"#8~ApG- f'R/fxf 5Ӏ@0( ZiR\LFSf2A*\XLl$uxCq`d@Qχ8= Es 4B ט`ѯDyO5"&8/}F$*Z0XНwկ2sj@V3ۅ5q1D߀ y'aqP-2Hqqcz|4(8QAW 9æ557D&~ $28#T c2`椐LPeBKPIeG!]\J׳XAVYL(`E%&<:zBlz"^.dMB?M+$({jVB#0Uzmu N+Gd-G^ٸ~\Yr$H0D 6觽rTz}L, udg#Cso>+za=oOfHrK"h%{v8KfN&%|vUӴuή WK)ruv HX~V(yv5TLӾ=.}q$;]I;]}{2{0#1SQzHr>|WOK+~L= f&hWWof< C !^7[#Oom Ͻy<ĩLŹ~ IR.y 3u| )1 i x{.> YIݒO$i|Dt4!ԱYUP(%0{{ T-)5tx3*bc<,Fb.&E*/yJwgx3V .$].蔘Vtq7#^Jlu[ nQ(90S(9;&QrZ(9BfKk$6 -۵Hxvb0S0/Шl7s*@EpMS$n!]7YDRcvP,YG-.})pL*6PYGA`{W;Ǜ %8~$Xqŀmoϟ"D4[Ը 2x3嫦WfFCZ3ulrq_t=ڏ@$mos$Eu`~s$ӊ,I7w žI`Աif][ B;f_ob}f{# RUz0?h Zo,=lh b4o#=-I oq{OQoU g)cbIxiGtkYR>H͂$^>*+`蘺kDcoFMhWɌJwM74Qzcu2>zQ}u"3G5V= MA.YESAS਺F;x$Q7.<:MSĊ;CKOZ4ᘫь ts,M/]/?Rlz[D'T^DU%x*YU1_J-8 B8NZƁZC8Z*"؞` N 3B`Dq!*V#ŃN_eT$A8a4!0)r!MlNR<( zRI(Bl[!#v/C2đx3$ 0'RY({sKQ-Rɋ4=^MeIH%3cb,N5"m7dI؝ORD `8<8LrE 88(c.(lʞ>1{Y!lW0'qUJf#)y8!U\W! Sʒ )"P:K($hA)8U Ҽ"dgN6%Đ (,J|IfDF*P8]`H`(Kl?iw !)g *82F c:, gćyBE 9*Dak*IZţdA?R߸GJ(\Oj<$awr /%i `HR G?)ݲjqoVs<3L\/~ݠ>]urw'Œʗ8^Íhw]qإ|}:zwz$u﫟~-/?sYΔFOPy3:hMVJ1H1:rz#рޭޭ&ZdSgYޭLky"xatuT[qʩ:AH\0dV:dv+ &%Yi. *<[`[6Z?.K+9Z?h//\?[OFAcⱱZ)@-izёY ){36 ?& +ťuE#_\\~'1.a!?ؔj#VA֑[5NnMn5,7"ޓqdW~fI} 8N>lNDixxY&ixEVMc[fջWjҩFkMԋZn49vn-fqr֭\քRdNsݷh.͎g.E2Gk֍ay}Ki#:sb"1.͵nMH3b2z7rg 2`Q:)^*kI(!40,9Fpc\Eǖ3rɤЍn9} 𨺦 xqqdRL}L*Q&g+5x]5 > P.G Lߐ7ϻ&Ot =DzEW]0t1fWW`.Ϋ\E,l4`Kե/\_|K9(d?}|?撌cbPx$'ƨܢruׯ C)o?]ua`괻 \EL(d*1 O'y GtSE7lzwCr o%vIJq,X7[7\MQ$h%B56fϾϦ"HD ̩_VҤHS)$MxYR&dbfGxW##~*Z],xǀs? cO^_+Ʀ kK16mȜbblNCy JPQj{B˚YInϣ<[xܢ [qm_ap}pcp9Njtë0;EoFZ{mÛћ?%9o+M;盫Y*oW4n?(;A1osgJʼn`=,CdOHuDK_ۨUiʜӀG` e'gJ3΅ɤ??ƠMҲe',Yxpf9~r@6Gk?i-8\9ݕQre1)LW1(@^uawMո>Ԝy7(-rfÃF? 3+ӠI@x8?>Hea]pFskx#['k0pۙ̀Xr.ԯh /)RkִXt?:eg \wҢY=9a"Fy@MVKreIc<מ,l3?,}?6H~; EeQ&r)}~\Vg-g*iEb5Z 0>pROU0 s3X;x۵zL?^t>О->Kkg9#-.@^"8yI!3BcViҟ 䩖aB>g&Oa*B>gf%&ƸN`ӭ e\hm3Vпu 3ټɢ{I.FS u j /b=qF1j,Xy$:Hun^L)( uIFH}Gl`TaITAp-ʥr{2JU9| M=q/8Εf e*(cs @ C$1z*Fh6(n7}RGm/p]^-IW1BkmKDbԈ8Ľ #yO 2'$,%PHig ^YdbنH*F!}A"W5ER8z$\n"0,n^څtAXjjzx cfۑ(Aշ<Cs5?t& x H!RrÌ@Ia}= x0`( ޖ9+?[%3W!.vXd1j#g?FeŜ [vGZ9pu6i`u@Xv+vۑLRwsztyUt[]Blе~Շ-y;FS+{jnz~@xBE}'+8" J~)sPxջRYv:pPb g=h*6й&Z9c]ϊ-GŜqޒ|cŨ-s<vM?G<Ϊ=_`Hi!^(o?zvqNe' #O6{N ᣪ@(c-qV)lHPƌFg5X3d<%Z5U^#Ν3BC(1K,KPAhDN[l0WH:d,Q̢$+%!nE2g0`? '7tt2i8@wusIơ3LfU]/\(LKomr_G>dćWJāҤ; !"|fWձځ ])teBA=tˈXXnJغ 5E;Жi먧xgvJ'*:S:̵ք]./Zն4/wiiviiP}'fd% ^6 ZfM9)Jdr:OD$a$s8N`AF[H.AN$` &YZ]LͿJh(kY`g)"Vo3ڨ5_`jQpNrB{;ݪh8ʻ2W,?|&yw^ uD#u[JP֑JcpBㅹL#h.0pÀ CnNpfЁyQ ޱg"|{nS*0Rp4 cțR54@vvm !˞7a~Ά}w:#{& :z\va)$ $("*-L;v?Ngg>Z c:q|})lMȪ74;0]`Dl$CǸTҞ4e=Ȩ_ U7BdmQ33d܃y7 ޵З ݏ!D?mR UIiHVg|4kKNWZs^ђZfW\τ2ڧZfwuI?{r]Sy;y' !]WϮ= %7]Q7W|~۟p1WRuV=AKUIR a^VA2c'#g1;6X: /$x8D'KÁXpDN,fj\dY>6)/ R+8q8 qdvf>K"4ֳd"\'!z崋K"F.(wHF,øVSbu۳ByzzYg87 A1Um ֻ7GpʀIn+ߟ AO#%P~켖D +Hu=L&ZG/0M  K . ӕM1TDh-hZT;x4Igf㤰d}t: U`$'R31pI6{>8p>#gcpC֊c sCLԷ'Y|?0x WCLv:SH nL:}`+\ח%So$7 )QE\" Z|9Lz `}) sc)@?pi(!(9i=DO*B!Sb x0N$GKIܵi͟*musSx!|l0\>& Y1 pxG\~aE}v\A,[<+3\KsdJFrx,mrZ#5'B| ȥ^h\ubqs]US7o{yĪ/|jC$dG,_;߾PBXK郃ˆV4|jvw֛僣p8F&P1QAFJ+r5Wg/>ً{EE^,P/(d]S[a[KUPg,F K tb$A22 ˒2#[6 i4m !:4BY,;95:)|·YFIᏪ 3۲dS2[?gdwVedwedwaR[4\Fv-Ĩo1#ۑfc5BD;A3 VY%e Qe:&gF T~wT\sc*TCyw?ɿ9u+V r뎑|G9xɁ&Jap{!|/KTƕ*n[&;Ü؟1I)Arw37ZN>1(dv{PRMk<ڸJ(F+p^TTy扡c=Q2Lc,Iu@IsZo߾$Émm.ƻu0!R^§.$vHb&ٹyx $9rDދ,^ۃXklK8NaV_VŦHl7uE (>})mH,/}}eIqZMaRrV>!ʒo9Pb@S7Q}ӟxҪgS?|%.\觠a?5hR''L9&fJS6 Ӵݿc X.6I/@;)WNyIFK2k鹏w?VvqǓ'O?4~v~UIԉ!Y(U1,W6jS> R9)-Md͏\^ͮvuEruWGy=h!n:΃oh-Tq:*[%$}KzxF**zNJé #^QQ[q.%fB#1'F''T$iTFQ4IIF1-@ب\W"@ q7Rh` e0ڸy2r҄:zl1~_ڨOSvZ" ?¾4@jM֟Z6)g}5MZKN B8oa8->$ѽF$o).&œ(# zpUq}{A#A#`B0rSM,|Zphu7Dub,أ0FBXǺqO͈uAXkr"1ZbYT$dRd#R2Lr*{WJAn:ùa(Q 43q؂$`;.G8/;]o?ZF8;r2Ӹ\=\\ӢDTNvhAsc!a[8}qf5HٕQBvh zC%tVJz@0)70.eS M0Ĩ _ \򆇽/X0R]&e7< BBGSZRGǵKJN סTB)5Y^yuW15,K+C$~>#X^[O-ٵlSSx0}2goa71J;\Ѝ;3pty`6L{lF@@̚hI`įJW[y(Dv:jx@fhdj;щ)=x@3*}XIFnX#0p:黷wڄi'aH^47Lu +pOs;CLIeu1/1-@+ZW.`IuzJ˫c嗹eɨZÆ"WOI0_r #*mbMY2Yxn$Վ1Jps6V2Vs(-偺'&W (orv9IΡNvkle(J1c;JW;*]YQ0fԄњ`&YI&5Պ&UK4MV4'c)"^̘I:|`,u+ph?LRDY!lL?D+.9?+i$>Dx3 m*je &L%dK0iZf(`A EtAHBCFM9YJ#мC8!Ӗq"Jiujul| ;-TZ !RU1q$wLqGSC)XuFX@<lC""HyYCtdJwJw|o/]\\JoDEWvq=߿%(x3:%Ac ]k7KJxe~V < Y1@$I`AU_Jv6)N8WBҌ%ʼnPU!j̍l}X*b%N$DJP.9R3oIuyii Kc&tRj欤B/4H ĐeHR J'<_; P؞#!Y_q,W/Ɨg-ωa~Ȣ 2'IGxQKс T4 Am.7媸z]}{=ə!|fsr"s7~rZh(.6:Y*j Ώ}z+6΍O߲oVɚ9׋ר/0еׯa"jnU?>8pv3_ tq`Α!ū?)iFkBps6ڛ_:9feT)R('FWKWe.jN DP(Ig\b흚K͙0}o&@v/|T>dAiʧu4) 2&IY!~mF8C=D򬘺N~;bqZ|{Lv;#2C9P8)CvŷGq1l{ ]*Z=E)Et$.uowbTkPtvgR~kSYYz}˪@RǽaE&@I4lI`kg%'pk;pt(~̻/_.ˏj#gdeѝT*1ƣW-l֖ U@M}aākN2KF%%fTl0yq}K 3Xssss%uZD{Nq陦,B ",)$$@,Ps7J7bKlVFlW7b[(ж1_۫OUXܠ \< hN }![-Zez/`SiػFn$W;4)3Y`l[IrdO[V!$X*V'0PnK,zJSh15%S  Di* tHKg$c9,6T Wlq +v?n^`t߹#HptXRzR2URr VY*Bzg"8 oH3>nW+̩1H|`(-$MvJ I^Ή-11p1ql8o'Us~߼NL \ krna8qҙ1 ^kE w^ٯ3:? mS` iuYQqyɐWi+E) SVx ń_J4fwr2:hSRbr :^XX (D{)}Lȕ"#;[R*bW%FUX*(xR wJ ,%j=E{“)-t䨲A'p J8p0"`w# Ŏ6_":^bV³E(p@eSFq`J`/'U#d&D浨4=cwVpWU|Qh rUL[ӝs$jք:B'ŁM#I_CG/%t(<-3?@CЉ}.Z;~բ*^0D JU,9!Y!RO!ANULLJ0VjiYit<ͭI6@7g NZ w~ƎBN`ů+H# &^*ZF`MkWӬ,kOHH;Dv霘Iwp"Hgk[}-kelmN-Vin۟*}zF}\)ܾ#gS@՞Dƿ&n)f.~[w.~~F]<2sO7wnOO୏~?cpLjur܋vl loSin2$p*jW;|S桙|js֩Itܿ2E1dMv&pm}yj]>ΧWkEś ObyԺ??лe9Uwr<C a2G&gA*Iv~AìdV=sԶ=ֻ-; @l;/6x7.BY3GU{ׅ8;n`SuFzAX_vi@&F^`δh!+0T &Lul=h >SP6mZ0Q}*ؾ{YOٻG4ܪ;i5n9s3{2w _NWUصGzD|E;oKoOï@r_'csGjsH1NQ_>;}/OĀ *7KL\)d&޷}$;; aVӵnOe}Li+0NLz0bpUzGf^窹MzK^6fjӡg{שCHYjhQ k\m$1ܷ)lPin$ThCҪX%EGǩLL>bۍxy3=F~8NkDωt \k*Cml^}9~S_}1|(*-mkPc.YӮ~A4w_## $3(gO37͉Q"x@@B" GS ԺV#~#/Omσ}FɊYk:Bv$B=9- s|0ճ`sfdR3IGQ]m|B\sTCp!GS߂[6K9zrBi)GOOc\s|ERڐQ#7` s5 ؅✛P/X'23eo@D3o1 zN7=WK]9%6R뎤? =%Ň>_7>pUl ZB"deXZ5Iǹ{7H┉w:_BZfnnFx0o~;I bl򫁣O&ޭg?8-)DMEMEMEM XLaX$LStEɔ[?0-*.8f%G#>{K Kɧ*gs7NZ 4j:Ok~\MGMf4[oƜ|mP8Mf.x͘DvSQuv6\UX_+_^W,_|W0IF`C- V(]}YaZ}=8l,`b?~O&y㇆V_P R{m@@ 0c +Lr, g#B3#m Ш֫g85%INJ€;Uh,3PR_BK$TrbKL9w̙h R*bXs3I,FTRp[ P++LauJSm5RA)`k^JOJJz]J8,V;ąRE,qqxP3z=C(SBZ]zZ9=/ʳB3p n/.x Elg0`K`[@F4aq*Nt_ee8-T<t*8EDQqj  J 5h/e`F@@ h9P lA^,k =O$؁`CDXq,぀kO74(T~]..LUvsbL0[Pn_Ꮻ>^o^U&x'984^}4kv]'~^V~]uha11v c`0 e~qܸ .X9V_o"`K~~_j1O(lo v)k ~_{'vR[$ *MuAq Ҏ-^]RqO1\JFQ cc=rXoj-6G2g,+%}U k_Yk*Y7nf-zЋh%%["-+Q) ABKhUZk8=)U u!AO;viӇFH'؅(Ge9x]!Nө%2 %iL%n"K)d+ '8LPF-![Q 8>( &$ ԚHl^.͵y\/v˪ДhN&_?yؓF gkl'ֳHnQf#ڲ%4>~ZYGf-q {l>~_?]MV6lcV ?1\I*M4Z~=kQRLg-yXF&b𥇟g ~ՁrYu)ϓ|=wGk&&tUup|ۖ3M];>|~wSclg^-}rͯ]SN &i`'ٷ9Ny5*W[#!.!2k3ʺU!XP NXt'7ݛu O7n1$E4H|4u bDtB8ǺT|-\ȔbQMAXP cu2M #;5=F  dمW ^>A .cRFznE [ p )-{mTMZP NXtPTNe [ p )|Vut'P NXt)k҆&u 8кŐ  j[ck& r41αn +[ո-\ȔRjUz[sw Ġ8Y[K$w w !.!2[#QI,W~ϘWc=jB*Aպ:eEX抲\QU4Uc*if W꣘dc :WJPuì U ll$Cɜ2%O0DzKȖLq5&m 9v 5ˬ15ϙY a=ZάǗYcY˙@]W5S^ά2kp;Yά]Bf dw0ƈ(grf-&QZάIYGMșȬQ*2kRj3k9S"㻳~Bά]DfMQƗYSY˙ǘYRY˙e2k!"sf-g"j8Oء/g.!22koo%d8bH.Ƒ3kYHQ:XY˙ ˬq8ΙY \e8֌ZάEݶ(afBxάZLM m5afQNάZ Mo'̬%r7ȜY 5 %grf-&P2k˜Y J!Z ͗g脨{o*y|,w?NvN d ~JE%rZOPx-d)(f}Y^ϊu7\__eAi~|7!z KAy:Ԟ}M,PC+s3S=j4W(8+ b , D yp$+6gwДmaP䇠ѝDԐRN28A^*M)b5XaU |Jx+$ጥ(CJFm$&Hv;D; B0:Q;Jx"G\ e@8AhH3٨i !<[-J. Bb8I BbqN^"`@q(MCyIq1N#LMY;9z YsJ$cQRb]jBB$La3 qBܚTshKN7Jܳ3iOI:֋ux-ow7ֿe:6iM/{֠gkWۛ❻KT F1m|8I ᠕M6>Ɨ7RG5×qqҩfB 謵}8:D3MD捨}H+e,xקkGӤz"]j/m|/s׀9Azi}VHbsUUߟrnӑgm%M/OCk6uV ~}|=]IUH< ü,d%?/X &󢃑_Ak" zA`A2=AB D1_0DbIuMum!+ƒ5A`KF(6 PD0墭]y"i-8mޠ8]A?ٔ#7c{f,gmBxLE'\!_=#p\A2|7-w,ݣENK 2чT9uۇ(My̙*I#^?go+7GB_8}NlaǴ4WVƓ|3˳6^EEiOn~g8`g6%Isۭ?`Z]n}SgL΁ APYO;ɤ]]* C;J$CNբXel!wގ5;g}T^v3t[z:9k'^M&ק''T+OMy^k[w+ o Hq b 7 ϚldQ-pMެPan\gp־`uIt=dFx,:`+sPb21g$k<'8&R LYYrY垈OsSz9 [}Ca\ߎu/GwTBʤ%^*!D=pr" ɐ#(n5X>ΗY_O/v`/g#(BKM)~-+UZ%&R *3ؿzz[v<^1<.yV>W`}2JJYJi?ɤ}9*Mβ\*gq^O79y.}tcrX\bhȋ0OŇǫ&@UYYwo^p W"rkW~nѰ_v2[#oy-C =\$ P(&L I)- R1{YC(+ju^^=Q&r}q<,H?y_)7EˢQ^yp+.kC/xRTJH/ YBBkEN]yǔ@e<$ -)Ae{LtkU?z\dE"i JYʮ/Es5*w<|d\CGC=W+O"91063M34e:1/U-Q*Bpɫag J=ܣ`Kzj!Pqicj6^)B?4ELVD !Ls7 pUS|TI.R hMiG( L7Nn|G=^AP #!8ؙ+0Sіъ^7 sO`YXkAu>vm/F(iWoߋŸ⡤D#jF:c\Ȇ6C/(#8Z~PA+nR D>Nj LS.8׋54}lzYj6 |tpq;i;8n' [ޭ4n8N3.>E@MXdqi(~|o˵z~TΘw.0\zRx #F3u v:hp!h hUV qg"ɡ^9g_h[0@l}zٽpo7!kevKut9l ʶ.#-|MwDĂGYjڃ1cM$dgYTZ=2r떨͖[qu+k!xO$^mj O ]M-Ĩ Qh6 QЁİ@bd)ċ't*sj>WJ+S'EZieH}R,ĺFN[Ѽ] rim4NCm (];8&ͪCT?^+q/I^E;`'8z\qS :[rgT_tjˡ(xrBx|v#LIxJ}>SlќTkӼ2BCQ[-R9HAZy\KA{85 T_p81IƸ:C#SNJH3ƳlhQQ;\(:@஫C$pCni(D;"QJ\^)K[ W;3M>^!B)KE4+z:zЮ{fZ-qQ .#;<׺z[6؂C{P9KOdBa{% e<˓V20άlbcw.P)?A}Dn*#i讆2)A׽e#z*lL+pLt͖ uЌ!ҕlC~lHSG'Hxo#S- y(p XNsl05V+Qs/6Js2xs _ *s>)hC4Fr1@?%0UrZ=!W}>5 EwP. . WV"7|8!18Ǜo続yT*@i]\lV&:o5q#~/Q͛Xb¨$#Iņw,qWT_Ւ_Hb;qs]nkmɗ=K#Rm?hZ V}؋6B1}1;G;_Fޣ[ϡn5?Ths}::ge ow;Q UO6sy |䡒ϧ- GyFY+z6J&2OM?6؇˧7^WZmHx1gqZtH饩3n$wOU _>S_$5gHy󚂣AUx:{yar3&ՒJ]ӮH(Y-*,H+~ʑ@7Y:cΫֻ+l%ֵ{귐kZ.伏>f@~y{Lv]{l zuk DP92F50=K8Mn:Xn8eHnLUr\yRڙh&_R_x>+]QH+ x7i wӼ~tcJmtr(* K34-@YnIen5s޲QkN ]k|ۖE-8yBqlɛ-b,*mz CO+䲑{e4O}' ËMА, J:hճTA ۷TF1z9l@V0\9XlD%]oTחϊi:XYt8n!5w9hNsl{5!/B@i{wsadgɢsN< d+iPRw}CcQJ#oa)@2hi DO6*YwyߧelP1̛Ҵ5E?R3EaCU|(_;W;d|7^#J) J!9J DQ(d,Hj~p%E '[t} ' z$Pt?MwBc{3e\.ߙ]s7:s% R%008St2cy^EևHވ\7!loK dNe ,Dm8Qeب}p(@Y".g@eə1VźdNEm ZؚZ6&Cy:oDV⇧nzХ_!ȥ2A<" 3_(^,_' tr#d>Et|/^n)̷@J!ok}YI+Z|z _W#d]wj|^f:Er:)XJU ֯R~UN_i#RɠAk.q`w m/)R@`=,`vھS{3ќ#jW?s5=ѝYK4+W?Ú V3U$?PbiQ’eJ ^RĚ5m(x;@wesvK+Xb))|;|lAKHI)"P~::A.#VAP`wc+?-ҍ]kw&WkWWn?.?z{d&bHjmGriYmYqØZ W2d'~5GRa@QFbI=}A"]X 2HߩфR5Al$otܼ d[b)BA+sR8хRFiD`JՆPZY4rmb$(Cq6jfU-Ap%EwjMuz+r>vu89ωѹ gW>ԜBOV3iư5t}|P&D 3a9XNJ J8JNx.MΔ4;ѯIs@Ou=$BB2ZɍJ8EHVe$DzqrSs|rKԃUbA5Nڬ<񭊜VB*N]Mn* H7U8jcu".#)Xf}%y(z^# LMcZ(LҎL2@2maS@s!wi*@Xb#gT}OUqI֤RK_l!&lʝY Gi&v1A3 Q2X/=72LvӍsxw2㈆%컄X"_dr>A}cnJDOapsO K̻>d9]\{WaFDdWvAvL_ Yj{|jP\6P[q@δiH{K-N3z>0g:CyY'aoSr g[݈&\p+2H_Dy =VmqftMln2}9}1bLT;vȴȋqWM7 HE-e JKYhm'EN[ɽAZ]D-2B)f02𖵂vx14wҳ[I7oF6Z3Sp48S*(,wQqd,B^Mr.24`$3~~R&߇8 Pٟ6%f͟#@_!w۪(LJwOSzE{y$C:45<~ vM_Lr7̜JTtL KY:6eKht\S`Xf3*20 VӇzqО'oʀ/_@3ENpN?8z,/Mpskxo#&RKS*T@JXw CPRU|U1b^tkE7@Ewc1$4Éd-W&Mz4uRqud57|R?ewoH(*r$QxI$neAg-vpb"ߊ!7 0R싪%5Yvybg vqƏ7M9|4MĜ1jy3hU{]d(aD&ZiΒd,!9PLv̡#a2{ L355oFޮC0|ˊ/u0clA-}S r>87OK}FScm`?`>MqR K߅T2CAO?tkp492 ;l.JgC&"xʶf  R 4 Nӯ:4@+l i|۾TwL?hh'-MɭJ<GWFPR2ctT-kݚϺ@gGep@QxKP.%7yXQrm`L`7 44CR)'p~Akb=ώ p2y CEjη#l kl^ᜊ^250| VѦ<̴^aϣq_BK4@BZ$R:,J@HYR.5틞4֧i]xX N7Z2&dФsodsRL&&;9mfmcw{Y9c5faTaC'PPj 'gU: )y=8=^{EXO1r/f)U}liAip(y3hjٕO"4wcT]^OKuHKgܳ(1LkECrmSLck-]8i&DSu8zڭ YOcu׵[[ rX;h yM8vk\|,SS82(zíAC0ӷסHx??ms駷o~-&oߤ<ݫ#nCū_T4W΄YwǫZZNMEdj˻iw|»?Zuatunf8ۛN]~)^mC$/a$𣑒6D1Ftk 4:eȕ@ae^_`9i)PЅ[.D'P'J?T bL&%xPPfr;9/Y{~"G]14.6Q^O_"Жw>h x<%>qAcp4 ;h]%bߜn^8 yU8SIVIO hF'Pp4%B2HƬƒX5`b+rەO?@(Q"?u*/0 '\),-P04s-1JUJ&FE.RAS0ľ=yClV=5{qQb+vQ1!2M.n$uٛ ZRE0> B]\]mE&_d_)\խ=/՚P}HOX:aBWRzdؤYX5FԲG{5vv5itTE@eQZsM wޯk^MUs /RM'闫Ft3#zL5ںUytVJT5PWa b]Ys7+ LLTH q(BOZ'3'PU8D_bApdW2L$ė#RK:$w,5uõyQMZ2Aj!7e`Rs82R1! H}G GxQH9rdE| G&$R1qwDg_Φz8~*-|.Sg{ O"bD(V=Φ-"S fo!߹v)%b23Ysf!-7%̷*B/HՎ~nWAQ`BK@f@yjB+Y2˭J0[PCT>? bRK8ZutXjRBbWkSJN/!1a0kbcaCsBBxh@; zxkIB.瀮܊ȔcQ9'9Y)yHs\eŁ/ͱ4O@PNX" lFq/;AZET dL|s3 AIǺɖ,3`2H{-U\˴cN2Gwa|2[w !l S19M}GL 1POSrxw)>%H$m-1ƌnN;x%,lf2xwI>%g $[*!6*bg-:0Ew|*XL lΓ\gH G?!}p&(G=Y_`B$nj]aec+BKZ+%ϿʯX|YKf!)M0`~_YX&9U4(a (ohOg0K\мfV,f7 >g&՘FWa$&ኍr$0JKr [EKx>R,6cF`hWM gX iƠ&i7m\EM cE%hOVW!f,v6 rk>Qap,[E| CVRXApyǔ!=2ݕn(fٽ``?^Xf1ąmOLXz k>y+c|XdxXQ_-.7BfM톧KP*&K;;Z +Z7H aS.vSH6l)n?6x7y]`S3ނ܄"aEX:HKH8X$MeO06;S+:@(R="8>9<(s CD GW}/|<,F. bl/PNvͭG[R 5 mz"̆^qiUb4L'Bog)G.RLJα)S?A|O'o,*o,HAO |!êhBXh cB7P6 +5XM%֔Af {ӓT茴tIf*ݰtAS2 L95R+%+~ݿ P'q#?]2_<-e!CDBӷ/>|q>"aCLsd J_Rۜ ?~jjO DP.//`{M*ZN9 -(MQBVBAl.v4kc>rz,\T-OמGgU|uπpo/!WU巧ΏP r+Qr%6b_mwֶbEK]p^p ԖGޱpkG0D֎3ζ#1Ȟ~f0]iA9-wgL-H1P9jH29f3=6{\g>>[g ba+h+>#=gx'#=P" W 5@;jkh#hiSJu׉]QAޕh)I¶@na[iW/'{WIo DQ(80nM\@ĐF E)OL|=ON1;VS 2f A zPB'vv6cz:̳֒ƺ1;=뤭|գ7:=Y+)X[FCBA fY=O*'Z@}+:ÛyN[C.N^-n]kan6 fgl_k}^7Po Κ% q?)iE*[ue P,Xk/~ǛͲE\Dqwj ?<|lv+/x];Ŋ &`[3H.Ԙ!lq 9 w\Ⱦ=۰ A!ԍTTUmZZ4:˔%1h$ >P!T@562t1_+E[14qyS:muYG ϑ.¯/<=cS>|mpP,:y4Ō3BKB<^ rJ)ՇD)q=sFLmץ-W~ ݩKyap~rC:ywE:9o>,濭-e7WW+Wmnn ,BLF䐌±;ϑcYD/b`GF˹Mkxz}RU_ɶ& "UjZ|դB³%VOJ RA?WWjLťbޫuwjna12Us&4ΈqԖ n]nMT_MqE@}+&0?@ ^57o^$U(r"$JݺHZ~ km ڃԤ.E#K7tQQ&jRQ).B $ܹ Xؔu' o?)E:J h Rhj2H 'vsb\3+NpNqMa$=J ttqA49Ct#aD K 3 .s\r *T G2ل(`@Qȝ4̴kGʀMLA9m͘Y)tmjS%uS27Ga'FtݥvC#xȋ~Gد֕޿;n٢x,T)D "#q.r}8;)(@)i' S}>$Ba?pF e.6u$  L@׫bO6O~tw7/vV[RίnoWO\߽<Cs3 ~g;/LJZ k]He xv;vfv@xoz绀@t-)x uCmuޮ># *w F{bAe.|f=N;+NlN ?̸ٜh9S?ՁLA}fU]*"A7 m7hVZcKs*(3Kw÷\dÐ0e yFT~hW(y NJD*M $j9"b麒v7 ~ԴB]Pz<[tݦHiPSiA@'ʵnۅ\X-Ox˃5(\"XddC`4`Ã/O\wz4X*i(BoـK~Djw6ɁrSuLkݒ3Ӈ‰O{خ}2RuShrm* io~u"P.ڎXfz'vX5&tW˝w %.-Cp!_} UGѿ?j.q{X偳pEKyO'qή*^/Jo愉 Ch,BHRj; 6VMW}J4&HJW܃K2VP)u"!2+.[c)Jܽr-}!bޗ?z^ akVCcYdԎ\E zXd`7ɇW-Žwz]"ts>'} _CŭH`Mɽ>g缣4;AoW䛬,r*WsKaZ;Mu-}xMh.f[*y#qr8?ّ,aq.ll]}(=1K}J՟ qKԫ-7}1Hu>G}4|xԼĉZSsPh8>bXt?_@SR3hlVc"z<Y(y}L E MWDvr>$@%z ɷ >IR{EjfS2Vz&]|OqK~e`b}JD`NPЂЍA?=ruP=~NuOژ7e]nF|O |#{KV fi\Cnat "sM{\6X}<@x9eh,3ɺeK!{JX! ؖ?h&rR2a\leKUf^*K[IU6ʳ є~jyokւX@>kn,(83۱bξ3+Ms|AUi(Q'ѧ |5DI}>v$ך73~n3 )᜵3q8O*qIntafTZ ABю)=gM !ߵpW/+mJn͙,%4d8bB=FN k uTZ!>yU OADJ jbgLB$I 3>9V IZ^<*E R-a} ᤫ\np^M'6wݻ/jݳ;`Gj5e aO?/_xE?? {sbK92Ƹz4yx>Uw@ȻowaK F[fxZmFZb}* -cJ>i 4 ~X/ԛɬv>j9_P#x>Ԍ_y|m*yf?>xRJX]m( [Rdi%cG@P/z:8r@hz0Uow!'] P,͇QSvo*PAqM%p0L" tLJ0&0F;p2%B: ^~6sE2;< |A24umbAX;[#gjc4J6k :yR+6_MB8>2|cJ_EB 5mkY1b@1KIHւrHJ`$!xtxP+ O͊un]n rCһ]s-ƽGfx)zJ1e I$1SGx;EVTj)3 N;C)1!U/+Yy}t;uyṼ Ρ'7waWZ˻{(Mo2P"7kzb'qp8G~2޿L 2K闟/ ȎS?XF-O&1=>di=.U.|}d=RӸ͊$s|疦YPQ 9 :d^'*eK$xhvHlyNOq+1_c,k5 v%PybdsBTTa|Btj͠Ydqpb&8A̼A4Uqu+9jH$%MM#e`Y:==4Ȧ4\wA%tor8r vČljcZ+-sZ`@]N kIDE,bJ@Q60+L 5]L(玎D3-H(/P)f##*7@"0S$eHE^GMX(86"G;IjFuLKyԶX ej tpZx;sSJ0g s%Wpը*;7m;3T4U"3Rw04k#T;MIn#M.&F88zM0hV208az0\{Ls1ϊE9XZvy{( {a"n)$ ckP; FׇԁUS$pY5 W)\?І^F$^GiH):7*82jا3uO~9/.g;&wwۡՎ`ZYmv'!_4麺EN@;Xֈ#!K|^"do,oWf*b$-W.j;^ezΎղQ:VFћFm@?mk'-:(!1: B IC7ng3/ɫ٪,VټQᤤ&V{~h(0Vxw57T"k!^$ v Y+'$?|c_N4d?|'z3o ciqDWF"Ηg)rJ| RbNZ/|ʧhrPz)g0NGKIh WٽA**+t!厤j;^jhx%/(3ugy[# kXQ6bH(4/|[f|h#EfLo5MYbބl0&(XL )QF)p ״&r&^=Y+z-&vp=>eN.Š]/ tujɨ{wy\?S}F۷mgrݛbD?}S4ӬhE]=)8QGj; t}8I ee[3I2Bpd!>RGe[li>1I 1A|Z*O>j!JYHR~vVȓ֍_eK-UWMKlfIxD JdtPJ˨ hG&+H`%h|q3/̪fh)eMnyQ7״+g '͑R0[CޜAYֹ}J5^;"M\Axh'y=zDD4Nf}~ߞ7X-s}Wt"c8<~.F(͎0 #8!aUB6|1'ly(cc_J?)4ynyRhYZ34E#a*nDQݏ$NHb!O5_=Hw p_R3]VlaږLdwmoC}\#Gm}TlOrR@]}T9!Phq!6S&PB*N!.eKMIz)9h6D|H%z-YN%`1Gr5ƙi_y՜4[kśśnOpŜ`^t9N&/&~J> <(K&rZdA>b㇙DOAw,xwM@?e!ɹ-Wtj^/h2 hr7R;//hvNN.#w9eF(-s9[ukoGX: DZ9 a(T` /-QJ!0ٻFrWr 4x+*@^sy>hݾ[ػc[$vSff{ݱl?X7V)ZiDXxeE z/DHA_UQ*צWv7Q#?dEU]/_?QiM4 6Nz琕 % h+R:k+$e$%EoAyīuƤ1DEE(tWZ lk6R0/C%F hv\A:e6?Ov{ߋC<7?Yx+gZjeͳhL/Rk*{?Vby8Rf[Gu3-Tsmߝ$4meh:ŋ9|BHL`Kd {ڨ%ؙm(qoŻ+o5 G+Pp,47 97 54}.x'}RmV^w[M&z}>Z#"iQ <3iKN\ ]lɫ8E-{ɔ.0isXB7*<@WGAqA֫w1]'׻)|Gw1Q!vzQ4 \g#)|G+82ToӢЍ:z! I4/YTx܎vP]MG_o~\?֛~Dϡ3a P>c5]./7 !x͹}B Fz羇Y?2N{0UiPtr '{9Wm>gLSt?x?LfyR{@ۗcV|kIqtnN7zla !kŃi3hۖcOC {TF@ VC';Au3i,W1jJoqksDp!` hl4у"i%kABDH;;g}S#g`D'BD#/s ld Z&-.԰ѱ6kV2ƓminY[Niϧ@gԛz6]{@bwYv/ځs;׷pBq+)2%U$y}CJڣoi[^&k!sȹD|@iJHʢkYh{ 7:NQmHBm6!PgHg@P\r&C@&,l=7 9:ޟtuuu "AF&g8Q6]8%La{'vϭ*,p|_[5 dY vqCY('`ˀ-к_; 1߮rN.u=0'`Kr=,,U5@cJkJQGذcV[S&q i6\f#h!xmÜkH"̟xypr\ُW}%Z7kr D8cǻ L5^yz0F.ǭFtK5&ڎ)) nXגɳgӜMu~sG(R܀@0 N k}~r`x=cxC*aC^2AO.Z*돊l*^|a<LSVd=*Nr( ' QX]\,pNNx徑xP"hczPCI\74>N wFvm#:fE>Z5 Ӳ Rk 1V/L6SվST5̣!1|E1;>@7?fp֭'ځs *E&* ϩ,?ФŠ?Ӣ=zޔm|9vϾ$)epֳlqř[?}|>ﴧz~MO<\S7? } ŗ:tCœ) m_^{ p~/V}_az탣ws| '—Ct0HX&hTۿf5?@]~O8k v;{͂XE8td)\ spa((¢*rfwB[Y%4b@n~Y(XtIlmqbu008'#c8 RUJfRF 7l88xO2}x+_@aXX'b[SP79P}:f=%x}l/l[nS:#'#j9&F ?!xx"cňVtL:%[tw UWް S8.!;vF#? Mmמּho"{Y"j l:ps[+v]&4x$*gYP{D|EtXCλ-~5g,|}rmق {q/#ch 쯷rL8}^>KWFJ蜵~DF[ӱ.7E8oo+2~f C>h\f(+),D(U< *\S'VzFPg 6{8kU?U1N] hn&=#!Ch؝_f4MgШµX0m%9@"U4*4?$!sc !>x#[Qހ*m⶛TRRi;"1Hw޸Ҕ~9MKz؊e?SآT$DgIpl-9԰7V0Xhӑ.Xsh4mEQ콭TFw)k4e8ekMWiqLH㐛Oc?Z}ә5X*ȩC4B4GյAm:):dGŝlz :R> ܓC&qډ;/-Tpn@#0'e۪;3>vZ9'Q#p\O>#- >'KR:e7El2V mbwNJ?G=40Opw{K>Ǖ5zN)2BS*&/[m2J*\M^e z }*+j9YHG:M$Zuʶ~2]4Awb t34(XF5ϸAvN NB\o-ٙL5LusЦ`Fx.3[OoR2aba,,Q:rV(` V7%&)lGw!) ug8 Ȝ_UwWqz2-k_\zT`>L[K|I0ҿ) kXzIh_ج8Mj1eR=0R$1) G63Գ(lfƒf^|upQe/=Kx?A*)%u u 75~8OsR6Ҳ̵p_zS)͙\(ri&d@Q ȅL 2,*ժV6l^~X\`}FVX5]__%\c!!II9օ[[Ө^0}5xzv549*9oI#Y۹ZY]*N1g"c^PpXM̈O"Y"ْkMAIkYn3XF;+4ъ6 c@w%hp𗼲O~RFJ  ^X]*f TIڀ#sVbRvsG[qa%ӊ.B9(z]gHS$_j~U3x&VӚZXͮU-s&jgVo0 v,;y6f䬱+UЏw!U!:o!7wXgxbϝG*EM1Pnxd[-ʇDVL zZ[-j8 JRwX)nX2BUW*gzpX íl2Rf^13[1 TH2 1)܀qgebT`rPK) mkh# +UWXɷ~t+tfV;B0E!"S ϥ"Z=}au󩥤?(1=F')D)Ry EԘ)p80x3 je0qyYzfe\*X3s *ZaLg ^QމdУ_f碰xO;raxb$,}?CC #IssfE$NZ} w SJ3ϛww>[I,だ!da0alCfA24J@]K<.ah'vԇsҤg_7wpfgrmvkכIR4xt Οyz]o{gk;g]Oow7'N?y3 &~^yS_7߆L_oǓ8xqU2=s'1_;g߯yvp#3{u(_;)AC3L_~䆆r;fPͼQE3tikpxV}/D>Mƿt㓰߾kfT|N_vtǧ0{'6jvα՝ oѪ[^0—&~w97'n,g)Ll٬3U+ׯud' (s扒G9_:Ӽa700_ROoaLljTn$!X= ')" Ü? ˯}Bct[+^'^cp_ Dž'~QSfϯel= vGt_#0 ItyBh{ sll]las/~f| .O>9z'íɯGG÷ь =K'~7M23[y;g3i<:9o1rG`h}:FGp_)u|5[WpxT")[qɈp,RZ71F^?%xmҲ-nu [x>x} \KMs]NRXR#ͬ,SZXw$Au\2CQI}z)fT0f, 1a%bH,!ŶZxU [h"ש v [hBZz}zkY~"&m6eS8PPDV֚i՗y2XgJYyH@cMptjkơp uQIy@ TaF(V?)XN?q [XVz}z \"-B􏏟^174P4"*h7!WDʮ}+ze\2X//ӓa*!Al;#n('(V'8Ս0cu[.fJ@1fnjN%N mt*>XƑpQLHe r>8@b%4*Y0uhbwmW78֞ + 8ZP,:.7BFIJ[[ڤ{?&8Ϙ 6cLYb%4nhcԡ2]f轌!6Me[mѻ GR|//:xZ+y>2$$˙U `p\#kʬږY"[|&Zt|u h[|eA*F ;ZhhHTLz2Ikyo8o<~%ҼjURd/תXR,R鬘gLkUD^o VνDR{TF#ƉsBZ6"BEHT>*j(;MW!erLτ\yc2T? 1JMModU|Xd/ `?|:8c;LoQE֣-b [ns*&ĊI&#Z‹AifJ0s~E c*g],SMF,HM2^pf X + OkÚ.)Q) &=Lcp#E:Laê&}n|fr"Z>zCLiɳV*Ԥ4Z^踭3`hK}&OF2.ah >|8b PcN3z^ ǧs륊)3mOkݓW;/n-W݉+$=>% Df (^o% h D0Xu.I/R k{{ػ >^U- shִ֦zUUL '+۰K5 vzsƤy&Aꂁ%F譸VQP A=zy7 ɫO\ݜlk%r:DJjr[r:kXu$9:k8S # ;, % vx{Z3Fib滉vd ; nc櫴K+k-f. E'],boq@.C2(l1=( } 1ULVC L/YQTVc涧m,IIu%J`zS=r.xh8jÑFE :T.Irh@b~ÿ?mK E'iB":UW*gلHƛr/\~CtL5o!yù=i޾k޾k޾yV-ٲקp}Z;V/sUgm'εM a *[ZCۄhpV%#G]G AB Ơ}Q!*Pvل&L@(7wnUۙ;VAt74V{sRvo6ŬZlV ' "AkkPvc:m#'{|K) l!!ss!-! I 2xq[B9Ps 5Ps 5PB d)_vR zai9X@ %bQqNVuͫpGGڋ<@Lp <pZ@MX,l~atq7)$\ұ!9 @H"`@)(pYMzG G@m$nUmM6qR&l(EBhMTdw킈R[a|avnWav~ͥ^9 10$Wb?;" SL1:P+d5g,ߏxLWgϖT*/*KEH``X-LUʚ%gurudȤke] oԃvOZꜦ:$t&֦j )m*E  d"./*IhC; ;$:Pvo|\"e)O_g E zVf"1yA!<@{}pEN#8e&ܷuF!Bˤ(huJ C kH!2K<ӈ>+^`.n|aV$EƨJQ6kQϫZ&3Lxyw ՎeS|tYON UJmȥ (Ղ%,-3O^2@g^/V5?T)E,Gݑ 7h#>5ɴ%GFߦmX7 |w&癙1?:FlDgFi1ڪZ;wKJL6bGƉ<=^ s_W.E'~]}~e6|E 5Sam䪍l1O_vIR>Z_ۮëss:9));e頸ء3:A-'Htsӱ{OUa#/@R[^{,n7E31ڶiT6pryrN:ƽ^V^z9^A8>딷3#E'a"M)D9 %+#,I&"Mt?jkfQg?f .r|c).>+rR]cf:jի<Щ8mF;?~yv=.W,ߨ.e4h,:x|<8g],Sd&#Δsu|?5~Qd|E`ɑqJR^Y5J||lYyl/U1)`mg0_IߌG/nz>~.9G'5puSL6adT!$v=|u R F-,H51f5ߝjÖo ^QAh&3vC^wқJmdmӪ :a=?3Z47e϶7 Y|I]ko9r+| ؖYd~,HAO >ǚe[=9eWjv;3v,o(9kF c4Ԉ?cCKM+EnEmYҬRv7E`Ky`p vhs\ ɖHkee13JYPQh] r+fY1"aH^cLz0&[_VZɯ|jYW-ey_ϯT*eJņdb`|%Ϝr*NᭉR_*>ݣqRU1jU ?zut0~BpVg]3 U)S CFg|tuXG3Ј ׿'p`NW tb/62^ &:ajÅsbڲ+H_ XL$k̥@Շ\xDF+ڞ zH^yo^?.\AuO2LhU%EGSAUD'gNj19نZdxoD*0afiũR0^g'QU/ ܔ/^Y8I)ADYE%f 29ݬ>C*킐U(>db"jXR2Яsj{, 7Ҿ W!'>=Wc*5_'iҨN2r1eȭ;]UAM*bHE9HV *LD x芆Kr˫PAG"m/Sn $DnO\Su6Js|s htx @7EOiT2+/s̑l(XC$rƠ{YSu,jk֋c{,Qr8A[`&/E]`_>5$,XBvr [ s`bRd)E=5.1Ox$[nEZ@z[PO3jˌ "zt_A0EIv@$JhQ# Y9XY E$葐 J6vZҡ$4 U\$CVoR粰[̩͜=Uym4Hg+TJΑ9M_CN߼fWiX=ԌeqGͶvz:|fZ%vq6-\K}@bvu!#1K,g`~Z"8~$$d?7-^J!8;iP1u@bG@֯&/&CSo\\b !r$1qNVC,$Xrt2O^BFcz02`|[!7IĦm ZGb &[&k'=1ywN79N&&l.& \7 PJ),4:=rP,ڜ>,=`ZYIX6$ ;5LG7w ;2F#34m;RaXީPƘ\K+Q`$?k|}>{Kn PIӎ=:-:11J"Qk *r?/=aҒ~'T.-d膖ꖔ=/xV`b_/3`-ꅈZ+ŨNgyI>I^;cOw`oYِ. FOɱɶB]g>r#da%Èsw=#l CN1\`?ihB8@RmO_XX`K>ցcr ( Cp=ѣC̐?СAh~QoR)^ᑝp,:?O(*&[S>n%HFsgaG c#56Z7-uƆE7?PR~N;axC143:$N4!݄O8hs214}H~<%,Дe LQ$;D)BZjebq,S0T[|5:1>$&ddVY"#KHۋBC]SpU,ɞb g]FJ˺XDX`r0š8uQc.m {4{ n景>A󠑲3ai mbί^'2 }e R2s)9VӡʉTrʩ05b0P[ ȝdj,ʵ l.W/y8Yme PKj֟op{x|^IZY5S_ph,eJV\~+vY3=}wg/ryy^ſ_|g:Z/np\>0A=՗7yrڎ9񥷍[Cfw!7lG)|~: God֜{W&1>ϳN*!rJ5-$Ulچ`2aF 3v >(iDÄd{ P+NBTMQ 1",V/K!!HhL v8owҫ@w[ps|o/_9Z-*A3bmҩ! 1]`ۗ?rQ0*:LCRpHZ ;\|Lf#<(y$ c W]ݙE$^YyoƳkaiֿݷd%ȵȃl~LbٹdZJBrK?/՟)Iɍ3 чkiLgoֶ#06X俪P4SsSXB0+`X-19k60RXO:,9n3gHځX;uC,jf_}iGR͊Ӵ wBwŨ;k*Iv!'i2cTf>gWcNNk7ULt^Y7kD!X dl)Ր D<';>FI2BTpא؞P-Q XM*e^:Ľ :b8n*JxX,ŋ)`LuB:T_Q|lpAeƪK&S;V1R&PO^hi{TډMPs뛷ߌ9{nLM}qT\9)T'CS_t!!2<~u1!RDց+%m&J,3ujQH˥2lc+c`ؙFC-,P/ܯT<%P pѼkFI9SٲuE=g֏UA1fNN7,3o}\ZkFp A2!SM~z5- 8G޶˚FpA[ԃm[~w=0PI^KIv,`/ưU#V,m#u?-0R pf@tnmN7xca[Vvw>3˙lI`bZӁq]F3aҬ9Ʌk.y|4Gm,g"T;e#.lt9.A8!_P]C@'"vpuz۳;xvD,Ϧb1yXɽj[И1-a_^$A_}h܄4%ܮN)emr'ٖ4~_,'wg3ٝ ; ;ؐdZYdϑ/S{%ϛ vl+DrZfkOU{?7"-ӌ=CZvhJ IqHٛ$[AtD4kk=CKXbD[VoH.# FDw+[i][8+Ƽ,:ɼ &ٷ7hNwt_,_dS)K#Lړt|Y,j[ce=7ŏ_Nڕ3 lpj)m}pYK)YT L/mrepϤ)Ca{D}wW.BY[1T m-$S- Bv!RlM685&E)S R@Iew]6]BTƮ _?|8[-kb9/Μ ?b2z6dvY?ӻ ӛ T\mY|s wO~j.]˭E-P)vi/~ۦo_~^-$Eϵ Oֶ܄.s򛫖_Z0s'-*ɽv/~)QOV$'&6˶>׀kmb1gPx]_%Yz-V`&L3@;µ7@TT8+>z4>ұ(&dCgY…z7(0Ңڅ;Ď7:c/DvYLy{isFJ NSJ(N(*7 s. ;HXR׶>&Qϫ|YP[KGd;󿚧-.nX%wES.\gCI3JJ<x⁕x{jߥ4^i7^^T2qBkcG㈋8i8go]zj`ypTcqt㝔/ay=pґ07ϱ.b٠ѧ;;Em٠сº3DIOfឳ:|5LamN5ߌ(=ښDb=`-}N6B~ٝC䘻Np:Bօ :I<]JJIsԺkuuf60=ºM31f2OIW{z鵳Ξ^>ZpxHSk@YhjRƈ.҂I j2}Okߣ|\,ח[Y6/z?0.imgj֔^4]Ny㗟=`4\x,ge!we5ٓ8횪FW(Sk#:eiP CYQ m\[fwgn@DLF<5ӰfH5 |e0IQ.cIq> \2h`2P0C5.gak PZ9e<&(ؘqbjZ?f\~[J4{)s4)@rIdKaqPTWXUcdiLsP| ťdYa b^+!{5 90"*B|,cE F!hsa?u9]C`*ɌJ%.&)hLi`gv*4&ar-Y=[}/i*d^s@$(Ai%B)t+;!U 725ZGd.]8>2,AKGo'<"53 ԁp)&09@Et5b jL1->r;XEo8fnv8+ݮӦP8ЩM87md2l.STɢk½251ffJ6a=N-tHV7;jPnq ifئ jP2 'v.ѣs˵±Lɜ>Zm`\@GyPRǣOD[=asCQs!cr}A㞾mE9>6hی}Cn~\%Rj>9f@Djhݓϧݧ'1͌"RD I3(T^⨠lEu` ͘8C0(z`r?yt06(u?6cFF|*3M9a}rkcytN DP{ydZ|ʐ]+C.!aHJr 4.sT\QR\ 0{%ϳwf0}`Ł/'~{TJqjAq\A0GJ-^^` 9):Yj$J`Hx5TboqUЊQn~z]_RCڬo30eh$ԳcU' =t.K#Aڞ@E Mꌞ]Kgxq<'R(P _3:^5WD ]&XOTeTW@t+NϴKpЩRasw;ׁ`@{EFӐg2]xV?'ăsv~ GyFqKyI-N6@YN߻vÚC) VV$ݍW8FJ9D`v/wUBG uJ,hwqJ:`(a7 wzz0aWRW|sRnz!r`[猱zY^=A )ַ"P^MRmw()PQNLJ)׀Q ɑ`9L1!1B* n 9ehCY/W͞>5-2ožRT.*dd¿ܤ34D&)wAƻz4H%V-Btza V({87㟟V0K%e> p R!\[{'Yh;)g J80T SŊ/^߆X*cNPDdʬ+yt#Ehǿ:fgOZR(J/W{b&^zT_r*7I@H)Jy")-jL {FRWMvt\'MP{U(_/_/^ 08ڕꩀh*gtM+'EcCe$Gtνr0-aGw4V;[2cK/t+XW pbWN(Dj2dnd@+TQ\LAq5WQ3}@j;7l}(sYq2GP UK `lQ D7Z|`nn~䆮\[X#U93RaЗl!hY;+][W!m-y j((dΫniQyE!)+/;/0(}O{sW:ʾ-fo@׍+la;\ RsTJ%Uq+#[K.LaOaBn9~qçӻIIN:o VT'f6;PZs5>MWi;x_re?w5>~S\}BHJP>޹OZ04 =dINzæZ%SrDbf݂zl=Qn8-}l${>HivL[cTc`T>3(` 1".-߭]ƈMr0++T %v%T/Iyꖠ%h?dKফ[?ߞ;ѭGK5:{IQ 0j}V?x~rD)^B- ahwRo"M'O6avt҆K{.}=s2ZE!IJI$7H4I;*yy8jN =ÒOqy^63X] #ֺ/װ{j_=ƲEɬ2;oO$ZKO'-aS{;(j=wV23Ƴ7L3:|mX}”h)qj<{!XWC8 v+{<C0N5a+(ϵW/¸ l"cZÍcLKqMF_di.ts~,xy_TŁzW^8W,\2KxV(up FGv(,Vr Ż)݇ݣj~;s2p-ԍ$ъ>/%nںF&iL[6![ZB޵ms" x t]Kdu$Vn!%۲rl9N*%"sxDF>ZU@ۤ dĶ*esN}QB ҳ蕝5_0;?.m"I8i:@MU}J_X6w,msChyA';^I y mȸBd?mT[o{w =,6qj{'fff0\R뾚_"־mlkWRvr4@kLNcsw/41G.Řp49>VP-BzFKOb ;^{n=~ݔ#Ge9ʂ9+Ѯ6Rp.A«Z4~={wYoe 䢝yye!qiEqaufwb"pJ6l=lQ*R4n4oM3\lQpJzvM^5~kJ@ jRu~K釂;6vQ`s*[:0* \Vsz@._Ö*|yn׃cžkbC|7&}^/(#WJVNj=G%%ܻຓ{S~YHYPxAGNf>[* X/76GUNUxxKс ɜG+v"qm ky): qko-ķې ݑF;6B3wnW tf +\C3ezIC.6̄ %ϝqsS_98dyUD DF?üǍ)hJbtS.H{6ttAyUbu8G>arRizB}OVؗ:^PAYapl<ԏo}צ e%0{OcYoĔ۩ɰցC%4QhIW]vHhz2 ֋}=I~ 0֋35 }p{#bW[vh5Z*+>Qv1y觭Nƥgx .X?L*UZeqn!޺¸o}T? G&ĴV7Yr@o,k?Fy? ½qVQj``^es>L8iR:GH)Uy mUFIZ NlxFmE}nt~bkf͗k2qN$=?0gu'] ȸp/峔tMtɢ+$isݎNM3n\.re&6<VA͒jR2#k.Ф{hZW-b5,Ρ^ #xE(;`^[EaݞfhtK4vkěteМ>:LG0ּdH _O t2:K2tD*;مJah75< /pzMߐ͸$n~탨*STЄkjKy $峛#)4r Oox4JW'Iotghvk7:㤓~2kh]SQ4U>G^dC sGғOYY&x@()> R.i_χuSşL?'N}tn=[EWSK~~+>I;.4wI[Y3R9ն1Yu:]r |* }eUM$ <:欓F׶6AA>AƱ(6O = {H:f갲^ ~`V3ڙ,Gmۛ#)sۀs?aBM>0ksd"΁HSy(@CEuݻS<ӴŁ&t1 Zb;|",;5L[AkQ7͢m[&լnkiU,ٳ uzk;8z~}}s֦=Eu;g?_NON._ӛחo.^sw3R헧g^qybf|v/qvz|.e}*ՑoU~qj^"4*3kZT,۟ՓҴ<QR—9BqG$&#SLcD,&8)bTgau?8t"QL(GcG!>-PhߘaP6L_* f, k1LUS!"Lw=?e+!܂Z֤0Opݜ01*I^]H7(mb0l '- a f7cK9sWjݵ_g3˧ɍ2|qqZ&$D-v'GL\72SLXzze!s:T:@rGpt$qMx41!s:Y,)1!\1R ǡ19*Pd1[CsByAkǜrT!n>ci.o\i M7~O^z~G6Wh}ن_2 fYlݻxA`|B*3s$U@x+v&8[5SeɖɇË當nt a! ^J)c htDqHWá20}0) )C,3^"c @kpi6!As0C6|+zr>U25$9e?4< <ϧ4ӈ\1.|GUmFz;ȏd> Kp9P0>" QSHf}pfg9ԜRƑ04F@<)QTVJ8T<-Y_`a 9pаІ6,t,0 $3 qdb+P*/`XS>pOlΙgKbA |",}󪜮6Wz /kSxk ܬnK]j %8 fIˢ%mQbl+mJXW8ȿH5@DAE95_n^\E175TcH!ƟicH5TcHmܐb\J*enɎ^FМ+{|WG$@>; =џWQ<^EtJl* ,TQFY^Ag ]ͱ5?뷳㣋W{VN,!ÿfp~R_tGp0ߤxs qDeءoIIR@Ɇ+cd2i iQMNfKA{.xKyRRB<"8%CnEku+kІ#ڥ|l zzm][1~r4՟O`u=U ԽɎ#ot#\[kvlAZ03G)ifPc!=# oc)f3.NX(_du<Г>VnSC4;F fZ߻}B3 ==+:ܫr]5|%f4]idF?+nDw0}WQ tSna0ݐy0;]K%OyÜM!w.gu>ڇbk}(vهRwuXGcnv̀W.0\ʄCu` a8"Q (BrpaЅ>`,Hޚ t=6.kMXEjhUFgtT(NjuwZE;fgz]JuqxV;ZewZtGK-!& cBV=Y%{;)ʶeb~(||cTA65—mFiλ^?28^NNW?c7&~hϧ:=MQI_0nFnoy:.7@^:/էi頋Wȼ9`Q.93Dm2EN- `sѵD,EBK)9*ѽ*&@'0]:>h*h etB۩8vP?j>٢ YF3[-bfhI؊0DuhH#[EiaEa'NA$^#V$o㣍3hGǔx"݆gbI \gv8H4$$>$wD*#1:@>As S(Ɓ92)Tw P+ da!gD81hUjOQ1!TXӚXN,794@*.8^Sڌgm@C&M!4sF*fmACՐi3oC1L 0r[vRKEČ+Na^5*Gh PM*K7\D6ƉػpT.-Zm|@9c8ZcYPƵ؃cy;mU9h0RY؋xsKѐ Ѧr<%э@1uPdѪí`t Bf~Yn'$mPl]nc5ƌ @+?pBf<4BѕH@ɓ`Uу,6#u@t:i:/@v\dTdy1P>#- U-*15&:b.b;ѹb۫G*ڏגXfd-EabhD+"Fh@#MQHC`Fi;G&"rlLA VܺOYIBrB3:w6d$X9EX88W|x͓Uu@IV,3h<-HĢc!q,$`yLY>Ndmp22 +@V"$zMYӗ+1\( &O$68 K71F" x]3+؟J!c-cs2Rl)'?޽ky,.BGN4jRFfe),¹4knV/AbnC;;0`6k}xY7H>:>B(WۜdcL9jEc YK͈tv 2(Z8*nR z`RQ縻Xau1-IOݸ=V-k5gkiXU>SƎ=Kdd(LBǒ{E7|:̚1g{)Th4-ݾPy[!O&5=}p2p<-$B/-*9pi{(MO^ +djuM^TkIW>׾bv (`5t!zL4JZ[4%kPЍ;{63g 1_a-g̠wA1w6p^=}k )5mP;0x9o̫Vqh(Iatzk2ˈM`{;8]>.SjYz օ-jZ2c8omPqE)Y/pa)N?_5:xREŎ$}~oTco&ҍ١ [8;ІNr|P|Kaŵ+@\9,A>+Zw[, n)Ӿ?hׂ,xu s5P/')\&UwH ߡ9-HW ڡ6%Z}[qs*ŪqwНo-vwU 2m"Tb; ԧ~)^Ec!_j+uC9t֐=rIJ+%+A[f5EV?U ؗ Ɍ;FWę疂pzi/]'f"`+dnXBGI&CfF\ץ cȱLzx=U39En'#? <z~l2tŻ}4Ogw{~wv׻ƧyAUl_: fRJ5߽u6GQ闣tkv[BcZW>%5䊘"YSHF0B R@@ph0\{$RVQxaSXHT(dBd##cR WAIGl >i$JE X(T{4]6z-~O$L#6RÈc/lw]SlmG %e}qήk?탲}\W{}|~C7/cuxyzl3]3P+7%`ӵA[+i{h;{P6daFs oմӜ6JoTS]f5K@)Ƹ67jA&%x>r[^}R\ۜly#uXoG}Jhl4+9y+' h) $MS+iȅa++t׸|\nľq*bBQ]gO 'r݄%cr"WP} o,CU1ڝfD7?y@/zFR/1Q8H D^_4iˢ<w%y;#WR!hX@}慯>BGJDm w+_8s{Ҿ|,ܩߵf;&{DVo#i 8Qߡ͕3Uadj l77].%G˫c܉G3Ħb _ohfO];^P]:񲃰*fWg _\gvOYyQK諨O9G9[WKD(y=] (|ԪsͳLbdeN^TѤx4CIkGU2!I"Hbr*8AF TuttmE^cadT)5RY%adA&8SEQShH T"OV4F&)  AI I[ٺIvN}=zAĚd"KiP)wpLpp:E;g؂I޳WƵ=T5@R-{*J?(rJITb$wDosQs99G 3()WrpB$(#Zi tA/ 4ܥ蕜.AOGkQRTˍ: tIS8D…&r/KQvAOs6}, N+ܴ\~+L`4{PQ̏3nj{Lav2v,z~(Bܖouŧ?= wwhᆛo1ȗ~;8C^V~ gG_LZ cs{3ܑ.gv8B/~XT+LzȋbwkeQh3+Rv|SzAhC q<[5 mTsCFw~hVC%%!U#I`X-$%ch2hJMr<3Dx錌v}!ՆZ\DW/"A2'Fhƣm4vXƦ$A${t853AuUL~"5p愜E$5>,D qG rn`-"m>=aFȢ Zof6t_=j޵] [VFD.;&[Aku:I&Ƥk~ndb*BʼnPЎWx~pՎ5 v_8viQ5TDDowL(IRR|iJk.MAggU+z閁5k &XNIL {97RHPuB%y%@iBn zlPF$9rzH̃oFHMJ\ W)F9=͜YK\85 ^K::(FMO/d,=\$xF'MQ<%[%Q]&"Pݗ.ﬔVӇJ4l54lMobmn7z5BV<)xi'Xӯ[M8s2#!1BbFyTS )HZ{&(`ui^մEhۧi[#V3з lA ̖dQTyfS%@@v'iPpw-yîpꊀfYe @W> mЛ<5bkh`9G )+h^# R o,:ED36\}bOPpPxgiƃ*]UV񂆂F$Եi Әe >ipUJ*Ӄs;]U)5+bΟ\ncxre*#x2 kDiya۠>G1䗃>Q _{CHy@ n90w^i}T])Ľs:FXԊww1jGi4eO7Mo~/OiM)W«@"QMBlB# F[76>@H-.w4ݧrY)=y6E%,bElBBWW@Qwu*zMz>4kP@zd {9f0 {m8H 3;kzmkeݵڜ:wf:35ߪC w]'æd)j| [<k pSFMc>[ 5kg1V_zh5_Bpǒ [Eb%,pG+edA9&$jrWD{fI !bE+SQ\$5>]XcMO/HY4.zH7h&9fN.fuހHZ#I'GȉiՁ=$aD kXQ!1.&l?Hù Y*zEޔJ!8':W'sz buA)- g0]}ektp;A 3Ww_nߣyOڠO!)Bپ;W+i8_'$f@{>(͕K7`TR)FN ޴S/5#fTT4~9 cjuiJȑ7k&+HdU՛9͕Ę!H\!c[' IqH LHVQMu2"D-\Q|G _|;hBh Gf$4OZSq6ZNcq?ΓFAf,ex<儞?N qg?]Wly>?6 y샶k,/q2dVi3e>i9)pСsBgc6 m"sϜ(5RΧTv!8)U;ymSջQA ]t;i`ҭFK[h9OSZ׺צQ _^4~̡ovoyޑ,06/ Ϋʑ+'ӘБŸ^>׹u{]xrcX ܛFRWº2IT!sb2F:ť3>)Ǚ4sьGVGd]h!ok1u-rzy6^pC35j6?{䶭J/9g)~qռMmmN$))=J4Iػ R."GIj<(uht/ng3?ˡ |Mod "#3~j38^|{?,?['x2%J~K$n4qWڪmDv*݆JdJdk+"eP4VHG+_R8v x> "| ܸ)g™K`%Ѣ!P~ֺdT)CRzG$81Ip9"RbՇu 嫭c+7FFU9C~*jkj4axm{ ,Vjf"ŝ( R;99{+x]< wX2Z ct<˄RQyhIV:\<ŘRX;RJ(jFP$XƔRMw1䑦C#8b!Yj0Rb?D7' RǒLs="IGv&,NasA PY`P`k(ȪM$8 M&bf"</6w|>QE̱nHN""\}~y<m>yODDHRvw~m |ʇ ӛ>|Bo⿖yXq2;h˹"C1Rww~|&>~gp hTgVXbz%)qfaDӺ(=sАȦ(HQ"D`nQH@NЖY^,=N-5ka-]ש,IvF*E%4O.[4Qpє܀i#Ɲ2"Fj 2yg+Z~1Z|a-QUJ ^5u ^$Fr Pk!33ڞ2Y .>RRB`g" $ _y)psrTuGhC^ےtx?[V/v0y v4)Ă;ޫwtv~\xОep:ʒ{ՊeZ}zs";˸ݐ[W$A jWol ]a]/, m' 6LmbJUI>-\NU:!iYv$9R=umDr.Xߖ_ә^ u\cZdfLBF] Ԅ'QSaiK~W9+fo;6!Z_e1A/B G|(=m0> S :jO_{wozۋYWikjDvsƌc/v[J!Wr6j{J筈,p~N?#kvvCj *z1iMQZKr 'y1 qv+ @B&2ۿ|mTd.6 seqUJu٩FfBG<@JGsgQݴEj62fRh.UBϒ͑5.]^Pv-5.P//KMN}PIhŝ_4sϡm=V9D';aI0$1KaFlFHvyT +*O/a:pѽwUV.^Du;X9*ݸA?'wvEy?M[zpy3'UN z=lAhR|\C?z@?#<و,jϾbagB$隅mC|n2^*msv7hO}y&kؠ6H3d&meK 6G sSH:!${BBTKv|wo4 *3jy ]T-@r$ѷREJY Zonx\kv:)XE-"FCxjWܟ,[SryƝZzh5@T]ང-ߏ:m~SHIq6saYL lf MG@ [Ř0La,G6 sXcᥱ a+q$O6 i=ho5Xgghre]PDcսniP%F*&RY*J3!u&΅ւ+E P"YݒrX c`YԒ z7I\zT8og.`OGY)쪚82̂1zA2KBh3 :UgwZ&0owlE`_luzN# ʹ'O 5T?D#c9ƚ3>#;TgZjO<a+'*ɒk `Oaۉg:u2HRfcLEVyit1`EDeک4 P%q9 kY]yf<~&5HVZ :ccm)I`~) X@:0pʆu"=^!ʲ&$AWϮQimmq0#!`lRH8xew NםV(;DU,YQ,-y-q~<'6 Y$l C}[@4 |ރӡR4^WGK6޽铊w\1J_I h~U%XA3F B>;%QhW)9C9gt& Aվ_zQB=# ]Sq9/KF=|,P飯duG˷O;V4GQNupv&tR56c Z*bs s*s(v7ǘgޑ ?Pf&[z8Fu'xvM&G8VH"7хPͤSiεF;YƄ)R O b0 <[Cy@om7x@ ->O} 79J3H}A2!ԧZӂMզ>霯R}-g|w>dNDjX1@)4 ׳B@B9 ̀Mh>}cS `ËQ!^!`nWKĚ|_r,BIiR=تe-K6Dd% ,Α~1pbə1Xs0/(#U\%ʫh>߿â/m%~ +Z'%U0,HK_Y:mK>ޛ |92 ڟO 3?'(WSƫ4d x\ gU}` o?;cO6ɄvʮN f5s^3F:TvH6YWJ,Ҩ!N4m dQ{s]+=Q2cԙMHgSy}-src !8C'K8Jy͍v6P6yYJKc*ggc@/|[m8 ane}ƃN lp{!oz,t JfZ't)@/-] sC"nё9ќqNk+:`LK&h>#b=PB(L&`S&u#H )) %;XgpDQ?yLft;4PduuQn1g ukwL:Yδ4K9*DaHd]uM}94,A`1QQr4|mT|n5VV>DǍ-G‡H*ͻ5&+6֠sj/kL `II >Y^o'd;l<l-hp_=ԟV_eJ$/EU*6+"KH=^+0;.=WD_=|Dzo_RV[rBY]B qM N.;Rك܋5T]ƌn'tf#R3#g-?ζm llO^{'Bi=۰yZH6)3g/hU)v/C:gmEktr_8&9guL\jl+2])ݽVs$MuY$e4!Ѽ?}ne~˚!maVõ^ ] "ۚxҸv^SU _Ws};]ӾO'w5\e\Ikt,o1˴GrF_ѷVW>juMFx@[0=krcpBZ?퀫j(aJdIpOUS_l#mJՖt aA^}*N7 4[e/p[;;} Q3x[[֡}۱elu5>rt֧K$eтhX ĵ j,8õEae,^YD14}l[yK;/:^p_XBkWtgoD;Jޫ VG|jBx{8ަYO@OoKշ qySYVlCpO(Z-NH5; hHCRQ'袱ԱT*bAj6 I]0luE4JSM_ri<5(R*9AK@jz-cDi{r Fz,No{$'// 9+Vƻ..N aBfn$B:;1NiϤ`0@8#Tf^䘕W~4~\ +ݟŹi>PV}iřÝuVِm "\y\Vˬ4BON=(%aAîS.d̾z PcLW} hV#nz@c;Xp)_f/3?jqh[>]A%h;3B[qԐ+EY̆TcV)0^dZY8zq_'8\Y8#넕B^״1MG->F {<1Y1lV;@J?.8Xǣ}W7F'm O!7#ov`f!_sCG<آc *bș=1`6`zxksUF8' h~a Zoeft 6ĭ3;7_?>5"b";RJ] +`RdɒoC7]}2쭧n{O שN1.XƬ9e٢ ,FI0p,g9<#V&a3cjg /)]øhd<1ƍ%I9`pZˌ#96\Ӈzodh˸Đ2BgH}R0[@ƘΈ _{Hy j Ldv|fCR5J@e>[,;'A3gD@&[ CǨ(Apg=[+qtk ȑ40\ipVEY#)\g*0'/0)y1hr `X |( |z'$5qK C#?}99SJۧaoL5#~?mo׉nz} g\O_ռq'F.~ۇ7~WW7o T =7Wnh/mg륆ٶ:~pj wάK3ڠZ\;% Cz׈SJBڐBkݨZ_ҶwY/1FWݽgV/eֆ3 M dμSjJ>Y;]SгVJ+cu Q )"T{.j)BȪâOb~JČ[ (E[6}(޿yݿo~[Nƪ%.x\˄L<EBIgf&b$8'NB #d􊶤l`0NAF/ϥ6wK3$5FaE"[EAdus''acVIdP%GxϚBjD*fҒ$#vL]a[(p9+?]׻}.`4ZZ#9u=FXŀ#7>Fx% =^a58bX Cuxުy7̡vDZ>x5VcFy8[g,ky ]\d;X$&:A2$#|«ӘO+80`^Z ]r\܍E(ZBfi$xK p~ۦ! Jrڡ+snm)}tnHjO/w ]һR/]"!O? 7]zsM _46jkwS9:smK>^M hKXrLR VK)N4;+Ex_'2qLNXCl,獍&4?g+=]ƻ+9Q2n[5dR%6w<xxLSG8|":_|(A/3>~_E&L'cǍ'w/ۗ2M5"̤ژ,s|XxcQfa~@,rq"X¢`B>wp-Pe%dnOU!Ȩa`7][φX5+X9[]x0gk?F[elONӶt6v R#--k&agꯞ'4~ۜ{F|⠕FO~>[-]+ig+%V({f|ʖ jDq]nKYk4E1_i Kt3Q9}*-7U^Df>mr DKLEk[ՅJI6qJ\r9/Oq t(NStWy*ro.߿{ٻ<tvuf)A?п}*>{JO{+bj7go: $.yw1PB+|U*k-#R[zB%=eٌV)UPy[[zJٽfDy,jQpˊF< -Oixuq{"a*=io#ǕЗɶT1|rƋf׆|I 9GIj&=ԼjV7#caջ;QV3Osԑ`47R s[U͈:<缡q{|[}K\&cڌ,Mūo2qmV:5nYްpQ {";Oϻ4]"\QV\)ZhJ7sݳ{': ^# R.񐨸jI6qͣ{MB"v^[+V@q fjc6bjt 'jIN2cRGQ$<UԒrk%+ZLr,z&_7FtׇMB 05V>tEB5e+06i]TZޢz+,JY i#Cs~*odQ8f4'tQ˒; [hRORb\ )!*e4Z18Qd[A.!Yh{]DRl+4Ug,2mX E~y4 <_m9&b0<YxH~*k2P9ۊsҺmg`tQXvtd(%oZa^`2fGoί5'CՆ!G"([Sx`gH@ ߤ$n<#5R)۵H5oGRg9Lt6+s눓I B%jd4tc`j2K1I_Us/+ظ0e[>qZ^~1-Z>)֍J>o@3M[5ig[f ͽd$ף̶ּJ sh.!Wox \+gS.g==/XY疳誜EYJ"d#); xxF O F X&y69M#aX(>0⭛,ͳ{': %ˤ%;GHEHu燡,|:q,z(kek;lM܎=zEc>!,/;D+>`Jۧ"b!LN: ܻ.aDEd*q0gIk}Ԍ^2D'#'iL҂@)`JgV(yq{dl[b$^:%>&ԭiگQ0Z}d9,@Yp%Ij捸VbV FJnQMk4xsWNdOpw}ۡqCΧeC]c$J=w/}/%CIWڋ @#6xf 5E |s+ )(oGQ:gYOxI2~pqA??-X hYJ)?88ePKWw>L 7FMRWj?ip*3B._O΋]kr6+%1LDC8ɫڍ6nJq:kn5lnx2[Y$b@m[@rtO?pk*y9Nj!p~wLLJ>Emf^?l7bEyV\b2+sIsGrkV;;4U@ fƅL"G)ׂK`G#<('"6zVr?cʶY+Ɣ1enO@|2F< '+=QĤ׼3;pN%%u4rWl%<$R.`!]rw௒vpRNh)> &@͎*x7衝X3+-דR8V@TJ>׭vSV ELu-xyCiNa[Ma:Dbi.hN„`^WQrk AXyZ%Шy ?>;.7MSka#ZKՑflA6S{] R,pRgpmI-f':%(Ň3PZsTx B *d0h~X"6G98cר՟# c˗E[!lûiuiKw6Q@fa)ǠcU6*: y嵎avH&GI9+@|ȓRdVT7WӇ|eц5V ;8x0N2ϔђJͻ#ZxxDQՏ|LE#.sw+C5c5`]Vmݑ߅GpU1f"j 䠿Q5$"#k+4݌vra!R~h 0;LʵLm )?N"y}7 0Ku[0wat&4aKFwEԦf'ڿsHY4T.8&NI.^#mnFS0Z|K HFrr'֕}ɍ/]Xo謁4 ?{B@S_Cki$@25ƻl+e-6J""~%r&7Y!Pl)d*gj<9[K>:X>}r oHGQE1^AjQ{X$WT]In&L~, jiS}l-#GsHd#{P>87=לcSBs,[ @|imE=עSE>V :jARj7jqՃHC,hlbda8l`BFj`_ wbT x s vF (1'*0үTlnr e,ӆs0mpy1S FYC+HC bQv "DzP4iQ0I09ͩb@EcJ8\NNZʃ:iBm xxmO."tMzM6ʨ$ejM' Kc{ 3A!?ӦI%B9dM}^o0SQ 3A& 98?iC+wޡQyi'瓛k8e ~~JL))2@\6S oz_3b=g4h (Tsf4Gw,hgEh#'sMwwiu=zQBBF1 !Ia>.{z3\wiwgi@j@,\&2y#lzM P$!+#+#*<3YUD[;ع,]PK.3Kۆ^ւ}g+trP.e BhF@9y阊ZE96HX%QK1Kس# \P.;+iqEWU Zp6EczXH)dpămCAxF[ UShw#7*3#o c t[N_Z-Mcͬ,";7\%bF!C!ޔ:RYR QUtGĘsH!X QhG%oK6XE܅gi+G^YIAeg_JP5ēs%pT=șVbfQA+% K"WTE\ +&.'1q#£tR`0 6S9IbAQ`_ Jy W*R0gWJ;eCJ!m - kFPf )aOӞ;vC!JkPOPH& Z83i?>\ʃ;r:6ͬiJN3 RX|?8]xFwߓϟ YY;᣿^ߦ "BjNF 3L$n >U5B]3~?O$9~z`K$` ZݾS*۝m*] lQsnr4R"5 7ߎ2JP=khz1Z %"As>g7B|7<u!~㚗b|FFe$2Y̛32G%|"Qo *8!##GHy% G-.KݱFDqrI&)$5%# 211l"% y N AEVu8Jrnrm!BK8I`:᤬"'$A8@"t]o4JSt7YEQ"3(bp4q"VʑE: F Q1G]db ."#AC#<"ň aJx'UK)3Ž'g;lgʴ^l}q7@kb֐¾E=yh_׀6{%:zPEȬ$}VS :q;ᢏNxf;1u h[-zzfȾP+ (-&b|ʟ$t,p;iY:$m#}ּ% r#Ť=N;N }wyuTE?rydlEw>?B_ϟ&?/I{<}VO"QȂeMYLBM!\!Mlc%'Rhp),FlMY@B{M!=)l1[Sh鹣~SHiUȆbMY܎B{MDM꧃bct[FcE͘j1`p6wz3}:Y-F"ݍYcFO?BQY8F;8>@P-FA~ocFOTQN)6oٵ|G!%OSi{,?>Qտځ.4'hWp.A Na^!!CHTxD3JZBACñsp)xGSpp~u{?轂CAqҁWHLx!io 6ߣ2 % -MmIû͟?+&sߕ@:I=Y?#ZPq}c!UOi:MRIeihfVRK`.,Kcۚʜlj7w>-4STd|㦀Eqt> 5H$%Xb1ʉ;G9sjfFO{x(df"Lf iFrqY@5iDzL) QTd@%wُۡ4L$hs%yHXUK\vA1 BsA)%C*}!Nƥ֪FTl=}JYA 0r8uήcźm1l1ytZu˰:? 7Eu~v:U{R6*{LַGH5O^4L@(W4-aN/TF=~,;4R*꿛t+_*"FO|A1f(H%|$?1+H|f=>-ShXPt}/2FF[ہ%Kb:Fw>* 2wW>sWx?חT ,poq7 9@ܿ1^ӯ^tΫaS/V W^t Fm7IC~ ]ZS]$V_leV/U'L-ykIspv OD.Tϊj*Wp[a^· op0*!5jɃ #0i^P&BcI\4 C,L$e"I8yZ*U l&3={]K-ZXIMcԄTƼ֗ cL SPpPVN2HD")N$8`o!n̥#ѡP~I)Z^'̫\`ys3=/l Xw/a:Iv~(n?}<7gPxz $"hwb搊{>1_gc`ޓ(S]VyTd.D[ ~S=~jѷ–z$?iq4}q:dQ$Wvէo/Ґq*OWTӯ=հgqF W* Fnxb݂8c=y482cq[ I}}>- ԁa=di1ÄGzW.hRCX9+*cRd4ru xJDp"01`(8,;C18QZY@@8-Z\QhRS}tj=V}KC=(.сULIATp}Y~<1 ñPHyi@ ؛[?k,˿9Hz&;JIP/7$^廆\m͚2/c϶Y}Ʌ?\͊ji>SBCH;ȬMsE "ƝR69-%ThAPaVfU6lGlu:S|n ,K+׌B.7lvt: 2+P.ҖNTAs;ԴcHr.c 3Lu3nz3jO˿t0%,|/Lj=N/ cw#~t7:[pbaZu֣ϖ.nf xOcO{?26 #~a'OOZ}8"lc>&n؜,y`&}s[)AeM_^oo]|=%.z(}J6WNSZ|g~M[ nZfh!'NZD8q4$ o3;RRȰ;)=l OT֓Pt/W~ZoN.L]%'t 2"OҜ,"PrS^5j6-RHIBɈ9s\ZT|0AT( $Qe9`gPJЙէ76R%G^t"En NS1]P(T# Y2hJ8$iڰ!uцJ Hao3_+==ݣ )`g2W lYs[|[nNu #FNj NBW6VEvCREuR#5e)r/ M,b)#pH'LYSZ0R ǏQ)ssoǣmwj;!vyyT`-_\T@xSZ'/_;ϳ$%%^ܹy_􅳯Ku& PxՍS9sZ)ϒhc /CK@y^>Wd9DI(iЍe&eCy|t c>׋̅DRr߫@3DvNvbCTc(H\FI4M3erUE{ VnHKe|F3EOuT3僣B\liZ ս€`I1I&pQ?Dfnj,.yHbJB\k-8`l0+Dk/<^,"MXЈL܉|oU0eW6o/55c<43.ʶ+9Nɤ`Y,=fCL ?Z]N!MXJ [SJ7b}h&g_Q6Y>,4b8"\i1)Df$ھu[[ `ʭg]ު1pb 6,<(G`Eva#Zj1X8 ؾio2]P4CERk59Qmmm q~Q\&"JK"EGiB$S珶9addW]ۈj]{836ˈ k0pЃӵK1 8*lxks4j$a-X=-$E=QR{ H_[h">ƚRQ #G X7A7fekkGF%FYV"aqL%0a(`ʉ-D?ly] Z`aEd&~:,EXX~\χc -jLs!ϋi|K'/i 1^KNR#ȩj{)9?ҽou8[=y)rP%{ o}1Ra@@NREvz Xq&#F⌋<Ÿ0$u:a߸XȝjZRC:Y7:2lafΈsEO[Ǘs +;Ks< `7ܧv4cٝq9#i& |cDP17؀ 1؂g%x$JH,X$|[L8qAK-8Kă_7hѡdg~IOܳ$@#ZՁ'@: }ЮC_n@ч81 K$Mo/%&įPsz=s%Ħ\m*h=@Liic)w[*ckىM{4ٳ =<`NL84 2o5JR/ohkȈ\)feB ط1öt /m)ZC"Ӧ1~h;yn/Ț*䖣},YTMZy=p xOk}Q:ݴM%H" \e?SG :vv>!HyzX-s%Ř >Ln0 ew{*f %2Qʞj#E$(,9qV}cqj4y}"4l;Y9E~T04 zl )y7fC@i2-(7n R5hkeNrm*%Ҿܮs^m~ 0R"2գu>GW j3\0B9 x"řLYc"E $@va<6L>d2\ U0ii g"^-zŐ((`9az8_!roȽ/XNܜ#f" I o5i3.#tWյuuu]{X19pHS>QѨ9+(y7ƅ>eπUꐆ^gv%ϊRjJx4~> 7lb&:z{TH~tX^ /JU`MwlT`B?#ͥR垮! 択q'JbO{"'vHFr9@\|:GZר-Nf%VϹJ s":_8;O+*ЊXjlw~sLCh~;@Rt#bYhpG~rQ0B =p#͚9P PPp2ܨ3vPZP,[)}Nϐ投A:6 F[:$sZV,|>Y#sb}Wݗ]uA f/վ ! Hn< 8|r22+ZG<2*1)ii!G6` 1!ci”F-sXKYNZ9㊴ $dNu׽^n3 k#w *sњN%][mG++5̹"꛱!QbExkgGY3}{Gf0Ad4|]n}}DcA ~$rmZGзD2 7GSMzw=?B/+t7fnX:_̺ϞG3q>߯_x:SE4Q1pcax"O yQKE]TKE]rQ nc-XzNEl@BN2Xx(G.뤉#Qt~Iy;%ehV6u±n8t}( ?ycMfy Z ޚ Z>&PV( x!(P؈RyVt2HycF1r(Ւʬ0kMnX[lwQ` u[QLaUjp|Ŵ=ݳ^=ݳ^y5C%!9<[0< 3B8ʊ)w,Ŕ;+K#Cp7M4h 7_ܯa60[\^]v ! P3H3:Y8Tz A 2`BDE GjA[xipF !DVadP3l'>\jP<x-J뗫~YdcOruf9J*Ɩx߆i+ ҋ1qmqS҃Ң&T M-0-e5 nI)H28E#+$X@A[erpà`yZb9Ncdت%\UC,(8UDHiXAҨ7#xA=Z)W)XA/5bRDȃ#`&JO hj2#.H1(P)AG@ 1˥; 1\ qqL/#׎pE@uQփ'6,々D9ض%r%c`5 V;4L6cq"8 8ҶR`z{mYp1ncrŜ: 9 ? Xa8EDZcIIc6Xp*/HH K$O5FG0@WaЅ`K h,MKI> kJ}fMa4r,Z G (mL7ꎭeWUvy fIېdD ւo{#d]@b0#,?-XSjYavoAWR7,-k D3CEr8 ] X21RR`tWZܧL1$`0tgutj1:A}%#0o5y}xxW0MAonDDH1zu_镝~|K0O|xsw}§ rk3|= U3C`'hA0~{r_k}KCUw; mz'Š7 \&%\ DmJUIxؔC"5j:mъ(¨cNN䆔pREP eEpvWT|KWT|+_!ęx@exfo~rmrkjgN6O}l`PHLJ M =5&.jFTe'х3wY҈SOoMq<)SB㟜^kjl\ݑJ3\OУS"mo%/~ ft58U&a4mp3mVIs+G]/.:tl#]T_vfo6wdG⚷V.>?8?Ƨ6'I)g{3Sc JSC~T7k jV:Mq~dT^Y*1+G߬q ݣ)'[7{&iũ-xa+V;8V?(7Ƽ7>0-;gU=W;wD'?HE#:8}XHNz8 ~fvүYr(Q=nu;o3=y[cgn8mN[CfTjY >v8re \f*8cf`ʎERT.HTr1?*I+pU$XUzXUzbC°VĈbx앧QDTF @xGG/⩛!To;;PޡWU`(/%.N5S!,&5utQG@[:FBY:jʻ{SrȂwG̃Qa$`8)-p:d`n10y\fqwq%]҆#P1O3.%pQ q9uB{sr}֭02"{,K@c)1wԵ#qE$1C]9#KInge7mlnU{٭e͔IBG; u=%Jҭ-4{z|ljħTIݶ4|*= Ky+HjD5pRFa@윐k]zu<߫`b2ݾMSfo_̇~Jqث'/Hjgt$#= .Jw'cL*B D(kv[ ^ֺ0i `ܻag,)Lt:S0.$C0rq݊ AQ1|E3ۛ.)AZ }HOoނu?@BBv:ZlVւRyOqS=]E-c,8TY2Z3y:WWBya˓ĶT,4VVul)I3qO3]$۵q_Cu TluhVJSaZz)C+e{+&…Qb'b=Lւ2jӼ>Vw? {S:p[y0Eץʕ/yNc)OY`Pn_r+glr`Np?|sNp(hs.1-&DaFs@>@8 SN3ƮZQta})Ǹ(IdJ,h7DpkF9r&#QdbWV]rɡp׉P):%Ou=-ֵGlЮ,:%λy AY"U{ągh (䐂vrЄ(sw+4mүBإfzrڦ ǔGzl9n k ?V w#V_8#2FU J߉+#d]`%ի/,馽w͐J >8&}#suq1l% ѲnaÞy*yǍo4p׭b19s P?v~*I? ruc|k5 >R/pjb-HV,;5?RbSA4jhmދ.к;1'0^5mWfRLIA|F}肱D O>zjC i:[_~q4SX'}#ewtfV+`|w0o5_/Wg^ҔX"wiB/0 ''O<%=e).Yi* ++"=Mveog[ +gg'.Y2Uf'(vŠQGϻ +,*j&$䉋hLq}n2R:l|+˸Rgl[ 0?m=L.>Kt~uuu0"_)i#Ze2 ?K,fznڛ#0r/Dd&j_ Wo~z njw0 $`P g?&ufR#3#="+r/g tGʼm+=5!J[GJ6O$nn"G1J- ..7]|*oWii ޥٖ|oGdЦ:6{dydݭbWp׼-RaĮk2Xs&m1 "7p1 YU +ɢ«,21b*n䜹ḳGaGll w"n[)ٞ0.)~׳f3eSx8uSEܙnxGErw^DfQk\̪'UDG}0>[>P+ #њ,sv>W<6Mhcp*`wΩLqêoX=[tG9G?ID)w:wJO E54`9- GV9k@G|@pp^AR7ie$6M1Bf\HxYLnTL^iYf/'4HY،0^aZa#0Bn ע3v 9fnH\m5{ME\@XdnTH0+PkC]a[Hl3'j2uj* LSSS(59q ʂe5?R 9E3f̏TSa@ߕgY1wMC=Tߝ;H+{90B@TGdVybP)^5w WȐ7IBrRE-a($ƌy(w/SpX^HMb[+^Cv]kr0i%_YbG˧y9c-tfWu>Qz (=1Hˈ-^ ]aUxay(֟' 6895:W0 }A< HpJixح[ DE; cB !C! c{]"9>蹈I&b8KidoJ w7"q \v u_,bwˉ bo<.ZLƥ5*(ly#Z6()\n9GMmN,= 6*Qov8D\3BG|79k 9;&5 AʍHrL`.Jܰ* C@dܟB F:T›0n`d4jF a>ޓ09oI10ɛys8eX9Raj4Zȁ7Dm߇42-)&M'iqF5L6%ߌc8B&iS,r\$[w(4u| =AHPQg98'[鰜xAsJtE@TDte:4_e:\Ч2h:.DRQ0q 1yd*8f٠lPàuxjD5 @=]^Vd |7^L[u`Bvy'<;zCCi M<8* e5 OB2R xLV!J,ۭ [UP/'haXZDP،X%P0(E!AǾHz x qbA I/gIuĮW{pʮnGQocw('`l 9@fh'.plp7OnN3v)r1scYStKm\[\bnU {#hS3!CBuܗ6s\Ɯ*mkU a)~l)(%\M~1|rd<> E)H (cOy]gHU챐m)Hؚ#Q#ₙ`?<D~i=EX80i6 `2[2q&5+LB$nvA,JĭCsx1 Ϩ][`\AM{:Be94#Bf=4 HǟV-,@@% c.6&:k$k)CDǁ}is!@;=UC(| -*6qW" 3 !…\ >էmWdYpx2I}^zZB;KUR·ͫ;Yp 9g<Ϥc#F7ZS!j&JδЪ+h:B-0ihEQ `8w0PxP " zhAnF. t;(ⶃ ]߷]T͈ĥ[a>u .iX#ESy' bbECdL<nnX  1) 6LtL`r2 ꪼ8{ꕁ0PVar@ͻ?Vܖծ~E@ "٠t64E)nLQ21 xu2EA()8Dd $9&42C(g4> \"oa e`Ǒ@wA3=b;/ƂD H|N(lgUJpv*'fG XTQIJx@DA\_s""Ɉ)(Z!ߐ<ƑbI>cup@NѻYmg偾lCMt=Vp}f'I !Eo;^g@7UaQjj!#IuSL n.T0}k{bcѳoe 鉤Xf! ɩ',Le}p{go_$BLxBGW K9Ȧ3ɥG.oQDAO01r=ZaUBUWb@vXީ"ɋW`|q^@FdlڬP:^0̲{ NoGs-k(/A2τ4;y̺˺]@c_1bcfr1Ay@ڔHR:)(bz`z=G(> e{ϲ Gug-jSql[aHDG94uGbӎrbMXdMjLpdR9#X:*[X`HH3CYM0NZ\h0g)2Wz~31bc[拓5[v}Rt]]QO^ajT(j 1mB&r۽`MWvMzVq `dw@zsJPq;d딚x9޷L}'_|[Os[~Kih:Je e}F4aDhHh&$ޅ:+y>b ; W0> pmQse*9וj]al0l]u*vT0 6_SN?E/[Uo0R/@_ZA/볛קFq&MF&ei˦>IޤM'G4PGn7en ڴmtiS7IY_ Rݨvݦ| ]F> x췷//n޼zv{1<6o/߿9̻oޞ{71xv{zTo3\}{v?!ZɣWݟxw|xuvwSG̏S52ak˸5NNk4mk._?o[nﺜP *}C@z5TS÷C4Yo,ӛ0g[ZTrŷgמ֣7{{Ò  s0DRD5` WO!ӷꢞW7:j޶|yr[oiuީԀ zx'^<5{JTo hZ`RM*_.|?jF-ծ~~~檑uw|yjs5h=sC .޶o}-0ݸ{6l^Z)wlLҗ~W* Y:+bXɓeq_8=' 'xcҨ ;N_&gNof84'oo|R]7P?Gu*NmxdޫuJGMdwlt@TdoZ8De:RasI)q/;Cv?EbWօ2-ASQ[H9pm }4Ոb4: $r W!&pojmAS©{N+ϭ~_}h+OŢ X&<`Đ|ϧO'sYr̄M{ZJƿ O>\|:[vm 1 PQ ܗUW!wg* D!iNd}` mՆyU[FWt"1w{_8.GGL7AW'%\,!pu F9-&˷60Gxc X!>H?Vqϣ&-Q BV ڤneVFMo1jM\BQ_гRoVk/$̮]#&jofveX ./o)p5Qo?\dbD PŽS1PӏLlE?VpTW}u`שITo+3a~N1e ٻ#>;f Sߏ;>$rKg_(H [R؉3Q#YbcX4KgɄXM܀03o@VʨL'ϰEl>t,][6+S={~Ԭ岈P.Y3!ld`d"LB3 !fR 1cJT n0-=ikp eo/t2HdG6 ?;2pd}"c6ϓ-1oe.7nj4Ĥ!d֌û 2'M<: 4\fp45kt8<˰R)[4e9FBNǂ] z7طLqql\ѨAY쏲rny|z$cJZuw*k֕)GqoN-~o]]O'?Dڵߚ'ХI`NNa0ɵ[^5)]$ G IM [h4 Djh5ojX/ Ld:zD(-cVI1DC(#(T˔TB E%*ΩJS{ig= 9/pzʅ^o JR#ܹ-/r Rr +1uS0 y:&?&} ;IchvLQ=c҃w]GN_o΃#¿YYL?xLw=gEH833ʌRjITdԑ,%xef,#/f2g'Fg.D?ls??NyÖܒhvoH ʄ".FL< 2N ̩?ݙܫ})ĉzP6p 3u={{?e$䂿sKwpzcT*ɧ(\f"LYڍ쑝UֈB[!Q hː3t1|5 mO2E!N9-#*˂u.ZƬ}d+&cRǤ)cwjh{JZ4\i혞e>ykZWS[ߠ_n?ܫ 9܍qe w!y{GJwWWn̶QӑR4>Uswr˥+r 8/^ qTJs3TxVa!υ))<>r$+(qHG@Sr˱8aXpC {`$'B@PA ,h} r;>vZ"|/;Oݭ7Rُc5c!Φhz}'L'jp맻P̬`xFYQI2!Y Q#33Yih6C25&s2$rjsyfdƄΨ ߞQNyĴhTGQC @#Ғnvzf`.M:L~8[r>$Ӭtof13>.~#\nV'3Lj0__lw/]&j8ͮ?>|x検޽ӟ\GP9C$`&r+gv 5|.r~cW(urbӂ UL0>(UJ#CU!H$#QS,WB@DO_!0Ixާ̼My3cf|ynV12+zFi3IÍ,T;ޅ,/f' rj4\3){|O64y8{|O=M%`UXylc=U[1#J%;댞j8f}!@<>o4Wԡu*R(5/|J4nҤWCj?e ďbg DEOιfJISz_uDBf]bWZB^O'f1:qyپF~<. Sm$Kmd!݃ qbpЄ ,v+BsLYc>&Wt;"$L6)reP9@P3(Z]E`w|{<>*AC[BJH+Fuث>j:} yJwe9f(I`B!F!eP*kN&}Rp>>5S\+\#?VV&W fxXmp$ksgddRˎ'Ǟԕ;4Q("#`YLfv<Rj_kT2jX } ?`k6ě(}h\spk)9DV{hLy+d>3NZ!>`,PK = .EoS.llC7q*਱>9'%T<8p􋓬?$~{+$ҚpK<'}S q~p!GJͯP Ԙ4M^Tor1'h @ |>MSð'^EX1l\p+Jƨ~P=b( U&ڷ6;SD+X  cTBFx' aXj#u;Q!1cbx`41"-0$)g]Zb#iZWmayɣ>g+i73߁=]}/%(h|yi Qz9~F`[s_l>lu:]r̈́ Ѳg7kz;gZw02v&JR,^s}\yc~cM74W7/W~jUWOoQ ^;ɭ[d`i-\D媃W;²?'O^p[;hd) 5JȬ-Qn)]ۮٔO^}۾fs.s?玩=-Q[O\wŋJN 䀖D +iSA"Qjm ɕqE%-0'Qm&+m4۔5FR d_hvoHeG3N|f^ ↔/aD+3ز `cbٟ S~3.BLt%vmq'ϪCGO$QeA O&f0~0wT[pqBӉ$ C+##n)ַ|})@ÍK>ٯc.#́l 9!>;P$D)Hjfu*X>uk]\QWA@G,BJRE}u]W3eh7@Un(h j(q=8cvaCAxp[";r/vvnĸ<M XUw_F#v\Aʼn`nZ l.z뫍_]~8x3̌K~qYEgX!LTֵ4kuzXz(!Uqܩs%U !锻45R̟ǃ8H RD:B9WN()-@b-a ,I'Ov γ^8ocٴd6t}g϶>0yN0e:kbBg`*7~tSY_[o X*$lN~rUF.aDdyRNH&5`Plq!f ZaFI*88,Oy~~cq:&J!\{=Ҫ6TiUm.UD]` `C=xQ&b nB-b-bQ T* T+Ǡ՜`)4JYFj%˰Uԃ6~t~)$ l o/amyb) ) ) ))ڦRVcE8h]""V6 4)({O@">߂*`N[;]ODAvh}2VpNнϏЫ8wTA"k\8 icc<&yC3 qBM %1E O8㘦A !:]J%C2KӘFqjC1BkñI3MǘZybidr@o/^uq CXB0~{9U^rwZCB}Dx5P p >r`G[?{ܶJo3}\)۳;x2OIDR$5{(1S.;@,ݾo^풓Gtjb=k.%_)_ˊ^H!Jʺ:VV!hf"`rbRy5Z|+/$DE%I}b}lW⛻sJQV-ߗ?쇿;?db]6C\n=!c\iy6XuBM9YH%Ӫкi &M!ba0h2jjv=Tbjƫ  tͳ?pk5r9G>HI2hъ1!fYJ1 J;u4bB(9f ޛpѡfTm8m X3#8+>kZD'rs{)=ڇ,D75Oi.* ou)YY(.^XSK@3P6scvt!S}V9% iFZɣp Ÿˎ(k3(\H*;Bt#fhv#Z& aQ8P}ˣzO"acWl0ArT<"Ĩ -aVbVxFAQL;Rzbi:xs5.G=vJM}GʲmHYVOכ3BZ*o!" MIiJPl.٘eu9K: e (5dIW U+w_>%˺G{zDuOyw%xZKCto}xf~:ch,EQnEQ>~hޖC/h jӅE eaRN|:SȈ1S,E`j}S~] {p@w_REǯc,o]*V/=s7vZW2bjeOׂUj@EBvVh 5~Ϗ`tST@@JDM+S׬bw\I7Jި9_kʢ$OBdcV b Y8x"5ܘBBЄ9]F8w2i-M.a+YXkq̲3:YB?<|ҡ[#JguƵW#KJ=q_'k&&I') EP?YsLOE'MfOD( 9l? CE=O9A\$B8גB&3Wi65I77ߴ膾lQ}(T-~>ʑ>ekAoS#E7)h~AHT(\pJv=ؼ`s;iy((f${3@ &YNXcg9!{4P/=Sl +8Hb>?؇*RND*Kh!YN\ b)\a|f41oD}A%kk-Iވe )|ApƴqW%M⏒Z:Vgr:=&(h!@ Zb4H8],0NgUBQZ}]AN+B ^Q**jghн,:SKu(tU$ kŢ=k0K@kYp8ZB8X Σ18-#iZԄ갯SN ,Q:.A *dhW7 T]PՂUQT^0fWE6(!Fw ׮ѵT  8 eQU_n~jahqr<-`UCڐǞ|o}<ïsH[I5޾y:PTpDE-GHY@k5AuJK&%9_ͫ]rRjxH/NHuc]ǣaͥ68aW)Ym{Ӌ.Z[D4v.rYmmbL's62`\5*F1k|;BO8v<7x F>Z0Ɇ_jbj̰^͉dSzm^MQS}/q 1|3F&\A;:`5Ͷ<5BwSL8^cCb` y Ey6 ܮgT*$Ĭ7zAF2m.ujFňӭruB!t̚ ]rF`d0ݡ-;xM(SḩTֺ͸0Q}uX,0ÜS2ppYPC004Etg\w)&[sx3 La$)'JIT3IB5fx8)An `D ָgxJ cx@HmMHC?Hp&:lCd$,5xM U%τ kD/iľZ'ѯuQ"'Wrb(w@dMȉRcTƜR!u"#Ni{)r!Z%o\ɓ!/Ԍ C7}y6KRiѫ A R樖_M5;McWرLMQa) 4k*++;n$:QfÈ985Z Ȩi,Ց9b8('g‡<)y&K!j*#M[[Fq!`hj/i4l+v%+Enmd>)? N&S e=qr>wʰf2lQZS}ցvV)K/*$L8w< 9"RNGQ2K=qА z. 쮧2ҊNgFv87qǹ郄Vɠ PizONyu'z蝚oD| VAiSxWF/w.އl?*Z*1^u=IKTƕ((dQ|>Ȟx޻(> zűC_qd/8|0FaqOS拐N߭\T0j%>$[Ob{c s+&'@iZPw^91klRpx98eM)dӤ U „I_}+?QJfύLnpŢvAl:czj q br/;6iAWfնMs_45)/ʘԜ큉|M΄#xYLhkGiciqjA}q \iw䇵A&.R{w-oς$& {E)!Ҍ[9AyxbPj>K㟱v !|9!c6ؗ@>l}}KF887^t9{#1-lmƙԌU34VYK+` `$hsAɡ砵X:{dz_ATJ*tds,=K95$gΫbYٴ%(sН9yVJBs;ff;шDrnp&`= "B>y]zprՃaH'T4A?UP"fb?EVqP{vߐ(!+u#a*ͮ/? }%#5:1-t"Pؑ,@vp*Nl0Ĵԧph<T8O,~ r=) Ѿt>H9vqAYo^uxT+SRS](MtQ7͇ٻq$r_;0c-L;v2s(q۲)KD&2;8#RK^lA*8x 0K&h5rKǍϵT[l3ƽtLf ;fXtn&R+!a.sNARUh6t/,Q͏˄aFaD~|$! }[0YM~"<w¿ҎGbT?~êD;RjYse}zhA Z|&\@I19d80"$) Rnd4H F1iQҡ@1Bk@ǯN-[CqVޡƱ׀bt Pv}0VGve~܀eX+ÙK]|#|<5J* HHqrnUȵEi쵨]v_}OH9/ ٳk3u5PЧ\5DgX0|ɂ VB0 /rB8U39`a ?Ar Z"Nܸw :Gˁ&m٨(Sq 4P} #-VRhS 5RtLQ g+j ,`&LP^y aaR0R4㌧ 9DWqJ{Sqt.^%%!|;:Z`(*5-(RRzM.\nyTn'@2XpKM! NOm]YԨHNpyÇCms;o<=d>>nJG˹cӝˌ vT3$LnƻfRempDi"sw)? _vAf딽&JZ y#*4 pzp ?ljM(%olE}}6CjG}&Rj*\vhoքϒtgD~ gbLa؝S`&k'_cLhԜ|1J\Rb@h|4Ga'tP.ud͂:rG߅]){(I]f Ү]a_~o&tPr!<[`N p#(M*ۋX)2AGUnS^b#8*v ڦ{;;pš[X v^orz? |K9ue"dt 0 k zP蒓"ؑ}X8xEꥢb`+;YOQ&nTU8/6(펍Vc>eUBv=<9V ?~?bIp¬6IϒvY|ϬEbĂT|$!Jv!~6\1hyo9bTɛő7;1y3V V'#O/s4mq.&P;\hkwc<#<$E"GM&L8 LemoF@P*;Fy2Cѝ,G[2^LLRJ o>2:N•@ cdB8c}I5v kàKXˠKT@䃳kb1+"Z~Z5Ƈ.em W'öE=h7)(m^,:?{[H)tE>+&ÔUC~~ {J sk=7oʓ?&AЋœeVKN!qٍ̻7K*j@1`)Vټ0Y <9sdBZLg_DW|[v TAqgJeXQr"xx)f!.vrg)fZT sv=i }hlڹNg)aY͚"f2)CjFchxǫ!Qu鎀P9eLFJ \ƕ(y~R.{H" bA/̿evG򑯝$琾#n ՐI }LQTLCz^$z5:_,~E^;^7O|"3^Lp˂2sNjαnf&-2!a9$t7#cDw 5x3-`dp] 3ޡ 2kDA[LvI}MwrƮ2:tIU!𳥆3tj)j :np`m’$Uz#c";>m{)@ޜ[iPw{Ρ#nrVGD-[̦gn§;:Q?xǹE˵{|wRH'Uʰ =)U%a%I}HF( Tq 6_j$v0-?'eL>m[a v '-͵RXS!VZcJ] nʗrHMW鵆I)" .cXe?Wڸ]@~ZTXpݥ9%t-3q(z-.oZ*M~*}wȁkMe1!e2S'RC'X` u8#'T]x8 {/rIiZEĬ GR(@L$@eY#=6.QaXPiӓ"v۷PO Mf jf @",DW^#!@5T:`ϩ$L*voyN!3fh鰌ɼLcf,.{BiǼhHjAB EKEާ3$HѺ9/r ץ:1D /~}b{cxxS%,yϧ< cbV"aup0Nn_=Y5m!o[[@W;^|7$ l_F/Ÿ_yyoj0ݎ?__l˟oǷܯa< wȶ"Zًa]moG+?^-Xb'XH.ѯaYTHvjRG␜P%%ΰꧺs5 /̟ԠEGKn=cڧUfCG FA$ `4,Љ5l8V4J0T)'=,Ҕ)71/Z81!R''''/pӂE2kdALQ,!Q4 Xyr 4xQ:~Sd5.~ -}yp_Os? m/3Xz|0 m,!s|H4 :FɁJ;7fYL0uYrJ"0vm2zhvR'm V`cZfWg|Xp""bN3eY3y *h@{o}0R<:+Q[!&Cimp-ƑPCWKƒ u89GN%:IN9è!FMFAڠ4>bP,%BXpTP;8e41󔻕4|g@Iyh^k,2wӺ2k46д0`m x bwkCꊌ*8T^Ɉ'+NMTr|I+.$fJKŝws^O螤NV=n^YSN9$AY fY/Q󑏆zbM<,mbfRp:\- m*vR>^ϟ/>}1<PfPkl̆GSq#>Z}MUMda:d:۵j(Q= z&M dY2C=zč =!3Bٓ%YTLllOƂ[I +5a>MOweէ-  s=)pPr ɟY0奝ky$}~٫Y 5~cw>]MOON )!៞-['NN‰\F'_fJtFb0WF.+g0^OA E(eN8W\*)0` [UXPxjIй"X`}*BJhm-oay#ɍ4L}<,.tc1Ohk ).`Z5y_^:vm)wVzj0~çB|&Wd_~C?Z{'w/|ij2 2$Od@]+D1h4L )8&xb3nVYZ^+{qApD=>h7 8yY70vPJQ)RI'JnhLY$CUti>Ic+az/77Zc䣦2P)C$_5S9oks\ҊI%dɤv,0U9KZeQҞ+XԟtXFyxŤa[UYeub.֔ZāfNj+ŐgU)ZkE4%^rz>z\_`C3ʝp&?eXK! y*5:IFWNXD84Q~$К8ҡDsjqD,)8e̚>oRe"]i15XM0~.# Ie=Z8njd5|!@D7b u' kMV$OJD ncm_6 {IPh Ce. o6@wkAQ:1Ak>[& 24Q wY+"NHI^|Fa59VcF&7Eʂe؞uK FH"ðY'1G"ރCi2-|(ϰg934Pf܃L*DiRk<&f\[pD!(pOy$ʑ i P4#tն;dRjIiG> aDtY]o}:uf9rP TPbUrq]j=FҠԛǠe|@.,hZ$D D]n>Ж0h/f<e6^@7D?a}sDe`?,g$cѧ=ԅyr6sJ7S6Dq%)BSU&A'[X2?mdMC!Xd]2XdC8N?wMoԦü֪ccwRV8;s!쑿_,ϊxў4ЈJÙi"og^8ehީrf5k 0ݿ^~ܶgh9m3鬸H2mrZ7h27Y ʓ1[doخ;O_>5($X58hȕќwi3iA7oǐHeнa=U%Z5y #\hA=ehz'rB1 5ק׻o?`ۋ^,qa/ScUxn5d4Tj޽kJ=kgs1 p)P, WZeJ'&MJE7raͻ!wkߝlmFK.gYRkvm^I~s AUUܺwgT!þ}%UE|Vw]blR*j8Va&z qAƗ|n`c>2-^_-ǘ[c8s w~gz}m=":G18}4 uߔQ`Cs]b|6ϑȫ>ڳ8n4\8z1$Q߮8_Lgy(|Jݍ_&.;X|7& '1Yڍb%(2\V-Tt_/mMeRER/,?_1>>/GGz:9"XMΦ,^o~_'ף˷ٻ涍WXMvEiUe'Ze9Im%.`0WR"AQ$A9l=OOtO@RosB3-[!tz= {ny6uÛ_;դIZ޹?Gu\w<\ŝx8 HoJ3殿l;x=_Ij7}}N{SF4meuGo}<OoFA$w)o뛏ouc˿wΎI|&/Mfޘ L͠7aToݰ9S؇vQw'eބCy ),ˁi7gQc e\OZI]<|'7\9EF#ّCZm%0y}s1o{8Jt3-""B5Tp(ID6Np:èQJ%uA$42FJZ!b[:. L K[޹NMͩ&"])%ӭ Gu{8mf{})>n1vJ\H+"#CbfqȄ$ I%(Ċ-?6e=‡U!A^hf` ;x$C0wلJBXP;˰՘EN2ZPMWi{@#& aDHXG'-Bh/x>CkqB`/+[HuIv]vzǎzܞ/]H oJZpgj\tSJFFUR('qty0Gc$6hE9j4|VrѸ=X[] C[LBiuy{%)`8VLHjHi㲑p2i][m l8s0$EHJ! DxsVB´(RmݜBv%b.U˵tEͩ&#b#xժsE /|vqN")?oaMPc| L2/YA7B'rx|KwqiWoȗ׿ܸa$IIgI<U ~*~܌[$V0s9W. &v! +r*`Ś挳fjFΠ،}ClkG]1VƟ qnpˋa6#F;AB ₙOذFRY@ӹ1>/ `z<{nӔgczE `4AEabl_t4aAo iò<Évj̋jy1*z&&|rk[}7F?x"3(T|0q,Wg=i)훓. &d:*(lzJBAoXj)S-V{0n6Z8n}Q!'_U`Ms]G~d=;zݩƅqN6tb':$ k :E:j~wi-ǂ"LXrWPlZr0,*9p2юXɚ` E2෻(^5Bt~8.:z)5A}6Nâ+516"K޷ÏŕS!IXAT,QDŽjbwu^6+TՉ$+sٍ~ܺkhnX5X݄BY~vga I䟼qy)˥XAnB=t1ghvOaU*wdRH1.pEnIMB uW͸8C ]=w }<\1EU][3RWr1 |P1<:gci}(1Q+ܻW_Rdՙn0rej)rjesqUյ7Z/9  s4bvUz{a ?*=wŎ;ww;VJrLjB9O]"HJaX.Sdr;53Z7' ! T:jz@+ҘU^!PB&Ch*iES$2d$Sabee B(]!Wdr嶒 o>~i#fY!kஸGQF߹וEiV{PPj +O -ݯT'sG׌FuFDYA8b G!3;,.TXҗ93bsdH֪f IUs^Ϸm:xߪͳT[ŵ2p[Ge_[ VA䞕<pu)ojk-L%{S>0ֶË oԆW3Y&ǃd3P`.@W `jO |.h?h1#w\:up0ʝЮtu6fلS64_eƌI!1..W13Pi˘h2*a3tqN!R 'dO~j|w 8055q/\:GpBФɈQMiF3Y?g{0u<=;e7{siZ-ǭN1% y@RXk/2-:zxv2x3 `9cRle$8ʍ/S,H! Zp d22aMP!THWv2Bjڛ+ʉtyk*Ԣ:}0 @in%TKئR.4|C+'tQ܍;I4 $r"犌|"Nʀ 2*2( #"*#D$$9ĹE꓆+@,Ht&z6ݻq͏{}lW'bwzrի`>=y3[[A9Tsj:,j tgT(2`( MLgP #VX*E.:(PQIa\1$ 4L˟ MܹTK>%ka M0Yl>L5u#e6tw~]܍܌MrTP]ޖ.c(BKrq$X'3^\{v_DD5/_J$Vs`pu?]ݍſ0LKB%fB$c5y!8%ϓL30S7r~7vYc=/UiqșWv0%fh}*$E%ǟOP5?U^t߼_\Š+ Y|)5\2j/N<Ѱ? o\v Fg3toi3᝽s'ץ09p0̼<~ʪ3ܖZ%T6(EY*2/g=EӚ>Nz? [? a6L9~M18IF7l2ԍzbtIsOYҎ|m0' _1m֧[wO?|^i WKL̝aG2tPPH0b qˇYc`xxZQmÆZ4n:F-wu#"7G%5sx-%~|v !A.-qF:fK897v_}+y>~w+l/|'FZ!^vƧz$%Q5T!-Km,iLcXBhNR`kܙ:(._ziF^Z=.[8~oyMWdFÀXgAQ18Ă(@B9E-`FWVDMu }is0*f7ino\ G\61̔=]S/xzOcQa썖eHX-0Xՠ>#PtSD5V:ER]ǐJC҈&8rr4/(~eZL)ɟUըO8)y]3G4lH0 Q4€K#a`sΆXR`!21Msˬ!H[G&kȘR \YyόBZV.c9[jHBWM"+%(PF# R@)@@3Q=L9Eyc9 %)%R #"~–̄R7lHhI-Ƙסd^׋3 ҕBhVoavPJ [Te!D0B2;B*$oVi ;&o.ܣUHh9 ErPl5`sq$63V, r5Ŗs@ƞ" ak8 gcZȑ"aoW6ߏA2w8L&i7Xrf3߯ݖej% 2"*VYY3:I"hA谣Z7*:&#ەFA)ؕ@tpIA;` {jp Oڗw I>At΢,x2f7plh12Pǰ0S숞blc-haY_$KTq،Z+Yio%1L LJ,IKq-Ă3cM <9)< ̷{c ??Wdn'-_yJ m緫=,$_&!udȑAt . e]*Bq8" m*[2\BQ CXԄvh hqV0JN8⓶HX). 㠝 ZR!s^xaɗLk yL>Π.- \.֩zWDKbD% k K PDe'HjFRi!oGρ#u\$vZqYXݖ%n]"B${KB7 rl lM9XaUE#}˲q*{)76 m7U}Αw[ˑ.wr<4Ro'WYwq95H89dSV-@F9qyLwLWsЧ<*nt&i*Kޱ:ഝ&}->]~3 ,Uo??I)S,,2F80a91n|'>N?sp?f_R ЧDObNb}[Ax?:1B¢ȝd+ TY>Z Yĕilb2`~ CmK lYFEP[#Xx@PQ ט52r,=#_hZOţ^5jɭwyuV]X솽,\jnN􋹾.!5Y#2qYך1b+/t3jc^! cZ;S4 (吾q*ZD9/e)b0V! UN@xK0wn~Ww~_pon[cW\㙙#B3'=m|lQ2Y.=䓨nGkFexC[qA7rzRTSQ[}8o=8T娖ז8 |?86cj}s8tG#j0< Béjszp^֓1$H*G6c ttW$I4.٣TwA04 W[ nfٍ_~`W |􏎘TR}Ťrѩ ˟AvQkK'EѴ:9]IL4 zV6w `]=;)7{I0/|/f[nTOD҆$3Sy{M`Y䐃bZ ,z =hkF3k-X!SjNQ$'5&zMɾ̓fDiU: 9r\! 4$Ұa#=ir24>cvs \CR (a"}>0x!4[Ux0#p**uz Z;J B#bɼ;Za*s0J;@uUP+%OH*$P8!"‚l"+#QqM|n$P1 KÞwF([eMq$(꩸`:C] R{Jj|̥̞K%[@\د~=NheVt~yZ ʥЕQy%R!g*K P(# <8"_~b:?Dan"PZ8_غCM*>&2D*24\j&\ n}xDu20} Bg|fMĻɍM[G.@ԁJA$.LexpJI*ذ=Pׯ~l3$"5nՆéVpw9\O~XQiC1QQ7e6 à MgRbÃ5mpy:5LO̊٢Y71zF0<ܝŁ=eƋV)yo-p΂&xh8WxMu.҉90/$ӗi5N"٫XŊ}`KƊ ^U x3_ކRP^vUB~.5|0߼ٛEbRAMXՄgTk-U}?~w{Va/ &!E]W8TD`C%fom["L/D2QfN7 C 9W<_s76>* N6c+Sh f-]^_{{}:qn=-mQ$!V.vWo[#ւ c r=P,YZl9f(=$PI knWB+9T|t58L.gG6Loau] s*9er)^¦dqj;q U[Pxa [kޭt|E^^I 'aהP?%X)HOGjN/ \֠_]|6wWb=EEN:L )z|^C%|NMǙǰT+/iu0=cٲ23>Ds ,ՙZ_>r @STx2 9!ͺރeHhud<̠3DLCQ@jN(u[ڠ% 'B&r-m5UC龂h&Z\h.:)ր KFb)<ڀ *΂58⤕6AvYAZ,-v+)j\?_)w*T:Yy>sxv01Jz,`PR1@ 見߉;.h29lҽ?x/+-ha5bj1E wUn(F8A36B1{0;XJ"B %Fh 34bTAKPtzO'DfA\,3zgB1aAzE @;3̧[cYX71^)F|_^ƹ+2:(}`|Pgx p䜕Q.uD<t.qN{F6n5[yB0Ǹ??΂'M#Ӫ3 ϥ3!Ĕ&i5y qw7ϓwU݂\.{Ȼ8P̡Lpm_0%Eׅ2@d,}p_,[nJ)ȿ=8IiSȩ,`@pFDJ"A:gIa*NЦ.l#TK g(b FOy*˧J8?Q*B>\ 3\.BH "J H2xPEԀSAtxߖrbnw2"%8\hLFy`J#F`kLYGw?qآp!PMDkȃ7DV*~󼷿_Eq˥%??m~3$Vɏ6ߟ%Hi<_Tp`ADLV Q?<˻bB_woE_O_Uc,*?waC.aڇR\pHrnzmczl%74yy@ow>j5B;O|qs]HI?>b+5INҶmҲvM i[KH{ #!ҝ!l:Zz2҂TR^VCB8Pׅ05謸].kknF嗭ڡp)ۻujOj]I 7(%ޭR␜N2ؖ84ݍ`Aki~$P<5L+~(p?haR*QE̘G郷a wbzPP?>D[ƿilt&k]lìlbRZIVe<[4ߚ8z eu:y_.\'n~i>A ED/ 'kϦ#,vI^,@/C¨2F-^da[".gg2#2Aᘰ}r}+[J(D!}ZݑsA-ӣdZ]tSL+N*ǷE@j)|W}Iγ{+Va */F,[zF,̷AA?+-Psʑ~sE΋s,d+9Ϯ"RJd VeYMJtK2Xn)<_Qcއy|Ycz5]r_GlHG_ݐ6k&Ja#U7)@N*-<\}9i~%ٞ(|su_]č1N)pD2Kר+vңO\1/%UbNq@]ʍjhTϛ^!xm~KgMfewuvz /`ކRYxku qEY`Mc}^Kh%%X.߂͕ĂPBnlK]糬3FC@/YaC ,<. Olo7o94㜡!#(HQĢYYEڥ*5V3 $Di,+DZ֊RʂD@z3f)y; b.ҦTv!jLVD0H\JXy¶ F׾ ;w3Ҍcg$3§8!t{~Z?g-vKD "RgVO833Tl`Hy4kԯnVֵn.n60̸Ěv$4}\v~zb7;gd}nJ{s2(<HSzG zoZ̴% o8 *1/(:S&LhDH\8!-Y΅TeBR?r8#/}1}H6x1Eo?}PtX_/nQ[K%c>ή~_|Mɼsr=}K4^O71lF6%Yv ht%f YOvg%.3p}QgWz H=1g]#8g==KylfoTЪ'fU A&yڅ$Xw 8 7@%LpD:aVcCʉArZY9!N,YR,.-w1sxg -NbH"o Eu3l8NҌc]:ʫdn&T9*%@%B<q,=ps հu`(5֥2Ri9Ba SOle-^f+q0/7k“2K\}0"P0wLJ{-eS*fDbz2>\:%H4n 6*@/Z\!\ q݆VXc.]n\TR0᳝ R 7J:T h] ? TӁچbZ*6Xk@Ogmk1車*"~ VF#c8 Un-zPN2o:x*)SYܨE\ 2Z|1^{+,ƈH3paYr:;C) M^X-$Mؓ]KFwL{^M֓ȭr،lF3?L~4{\?.ރ[~z./.0cZ],rNtp'2ay`rWQjfab T3!:O3}p0ej_Z%nC5$cAzj5DJ!T#(2A `EXBJ@FY r=8#w ,?G,3g;-Ë `Sj}O>%ɝB#p,%E zEt$Q01B:.#bj E\- ޢxʈnA1R0{ۺU6ͬd^IJx2\yR;u\k Ĵ5օ9Vފ|:|Auco6J,{5&-'-(ehoӜj&q\vܜvix밝TĔVGM={9A\P=c_7*INw j1p&y%xԈ;udu_eÉC{c,͸9w0G4qY~jH#2ϲ_tkz+8=8^:{*[~+@ /ڑWd/56ZYwW]ˬч6숐4mnUv.d 4%Tq|T]ůl)aM)EQpwgCsKU?ɆNr~Q=2JU9BcaHtLrȕiMМLFQl!B&ú{Eh ;}U%0úWLܷ8]j- a"%juRXIΓݧoPd1*!lǏ}i&Buխ3|F%L Xo0F~crؓ>/Far}ww9vAset͎0yHf#LuɊWbN]r3tBzފ1R1I/`FǴc6@,M$F"ц/C+u.% Ϳkfd>^U'903Eidu"yhuqk}D [R;2vSQJQ ޭsP?>D ~ ,9y/[fFU&] odAsWs*wׯ# '}@-5f4gSoK^5AN&)HTk" PI;zoծ:L_TB>j<] zI ZդvyE 4țwA7pB]yOت=+P3J>k`*Eu Y.G=u.U}70XZ{or0g-*r]7EQ= k^g]x bocTv.lƱ`Юx(A[2r2TKZ!318F&w}(Gqo7]_2/O,knt9uj1@.h S-n|n_'$m=xeҜ6Y,gfsfiq{hKRh_vY#LNl7q~L TB_7810PLeRiʱblw_uKfBPlaFKhjgӱ:fRsRJZXh@g6"jX6=k"= /=IHD-^Wzo%齭d:xfU*}evR(Q9ckF0 )`PM! ##@!CKrRh=.[E97+}Y,m(V)y]l$|I`'[$;oj͡aX [VU.P~N}ġgu-l}F4[Ƹ1ռz{|nvyKA'z]:zWzh=E ))h4n%XmR8ʺ}E(kް-B* 6^E/D#Y/T@$ϮQQb.bRH:DlF*6t/lG 6̺-g[K2LQPf@d3ۧD%=H) aN0)2Uv+54 6"6Lh*{{aK@L7d*(ևźcvyó@۾t\eis(=eve*@$mXU :|Ԟ4B|daV@(,Xm:bXjK!àr$Ió,Yz;Yзf}3voLzQS7u_~DnL'GlRs8ST1rZ8o7_&ꇷ5(_%ӓiӵY<|'W3F7)3XsMF]}Mdz[c+`{ۂ+?e"^O<%/El,Z𙂗Bk(BvլFJ$eTۀ\4k''oYі^Qc%TƊq϶)t?(qzPH~}кI\al /iycxh{N:^A9;gX#sien$ Z~ϗK¡e_tybvW160t_FyWNb2cmR l&eYP%=n2"';#U5Ijo@w;' #iJ_4f}=P\V[Mj;IBn -Ut(w'5'WX/n~2"Njoi.c ϟ؟O>^PͯnX+=?:}h_.Ϯog0M1 㥮OU޷LU^tMpY^:)@Snr~mH[C8)׉kD{ME6:FLS I ncpg<ł3a*@;/F4pH Ǵ^M̬ #4-?n:LT#pE ̹gd}ٵ9| ۺy=M7jYG/S9g'7~w77ھ^7&GD=tIUqxs?sਤwiJ:DqF`)z =W{VlsM}y}F,$_1S9 h^FGۧ?IQ3hBٍ6||ikiD^,^$`VrCu+8Tls߮q3B~}*7}GXY5?Wh38?mEez9'̊iywb儉䮮H}OEAT) *v"{;'ir=j nsݺ_˽QOB43%U#ľzVL[W c{]$5AU{FMuv$Schx lPbzv/1Ǡog#؉+G|h:}hu&ž.tBA~t0p8m+ +NNǻGsYq@~~da=Fж]1<{~ۧTx'}vJ/_?TcµXӻfA,^6!_KIߧjӎ?iǟڴOݴ7ֻhDPP`M".9Cb΋\4gKXS?,I?XԖ5[-{R0{1t `<\qW(47;K߬N iL(SA7F5>> Vj2 IHlP <웜aU"*#$' q1Z9A 9^Ӌ^|m9͗5Q?MCmF#OF DƩhU   Y1`Snv jQCPVvt UT0 Sek SrEV9%DJg!J̆SV;!U@`S;ko-ߞ qS\a*vɪTP`[U":ޤb䭦A3ZQ ["*- "bXa#&Rc_/ŝB>b+.4`jd [=`ڬ~`M uL5#l2/,|ӫ_mqY_I%Y9INh4U&ɽh@`A F(Kh(!"A!hBE!L SٙRߨDE k멼k'Pn2oj&L|bjBKzO;}mplZv&xSr*30qf^XxHŷm28lhѸJŃfV-j,5:[fոQH{9K{EQjl7&:Z51AQv{g%x{u0}~.S!ga6 zB^v{:ckYŞ^QWPBI@&8bp*H|$egdc(@Tt&h3Ed ]5C>N81y>oe:X s;Uh?V4ե곅s#!}d;VӴ9PJ/:/U *Ć H@7 3TPgRFP(@&bLʄ qZ?HGA9er cZ("GKf%v)8n+TBrzo#uRǢbb( v1mȝS(mq i:P|fΥ X *' ?xC`v, Ĉ6t2Fnhd*8c|u Oْ;@~NTc!fvSlCu`G h:qTW daTQp?8Nqh`Z.& $ 4jR 1@VRo3Ah| ,bM^18t['S52ƏNF.Tv46QJ)7Sh D34M)NanVV=IVQ f싣AN)')N!nitQC+CHQg~3a*WIid.c%$,W)T?IVRx9m[yi~[Mq6RKS* gKZq(6$f yr-r}]EStҤhhheCeQQ*I6/a3)+mjbW(V[]عR$c$ " Pde{ HBpY'ri TBX tF2~PˬP- CM͒̀z@!bzWs}}*oO^Ҕ\5oyOKS{5{CRwO/ҷ[3ВdGεU{O_~ejz_?ٗC|xX5B5%m[/OY~ c;cF xŗZw :Y_(}xkmz 48Y\Ao_umzM*u^W';h]o_z,?Sxjcݫ\Я$V}wZ?gDaGbV-z?uY]A9&KZ]$Ϫa 4>h%1UwrS0kUv?k($ tkS[lz@ĖYpW+QЁ`EheV-,DGJ)5S)&6S̙: S28S?&+Jٺg^@ "JbrD)\mrRǁbVT>D!eKP69,] QP}4 =J j-n"=Ůc vC= Ϲ"d'!;_ KM_L:0,cua&*6f ۀcTM sƄsDmHwL | m^(HN>c0e u_pl!mӝI ա#f 12LWh')l  F+YlR) Z*W±3nER%y Rix4tI(pe|_{s_ئb'Yv`::<'aPs:d@3f٥cktoyqKA%wP6(w.1Fƞ71&Ub>E)m agV\He}z0Zh#z}d~7c^0g;|lVHm ǓΫ"ʟ7ݛ4^5޿]vWVپ"mx:Z6b 6$Fyd\ITGJO0G?7*o5Q^J*sќ\-Vc"r@D׋rKh et` VYIUJz&x)YzˈX;fv+jtnLɕ y]koG+,0 , 6q/3aTHʉg0}o"fUWuwQQbyz{>,K)iQ3qs\BRr?1DȬʙT^g{`*JDAᜭ׃ҲK@8M#J2:Jc{Hp&=*#sicKЌ,eagu<}9@@)sK ~2w+Vc l۳f1;.PJ.Eݠbr^>P1 &d8:~A v=Li.H ~ܠIثa V_Zs F2X&\B U ثRo*X fg CձQڡ=9(a%%h0cvq3“N& #=_xө?At oG}?-~MgJճ^:0:8cӽ,ݣQ":ہYuhb>)?O]7$cVXUS>Xk(o">fL-k5 (\r0p0I8%"ԺޗkSo7Ks&69ۼ㾐k,oWHܽUtẗd"rWvW;¹MϖyTr&n綕br;"N(Q(¹m$^=@h~퟽ 18M#qso[=P(d)WQhprByr8ShW)ԸG [*a<5o'Vz3vE wZ=7ۧ{R%Hh{={_b ¨Ш?F>,*j:gÚӽb (UHF[_>X3}wY".&tuNa&>Ca |X ^|ކsKz̆YYkpJ_hg1lF sB =ǽG3zkY gey{t^ӗ\90I{A!)I{t>?@^%ڌFg\qnMѵPKYpBw:+G38AŦpSK54@1Z%P.(<8Fw 2)uz~3¼Spp>רNl:SHYHKHtuDP;$rjzPIFu4^``iK2 H4`x^*N]_L'6/e=~}ℴFݥ-m.o_bøވ5!:sGR1ben8&O_Aݏ"Չz8o!'z|wz^\ ,AL챸l{ :`tYP!pYu-dGM2vaX3Ci쟵#,SRxb8+f3=""yk^}Q-ߓ\7I2./=Bѣpw$Cb؝4UVgD "q"GW*LjO`a•ԅ4cJ&ҵ$Fqb 8T.>{C7Z ,wX3^wRKBp-Ij>Ry"=?Ve^^A_b%DGibo?ڟXnDZw@N{W7Iݛ7]!P&X[i~w6 $F{<߯әc d:F_6kȪa}=MӴ6l]~ւ\Lmymif>lt~)9P)BȮY <)h(՛g1w_v>4hGTHăRܘoC؟ʓAtLͶΚZ3cF>fڎ$Jt5eUܲ5Ԑ:ɠtU.N}֠:k.b[DR0uPt-X~D9k`鑤3YB4ܨ2y"IS{l&xu]PDTxX +pa V='7y?Njh{ZMTwoNOQ"լwCT\Mp ~>=gY H9pR)PX`OEZ&UIAwىxHzXWo GD\+rA\oexChWǭ= ƔhxD+ܿR:!^֗ 9^VѿVA'~cbsZb[n= E?9rrJԲw D4(f@׳2)4-w76=82Nt}s=.fjWd}wk9]i7B5/&K u;BOot?1bQ}< g v8ZuulNHւ&{&6f5'/w͉/JAàSD5q1a ~&@ =YRJ{("¾%ruzRd.@ 4M̷էm<*FN^)7ixvt?Ow.߁W)觛Ȣbm򏵉Es:!{Pbϋ\0\yfSO0*)pFʁٕߌGo&~e62ozXN'"ӊs& ~>{1m#)Gj(x/zLFc2;޷Ԕ V'Bƻ`ф N4`R y^ru?[_\rrClBK'~=pYύ+]}e]^j5\c4Vh1@ ݯP;l^vv $m-~ "湉)H?\Hr&qQnb/tМRSvsZ'b>.*tКj&RDEtѱyCcS;m Nbi!4P+Hx-N4P eVMA ) 7M۴`T}1Z[Yy|Fw\ء9WT )f/$f>XA,P 7_|NA)͐{Yb&GRP͸kwy#b*y1ѿ~^'޿B|"I^yo~%x^f%tlwhI$IR8GP%ItR_UX;mrE}!+N|"Xc8:џ+И>,ǧ{dh/'І{p{9n1DQ}g P&$Ē 2A#<y ~=)e=[ɁHEm ƌ~0PO\9}Rzoը\@iy㿾;6yso,{H5/ EQUi3H)eE˒ +*JU((tyIcaKuԔn`U&IU"yerjC٩tцx56}ti/ĊhG K5V-.rx ͘аE\ ɹWYFJ $(/Ht@hx8,S mNS RKA$e ^!LU֨J)Ga)gY&<+Cf!H&XψTWD9V9ALK.h*V*WB#p*qTF8H-(h^Z@PUGD3%o<|>ފ@#+^e>| ^ /.W~3Y~Nƕ22{t{Lp6yǣYZt>]|}|)ZJ%G>cbO9 }EZID( o<=]7?ZަjwtKI&Af77n&j&U\ ްbq_??׋F>3'A-{g0@\o/nE]T c7Ak- \@"3_& ~>?jR M,$ͪ2IE^d)T= Ҟ8w"NQAF"Yhۍd88a,I NrDaLbjZUnxrTSzWT)'U*Ѫs.4\DiVEΐ5T Jm9l8O暠l)%X8Ttֽǚvr u^JS붅 "a-6'{ ޵57n#뿢ɩe̞TlMm9ͥ߷A$Q@ %d 4Зe/#e}qf_(g\yaVLb#Hf!ңrǫ2Jle[aKmJXohUe8s SKNo p5D3U7?Ԫgjtתٛ!Gotn^?f1:iǨQ͹ٷdy%Bi"-lfy'GgQD2!1u|G㓸-Cl %9\vb锤;^G=t+. EkZWw/gX W ju;48h}WT#ЌqFwIlJ*qV-A6ۼtt%L Ǯ \k >jjkML{#_”ɭ&o+~M/ 4 {cŶ9u[kf6 %Tな" Pa'HP9ϰT%xMH䔔R4%A<-1- Lf\J#!Os6RsSYu{8%FVm/0K0;jеz}_?DShGmaS&,={||<$7ΑaF3;'X'k {cJUR{CM6}d;5e]ALY.0L6I ?!p$ढ़ TxEDQCЅY5"hɆ[W\ZYݲ>` q)X]xpblԂ :\1QkE;kdvH^C֓'SpS?/KX>R*1wC, #j9KM&l06XwJEMU;yQpfb>?f.ˋ䟳6ɋEfw"-.~ |oԬڐA7j̀ޙ3G/r5ٛds3kCT7T== UDЪ_&ԪdV0+*ʉiȉhN<-?fcnN7|p܆S%YbD֭ 9q )=mƬBxZP \T'>méw hݺАW Rjuӈb ۀg$Bow@CN\E디>",!xNL)*Kp1IBX.onE@x(}Cc6o*cE_a⡿?yN$ ,ofT/G4C^nghHCE'Jp&c GPo=?/Qy)ا8PְJ {eI!|XMOd$}jj3ʣT<}\lUZݠz6 J-:F^lsO~kX;Jtyϰbkq'1ʻY.M2}WR{cqTXj$ Eb\'&%] 5?5yU]<&YqUm4x2&/0iA2z ?iJL[Ot)՗"{s?k՘i: Y_UrdYnfI:IT)J-RQ2on7ٽukԴ;h9?׶vQ"L7wEbb 'Ɔ,u0\_\$j+3]-~RTt.+K5PF|Jx99J%\,KE1>ay UA-. -B5S f9JK%N2B9VrJ|f sIVb]ϱS:Lcj9EnEu,e{ɹ_u۪ͪ f,l2*660_7+ΧxʒLIJ!S*8϶,w(* )yY2K0)B&8SN$yQLjh -=ۈwb";>G~ @ V8TLZctE:J(ꛞ{t^?z2$ոo8IY lcot˨!gr'[:ܖq9;G8M>J^ãљl!є-D۶ӷ%At]oՂr[=3[}<6[cLAVz~ 8'%H(p +z.QbX-ˮۃۇM8ƯEN{1'>g=8pCW Ն1BKFTlcP(-go ]Mo3/:*Q"xzª8{BXpK"W5pH$VO Ij<FKc kEuèc"L]-R*Կ'b<>ig/SdO~2Kat%.;bkp4ib""ט.ce{|p1j2 [R `*M0-j쾄Oϋei!gZj'1__3q{=f.ɄwVǏxnXߨ%7zg^`.Fxj8mލo}{&P޸<y_>_&QƒkUɘ3+*FHiȉhN)6-EcnN7|p)$u uBCN\Et<7o/5{|anb~[j4|Tt\+xF CΧ:az*EtlȳwnpyU"Wh(2%~Q@NlW:b"Fhtɾb4̐&ڌMӎo 8UBt<Zv'3;;]o|?~̉8cتd{\HutA$'2% K4$$]+⤤%2%PPHJ/<L?#oz "ީWxS.4U4HM޶nTºb:nCYnun]hȉhk [JFq#K}))'NTj=Ta|IԬ'*' NUm%|x 螀?o:/iO/3wޟo.頺/ S֛q:I6 .#?_T$2섻‡kdq6pP#JUaX_?<]x|ANjh()xBEdg ݝh0T͟i_tE Q{$ϮjX3l7%T|>G m1 C*"jNrʎ|pnR;Rr8pl#]|`i$%+̊l*1S9&J%9R"M}j7u C]mр!P n@Z!*jmVuD3G%J"1T cH[$-/EhX_ffSj99T>W)s=U0a*4}sW$ d6Rh @!K2M2|]j3.xǃwko.ofbE]L)%X˟MQI( "jR-խ 2wwk:\fd>s;JLXtʄY<"ӄ~5 7dZkﮮH 9J2JTr"Z $iA],y*R(Q* ┦X^X 뚢5q)BURnz7)^s֜/e !ywij"C\jD:r͋\QMbH3+m#IE`0[rc=vэ>f^f )[<Զ{}#I%(f%5e3ȌCJT  ]+t"l4:&8\`V^"(1A9&Pn ( :B'՚R܀}z^>eIGb+|rq~͟T]I9"'9_䤞/r;2J18El >:\OHl7AGph(%}~\C̸XN趆)Xp ?zWY<|B9ǔRS+rXі(Z!uyL5W!BŐJuШyJ4Pwk|Z8|]e|a6).*n(iVɁōDC|VB4*Cn;/繣ߖp^MٝOhTV2N qJT$(ֲ@<9KBQp n% ýw2jP2 O&׋2GCdzˋr&0E ̷gui U!!;v0-Nw?.K)ȒoFQ'l嶀خ^+\~R9& A2<ڸSvk.A*SOmM`h%/x| 3^cHᖷ9dHBCt5\wpBPC\wrC=g٪o/{PLXwLFwa|uPDtFdÂc:3Qp IyPM큦Ƿ&0+iuE7s{|dn)%l1nS>ÒӍ_K\T!.rŵ<@5ʁ33/ɻ48$HNcBZL &vXG`-"]dC@o%ŤUh f* \''p7EpVXx4$I@N 5#T[/õsxOJDgQ h8ch$@@(J/0I4 oP fe-j-qI'2_OpXy)sxmrGfiՐb? 2YI؇o>S{`D6Ei/>aU=/;onf ~J!?_o#FKsy 7> l 9!%GBq%:[%|W<= rÙv#D-Z")x#a"TAhqϥL8MEE5VQTE;ՌHS9̯nk2BM2yM6' P<[ N1Q6H{w8|au~vƙ WP>! "T .x! /MEXƬ 8 x =H~XIzwO7',6&~F*ˣZz\e^bmoLj6|~v:ȓ]֙٤4>Kg~S)oNuzv;no.1F}0= ,l,tumU'6jȆG! Zf,^x лw2y[4LU?'?Le 曻ɶ) B% ϫMSݶ[+%[0rduͽn0vN*ޝj7Iz'RQ) TJ$%+B|-,la? \La kxEZ1beon.FBJ,JI7,tWr~l ܌ht)*'L`ޱ+O[/Æ90нxe꓌V5c9kS< 6t{fvWZܾl:[ Yx_m&H#_ {ԙxa2l2=&1ve|h:{[n_G ft'njޢ- c%jzsǺ̺̺̺:MoBb[A𽰒 9h&AZ47RX{MoNԩM+a@S>RX9YTqKtIINyW%J`nmYAdk˂6uw4/o)Ѵvwxn}(-m_o~x8r;?7+T7ןnq۷j N?K{ڛo/Ofo{\}7:Zdf_V7: S4i"sCMxZA@7]fY|ĊV3t v"'&RifQ\F*$Y R!ra]]rkxPzsiMD|Sۘ7 JϥY1'ehqAE3;6'Ph$PNLpJN B z -ZE-4VRI.&)%X\[iUDrfFhR [)֪pV˩)q z^(c)܍1qrl 6PBPG҉b8TO 8hY1ŖA0GnvFU7I)GY򪖁'41I.Qa\Q-| "6-!`= 7[x~NVɩIUp~׫Q/nUr(c=TԅK L&D[tQ[k'@Ɇ( 6{Y(%@T-8S-%$cXOB V`B@i̜'и]USPhQ},ߝ_C1$H~wWrkax}%O?/t sLr#Z= 5-Fo*S곻*(|u=~[Ү>K2zӆ t-7|mNZ㱬NZm8H8c8w6Zz4&ԂΠnhϯ?~mYxnO;ߴfT]kvpk~OfX􃢰5pmB{{\oݽԝAR.%PyMfJy'ڕ9L :T)1QN  )c&4o #L4J2t!#$5$RO.,λZkZm|uy2w.Q*]^/o!fK?\WD;]SLk}iLp6d$[- JĦN$4.(J%^[=oJ }JteJzF z5BVO}T榷bƍn.ݎf0͋?e2Uv +)vԛs@ ͭ$ ᤕцTIzڥo }JЗ_qftʕb WJCpA!jr^BЭˠ*L i 'z>JuUoyFd@6؟BLDA\>+ʊd5&T,NNdy;vRznӞCEޓgsR%%VX\,Dt|E>dK`\7-9%2E-P-IIpLxi$'\y6Ql=ߴnVuXu`oe*4B,iA^_GC7fvY:W+C9 BOj^'M)w-A@ɻhmKOi9"-)LE0*I%1A ~% @'E'PXJcb p}:GIA (Ci)\הs}-s5(<8VFVYPx "o5g=".>~Z=hE$1f:h1KDG 87nC, ѧs8WSXU(T H-" Qp| m [On{\'/#|8]/bІ9Fme_**;0moNElY+gt(1@3R6Ҋ/LM-, &чmi7UѕIFTM%jz#GB(<]^b~@&(eqtҊP(x4S1ЂOFrYrAnk"^nQFjN0 zIe I̘P?LW/Oਁ*ya*J*4!dLZh#]hڮY2]>"7OSC^![La\}dZԭ,9Qj]3nΛ w9yQ<:y׾xCi!<AȫD*7ח+1Z/z&E.lP]R%0ǡCHg{qC&mMt%݈d@-d$TO:chnzSdxAi?&C^yj .jeX-|2h#YkLzBr3V 3U`CJ{*AU:rR\[Tv :u6 g>lI=G<2Ȇy6Smƕ9F%>-S;¤Wr'Ck]o熨ax0_*O:)%_Og vo"}g TȽ|?sY/5mrWcz3XOΣircq'{ XQPR7Ή2%/hzD12#.'}UZPwK^fWeRm4x7cp2Y~&>|͟G9CȏG룯{@hT ^F% KL *Y` E)!Gu wo/^4˽\$}=ctq CmxFC<[I^_|pNZ\BR&(c>fYAS 9=9i-0}7@ּ#ȝt+'JiN!/tRcJ0y?BkTr 5UM6fЮ^9&?=%ti?8tbB?|es}rY9˹>g\8 &)#rgVJ0TQ Gx i )t~ -&g&q]ώBo/BBV%I8\gH/Pv_U#J*8O(UҨy`@ uϵfT?Rjy]gWr?|ۯNVb.5 ~ԗ/̌~< x[G2NQp%N=p").8C54ΣfH2p蠦G2F>zi>QhX#Aجy([fRε<2A' >TRc**`%EJbYW : 3h"Ys[&1]*p%VEq:  9d8k`bp;T%8ndX/2j$0bQ*%% 0JK^+ X"F6 aBBWNX^(UgvM[(ao[ {`jNWG~p};b2r`ft{gm~&g4y{b+rUCz7ZG *}y`%J{8Z<A`7&n>'t88[Z5o=0[? 晎A'EQ$#`h=A/XSrєn)Pm\F#?H KX~/JMX%O!Lv[sRhp-X.zi~ws+l2FT7`aaA6 1)?}oؔ,DGf'ߎϣ˼qyC†ՊG$_~~{e|;sto^!?\7sT9X?]]:pgrRǃ79sA UGJ\X3hIQ3痔P%tы%jXHYMyin^v;_B^]uɠŐ_W$Wq l| EME1ו9E@}ʌWßWsSEَږ-9{Ɇt1sǿP !a% ř"=[x5pFS9U@oJ].|}, vw[NԔVl'߯nsm+`ZJC4Lwl$V +/5(vb#4㱺mÿGpJU/,$}-@li {GB[O 8*qCH!9EOc6<bhr SY`).ca^ux!8!TͧQ&C Khg95&[^6t%nj梭TSč{{=\LT}]Q&@ޚm*̸/ZΖRpa~,}OZ4LQ1e5Ղ-OEu4%PGF$q[ꂤ{Ad2D: ۔DipɌ_ Lj T QhAF#1I4])Fo,ڰWMnxt6)Q ܆B Ț)TqA!參yssԏ䭱^9II EK [ph9Cd1'r|LFswpđE]j%Ps@X'5e0"#8+I,.8|ry$(D-dwl])*lʧB Km0h(^ؔBTVIΪ舍pYj'CM:޿S\fsPb`(0MġP9(|J; \ L@: R#gh3SRVAEZh`S[bg6UJ n-A 0`큻 BT3Ʉ10Ř6TNCA, 7DoƓ8mĘAƀ*%1K#@ՈlMbo~"Ӻб/y;rvt|FC"T,bpHQR@5HPjϭdcdvni:iJ*&=u)ۋ"b"C2Ba&ASyo'\rOA3Fi"hL 0640fA.642852!Q/ƺlo ;fJ#W ]&Ti9 6M˧ey2)^L>h|Rb gC'=Z7}D @&.R %ox>C)K~,xζteפ#&.+PW-$g&K}gat0 rY]boP"6]3nb&n>'UGW8O~{5[)9w#&e,Pv6Kظ*GNۋh{w3ߜOw~ 67vr\'e>Nn-(}mrq;WZWS3 $ġU B 0Kѩoj(Wf`s=]hU ækR_= ` ?gT}vz5OMk/"Ȋ¾u/ jT^QWSN=/[ ՕLn}No:׆?{ϯCr6Rs %aqڋ0SJ9gϻR Tr *"9Y1"5?"O*9%)+ybIL֓WggggUG$x"1X)@OI)% 35OS}N/&l<9Woy\O2Z&|wC (SFΎ 蕢b-d(L+C&5q8 :XL $ey)IBlR`ܡ,\>R.-KP0DϹĤ0{ 4ɚ"]xNBDmNNv kBUW˵j5/3!/+7l۬&$4UX(^W c! ^\Iwؼef{Ϻwi8?tuFZ4ſ"7oXd ;t$ ͡cHv % bx  # fz@XEk-"xQ!1\: 4eZ"kz{:4[;2$O6'*uJ{3`irNy>ZK!d1<>w+uь曋Ɵ6ʿQs1~wRiP&Pj(I ! T0),nxn\-Z]6!Yh/˰S6E.u6_O'w>^^G5)Waab|UVd*,jAo>,t~^,tzJJ`Sl+Õǟ->}xaEw\v2/4Vq"|B]5Zv ᖠ7هLI|OCc˔7rʛ?DC1 KM]=QI۬Zר$=ZS%eʢZunt%m%w̉}t)Alfgk>x w o=j͈Ly;Blj%tIdd`_qt_OM0n2qPCPM7:8J0ou p6h͎͎qjµQz b$[XPb8)V Zj(L Pc")`A@w<83 @v,; ExMt;$CP@B/9s:8M6(xY!6 0('X(jwu!XDXn5zjLhҳ0@-P(UDxh^&B- L ܲVJz; 4yu1F1sXq 漖ÖXآbm8`KK=( ^ ws qhz HVH0(P9m eh [Fcc !adrwK$$yBJJOVޒ] [)@p}r0@L1mR1dȫ969OOnqei)rXacsc|Xzwò >1?յ~XVG[AYlXjᰬOusXVCYYB#úkN{ۿĪhSp8yjF\&C`+^5V̰q ɶa; :L @!uvle'aO0͹@ 3ZN0WX=ه:⎬*䩄IaWk6f~S|۟sLĩWP 68N p҆WQ9kHb@I\tzݻUScOΏǙ@Z0Mh Yԡ~uE2rx7! Ikϙ=Q|tij=kٖ1wL]t&LJJwK5,Cp5yd-4XN_ %;zw.br9(PE6TQ2x<0#g*i-(+~s[bR^z ºMԼW:+ gVGwuʣbt٣L8MyW(={kʽ|&b;Kq d_e?U)١2' 's4,{2zNߵ#17@ЄVSg{K_F"q{7UIF+O5VWBxSl6wc%D=oNL^=^"`|,;3)o#_/WPunj~^LTX\Av&ah!bO9_1dy4Rfy5ʡIjy.t"q:䐨ӧ(;ȁN-|Y:4 +s2b;!Xv˼\G)N O_~a:/ [dTёUdӟ\0f1T#9/3C~G Jw8p5{pc.w@0d#|H*͍,?s'7y]rߩCa~s ̗92AAn~'Ƶ dՏ[A# SYOhH1TӫŇix օ,zS{@bOL24=71dX/:j#/iFX%=\dUCv_Y:`y]~*d~\^yo) Ƿkʁ[$1d!c9]xo\5&ãHm!"R";-kB^uVba8:%)=ޫ<0HY=ݻz 6=4NGPe u:l?}*MZb =a^{w-o"!)i5tgCGznHp0YCZui@x.&+PH–5ɳJX{QMV|WׅZj39jP8#PJ:f&JIn=!Q^q=U ?i8F`!.2Y͔"F%cK--Z>&Ʃ-VyXҘ8D#(bZ>%-N -DQi4aR2`3(KR GƊdDV"^rכvRFVG-M{d}bIgGi?wm88,9TyIwU8mNfy)b<5y b7oK~8&lnvE7 ݄?gkmζf`ml^ŗ(r@ƁHVcVJ,JE>gz. ĒogeJ1*tˊ*+kBpͱ)籧ލAx\ RL':툱Iiͳk>HnmXnY64J?,/>Hx)ãDA3QiePq[i!…@QəF\=CQkn%i`Z O1HՂx\-JkMŞ6NBSG^&m6֍\ze`I?E2ؑ;g# 뷱+~|@W{q uxCN[͖JĎNzĀ ( K4qt.cھ/=|~DB\M`Dfk({ٯX\p}f!5sw9Y,|f#Zl'Nf ٛQ]xO.VܭZ,VF,$  qX2=RVHϐAW(*a<˹u\c~ _JU,]l?QUSV 0_TqTU09P Y!o+y)iL X,vbIq96RY!aP(OPmo<'A){`.ydؒL@Q)F!} vG4W_12d9Fyt~, 6>RC1xC1x_ =o0MNa"&2P3!SZvajt`a؟w%ܯaW&>J6j. l^F: qy'6+D9 P!q Ή;TJA1- bkfkN gDOvkV慠%nCj,qʫt؋qC:F)_X)IZgBBR`z*m:Txp/14C"z]1wcR|zx&R^-8Oٻ3yJ ^Rfd77WX?mx VTT/ ~8bVۻ4̋q+,G*~fo׬(Yl>m Pl>= \q*j>8x)C$J4t&f {+X(|=u*N@nWj"iͼ[PD>sI杗P`B>߃`Hp1(Lgzȕ_1% dVy yY x]ٱ5^!53vlղk[Rw+XEV}E9h ^yn/lT(^['#ը9!,2ZOZʤT`{RL3+1eu|R%!7Zo6Pg"UP$&?Y ^!#ōȸn`f;]d 0ku[Ӝwwuz]_fCNuQjضw;~$*) B;,.QAy:l?9tx6<(xˬڭHFrgʻg[сwC+&gagՂ >7]{,H3gsCt.2M7*_$kU=6͛g\w".a%({Enbȱh1-9XF8S`MpSӈ욵|Ԛ#~)u8V jZчf34Q*JJ!f7>EOSQJtD#| D%Dlʅ)5I}4g,nl ϰ):=F"޽>Ja2l!nTϡGyjJ v!0<br/S;unk>uM_p`ΕJG&xT>҇"ifClEu5裒scm:0fZ\z̑w:`Cs5 ń(rg; L^f\$)R֗>.;ɀcrˌQ{ȏSI0 K1n|Gt54{%̐o|6)cfo[[+5S}9Y [@k3)N ޣg 9kTt; #M:A8h !Į " 0j|>>)EfW\xBe}_WK_2n1dkolIJ Y2SKYƒw-Rg"V:[ʐZ{ܓBq0| 10*_m9hߢBvpU*6Cj6CiRuʧrˤEkĪc=%:q\/dPV_;PCNPJrsէ3Itǚgt뾾r9ttXO[:Kt[|G1ʧרTϹ>hkVb 2- 1Y">PEREm>vC*_&dbV~m =̂TҸ)ZW*?NZRqI8 8⍯p̆kr2J7iQ1:?׵ԕ ׸MIbqxtWFh2xSp])^2SE o *oRip^z7 tXm.@ $"I*%H}^هE aHNU9o95Mi Ux>+nDWGruHm]P8P$3r yA/uC1F 4 :NX _]Z ;_a\̸ø.(سøPbdL$ݎc[J^g/b 1 H3'g_7ŗ)Mח`Qj<GY'*dp0F=?CZ-n!:MP(0G+or&cyC&z ({%J2{&cvl`ؠ˾yn+ *LNr{J)/hxi \XUm1@cyOa-϶U;_᫫v͉1r " Y?_tpP̟mAʄw\ uvC/JކilޭK?~nWk ӻ"/fO>ݛ§{:}}RP(Y{bqv׮XcftF&?Y=_ݟsܛ3*rn1"ySԧdeHmwoV +-TJ K:!ג|RCg6e[w,>{׊BvA-5a!QyӷbQd}}>?U<׸x-]A>>?ëׯ9[~ L|ƨegw|Tc^%Dwy-f*tی+]gdܡK8kܟ0EW!#}V+ixij!$+!.e)\-&}A5)Tﺡ%(g:3\/QI1Y-ZUxȰڎ4c19냵^ 8.5nkw 2܀nf;l5;t!wdZL,?r?<~o@uYZ-iEޡrc-,JKSy3H"hxRq 4g3Y: p.\;\9ܰ 46!;*]K~kMD&(V b+ ȭyA "7FԐFڬӥ.ن2 sE͖gAwۊA h;ݞqdkY5cݶtC̐o|6)mhqg;|Ψo1CQJw[28* =-`ሗNxOKuR՗kxFADN'42N#52ByVS*mV}ERu6Sv~8PL4=*8\o  j`?.6߇97|H.6lMH~k`_] S}MƄ0 CE)|A<= uky;>'$D5 4O}!fF \<=~e_:g%ΙV-H2 `!X2(<{?iϐh26gjQjHxɅHq>a>Rўw\z&Jv킴2%?=EK 2)XT׽$U_7#PL˹XSFQhT-R\lHƿ02̽ǧ(XmE\ O!7s]Ȓ#42=?ݯbyv3xRBÄNE K吚z)n[,>RgE53SN)rZZnJs蓖,T󩶾OeQ.ȫ1UTT@$ ȓpƯ[ H02| QD")Yj ̥:Zˎ9$wAPKR>$K'̏d?ӽ$ԺFq3U^?m%& O&7*`$H')f%%hͥ^=$4\h5{ ajTpŮ_JE'GDn:(D(D?]r4J鮽_K U~D7Q/k-»@J=aӋox`؉k޳3KE]U*<n`l>c/g?̪&}cV2 1e1M]b N^=Bzъ Ýj$H9W^ DxWgZ'Rgqw"]݈#Z4j.g؉:1m>INxsmx_j8{}#(םv4xu cfk[ D48ʎ`g:k3YYJRCr,wSM~7et_7K30\rЁko9qT%M RkR,5eiY .Ĩy&ӢR\\ a vt"!}tFE>_IɱDa!P{}nn ̫=dw^V!vE3 $/V33is~͆Pvv\e+p%Wpr/9ncKɗ&u˗^ Ńu~b_*vY|!rҭgb&Pec]bQƷ;*`}쩲;3/Bqob`fU$ynY0' i+ \{Qa=H*=ܦ[b͸ƃ,YHy-ɹfq-EtcV 1bmb^Ziغ X͒Ik--6ZüMJrR`N(RekG *v1>m؃LId34U"$ @vC!F 6.ZwR!1C6 azBU~t6g"1\,r-> :5ϛ?R @31&&ӠA :pIQ/kX 6kYUu4@Wevb,%j9 zںcM#v]E4)wl6Y h:*LhTOw`Fc@`MKyڐ JK,h%.QL)H\-:8T`&pVN0,R}SEG ɮ\UjK@hlXLon:?Ye뤠.ߙZP #[`aX !"op?gC\@h2:@ll:vdH0r56}#i>6va嚱&z?'zWwj2]\ ;}{DD:,T GMiN-{pvm~"fTgafř~{od]cLIA?& iNHkH/|z<#-hZ%QjݛgXg.w.~{1d%m/ Xژ S͕4Z (UTv+[RbN뱪zf^ንvNr.v=\h3r*@Qg2D&Z Ќ[kT\$RXmR%H!JPY ;jbb҂舺& MD^(u?*TDnAox3ZES@q^QtÆJpS:D nLL?#+got\xu5$*cnŝTt2?^MwSa- }N{IGlۻRnp.ך\,sF& S꽦:.xx% o,w`W0t3E}{/:$ZXqߩ1 ﲀgł_cxˆ' KUF1nQ fH>S7;C^M/$aC1YC0#JGѯf) CbMaYɮR P~<η,'OQ6$>.=~G9wQtʍ-9̀jH].?Nr P-VV ]zkAktF18Ot,1OO~Yo',[uRhdSJ9KI&JN0܎w}_U.䞕S'jdv{ 7 bqwwV^?-z {٥ogV|ā=6r- D[K_lDҏ<&:v-b=TӦ5`/:B8@@ϛ.#j@i":,l_if:no0`5'h ϭ,߼*T>ܿЅIGb92,,U+k.w0F^FPy5|'Μ}YO +'[O.K,r'.12ɥނj͈UC/ O0xJ! -?Z~T"l{LRXS]b>ԧH$%yLÃ=&$䉋hLi1bp-щ}F_"Pޙv0ݚ'.Q2US?.Z]".Ԯ%tR]"VB`3} ^>k{K4!!O\Ddv[TkR:?k;'?UD)Iix+{ZV3E^鈠]\f *ubdFJTƗSN7b gu%pPì>H^)m2 B|VĊ`>: 1Sҍ02"7pۧiKy~v4o"I6H/ >yg*IwL53;\O!Waj}fID)udI%,!{aik5q'v2b‚:L(MxBRڢnL%,JPCVwVfp=c1v>N@06~A`@ y"#So;ڍKqAYN3j7 RPm1E'.7eR hR͑EJds1%xye-EW@i76+L{32*IAuo#,LBǂr̜Ҭ˃㻲q`OW@t;5:t/CioZna&yxiL3i#+AKJ6fHRMfE^O|\qv? 9lZ8٫\%ϞN{ݸקi{tguRrٜqPR7^gaƺϖp%>Gxce(YˊěgmbEN%>/V nJ;,m6AJRN XN*J"dpl=BOtr*bѺm^ u XF}?J?_xFB?OWS:N2@^ϣ.yԥ?Q$;bN((J%Te2J*: E)!$V!f2G?yŀy/{΋}hU6*VvIE?h9K:] 'K Me'-rUjJR(Q-8GB+*GR ncKϐ`VrX)eSHWG =mb$xs0C^ t.+ 1@YVG|nǨhtVcCgt~6f7z,3a%(g;o[{]3m3LK"(}R>ZPR$fNۙ:W{ @ 'eD*Ҕүkj[DM%iVߘ|.)Y5Т`6?D!L.>Xc j;BO&8]q5ytnObjZh XySU4tG?PLAavy茶 qP6Gڰ˿E%j ԛ W,idEEp2AXyƐ!WWla^uȰj,;Q6bB7+OxSPr½iIS t'ߧfYuX P\ooS[zX!hx]dlM>2&7Ww靈k*P9rT\sTy(ke/+.뜇[ykf"mU)Vڻ+,]Z4y3|wzp{a o=|s i6Mq2}^{N~Osoo}}xw ǡcGo[Tz҄"5Q dOɍ"Ն+a%ɨN2搴R ;LR$,c pI$)_ r9p ч1T6K>uĩDq( &R4@?PjXJˤVe:u֧ ɏZP!]'LhJ [gYf!?VW'<-H:pZ>L7s=\\o>L]pȱa*,{7OG.O@\p^+̲j|]L-s V8sݟǭ矂>:fœ O`iyY ˆ+#T+*iJ碇^nsЮ,J>б".pXHsx=mvD0Ct\HA{1q4aHD1d]'x%6;xR~8s@,PG`c= i2LE:DE0 @F,D@e}L%03$c:-syq}5r`ΣV-bҩnq*7͓EEf̦E8]žvF7ާ=#Kׅ |Q >c5 dȫr LW7 "Y?~ բU++,xaYH9geYJQLyAi边>L}m= ZPJ{6 U 6<Jdtʇ>̊ьOb +Vޭ[!YZ%nG3]%6u =h6/w  n㙠-(_8i,}H%`zr ;<lfk&&4T*P^^ p;[&' +.khnN& N?F+w35S)?n\]yp)|~?AەO?h U !b4wMocIsV_Vjx霽hej2Ip8;RDrɕTʠߐ0,cȐ ]{o'Jc?]!y-Qm]9-.,>Ac3) ib]?אG/yV69c[ ubFs>{pj-K驯Sz6WdSJb[[ r*5ΆNWTzF"Q D ,H >,J$i-'UF'0t'Ni$;CMG(05}1Bz^TBUy}޽zg]HzR>n 2E(^V`3y)[pY TBv Ʈ&Tg^hDrT1tУO mW,N5ڔ0/?e +X:mqJ.JqHe}:Jv|,LK"ݬ/|D,0 2*{NVھbJ@ W>mw@%,ӏ+'g!+Dh)<CJ EӁ.j o"M3|a`T GUZ W+[Sŋ7]g%_/srv(<E.JYGr~=H9#Aq UrJ(ڸdgi5gm]TK%f o\O>nMM{/^]$t ]L]Kmp> ci*o~~> &jK |Kb0xZ1~0ꗰ?O`q=[]m&z%cBG)B0SO"3qýّENǏj)YlzzeB|99t.(}b>?T-Gq>vd6*bzT@ V?? qS YTG~Mu 9MlW uUgTlp"ؓaᠼ]Em#Ci)_$)iax[ERTGh#Q1BX+؀hE=P>qawmvy@i]|\ ieo!|,cȐAu̜TB+_v z6DMʺ87Y{CF[8*4bLFRn*) %UXrdZIܢHaϗƪW Zѫ5)8R!c vQL V6tyLK`r,RL0ݦf6֦Mu?qr<^B]nvo[׋"z)E.M3ݏco3uskťrF^2 ]Ȩ2(! n1Ncǖ 𘙨"{Gp.P-x ‑xd021zo")10_Jzn, g(.uۄf@0wY}_\"tI!O+R6y덶:2L)eBp iy4b&RH5UAcTl d5J'f6eR(Nzi6RnfBHS# $B r,jc: 0asBH(Qy퐦]f mMŵj3AipƢ(m/se0Â8hESh;fH*HjqҲ-(4&>8O 2ԳDqZ:1rlc,@h%yγnЕ@_36TS$Gdr!w<q,bc"(C^~@g"lKl% 1>w0A:Zg&W hm6mk) /1㣢$w^V5 !}2%C갖9XĖBh|/_f}CtXBx7mRwE:Ç>`duʞYݾmo# .W0X*Ijl.1vfʎ\=LFKU8*"}ܚԈWabaV|Lk7vFDj=ܣeޅmG(b޾a{}.흁h蛅C{]i̬L$"1ۑ)!]˫Sm=WTЫ^0Q1A(zK#9ˉ7Ni e 9Q"8Ri9Ҏ&,V"y_1Aࢣ}v62Vk|$:-& !DZRÉwQ%QNSPDp4%80πv(#tVz0ux=aw55R*/FCCp|G8"m0zI04xTi\?1"icd%ndLOäBCKI)JZJPGkJPLѻrkRS+ڗ)Cmƥz} =*22)w]RҌ+1KvCWS9kXF,:w<\n[>ҹBvL<8Nʆ3·ř 1QG$G+IMZ/78.4EMVfBQ+Z_>\/{̇ћ^$ȇzGզpԽ"Uc4F%[;6N0ү—ru>~R'!Y=jc"jH[c筧-o6N掸L=>ªgr3ڙ!_m2`<-jY)ҭ)9S;Hy>OK[J6fW>Ek),⬿%5c$ lǚy6fqLA{#I~RI3gU]cc.Y?3H f+dyŁ#68j+,4Ĕ!WʂxQM( 'žH5DWg!9O`JK='1e+T +˻K)U$Ay!gZ!6S8VHV@!Urt[@ aC^W!D8x@bD3'B));\-t衚II!4#|嬔f]^4JȢE48(.Z"w c"1G J¯dĊp HMto4!RFXJ27b F#)ibOi^ $r!"h8H Bιl&^@NR([:|рusj}➢DERϫC8x(G,)A`Utŋ LyO_rYRmHR _R%h[˼?$_t@&70m@1֚wzyb>"%y:UqNМY  [^}(S"GbJE).S8T*$Z A۩˯S,Nl O}xI@pejEҢ+"T8k]a(210`N!R,>F2+l`2;"B{pu&:!H(۽zX?3[} )4XjR0k"e@ar,) ̊EMuQPn{n}?׿mɉۖG8JjqMa>8a 7c:AK匼LH*Pti"ʠ@4':z`S<yD 9 m# `][6+~ٝkU);5^))Iv'jIѥVZMId 6y}sp.I+@'E[%R7̈6$jM ,3ٮ ?q8:RnFO?jn?"ڿMSu:u[fV0f4߬kcA|^L+|!8]c'lFcV[ϴ-/w64Έ 0J$wEq aK+;}?ݖu&ĨRqw.ȀJǍ,j{ -BD1&i(E D:v=g[S\۳[ndnd$FvY2@ۜqƎhcQ@_ޞ֦a{{ 0@2A΍Vu4RGA 8 \*DPɨ0J85¶b\5v6Q J2,) \MVl )4T3QʹHXTG5CXH!# k>54- N0Y|bafw)2NJ8$p 򈍀#Q>!1ߘG8~rVԶ`܃⾧Mpn00c^B^ n, ԟVb.(lnbTԈ0z_;f) Fp8aw.ĥ5°v1< RVyVD\猚w @\4刈ΤVL4\Rn#%9IΗ T}fKs(l>GΓtjnCN^\g+q.e7\ deuB5"M!`% +q̪49y"uJsB?rdHHNOV6[^?' 4vdӈIլc?_/zLn}nc;}_ ǐ&Pmk9Õ)BHKRkX \D$4тHiەS)Jxnx82%$EXS#2tg5J c BX `6gNۡ0q4$)p®!fqQ%6}|BFnU*&1"B@wӣFd4}o%6LO~qL'@L@1̬Ԥ+5 kx2d֐U>#PS81MrI/YXJ}GrVL)VbL!"T"+e&Dq! Bٔ e%f7Ѽ-9 dG:zQH[9^(o 7E2T\8:($ݻ'-̔-}9(Gmsȍ:̨߬CDE v.0|t'S0MJ0KXcch羻"idAƨ9r{ nbcv:Ξb4_.>N݃#{wH֟}Hӻv[Rpr}4xY ;{n!uv/!ݴ>:t޳?E90R) IL`& @J!a`BpB  )L@;q7PL6;#пh]aa3C_̓RSjE%].s}|1/ŏ=~?4!wȅYQ%>Y|!Z|b/)%J9fsTB}UP^ F{2}23y)6li栀{˲w۟jh Ji6UOmyj,쾨N)OPҮI.`)Td IXf1~z[+.탋C޷(*()}zެZe9 A-#ۈ*-j9򠅨hiK9#cz\R 5(޾>_#&Ge7[Xii"@<#I1l uJ!LR}V)EO p<E8P]gBO8JGqs\|(9W^KJ~9QYo/k-ymL'.{ a5ӗ@`IiiE. Fџn*I  ˹1v`0^YCn3)<笀7a "~ņAch A@_M^'yjc]n%hw<<W3 ⬩X}ڤ@mo ٫y$@a1WȅZ:BJD0j:>)z^ǝqI?mDT 4<(/B813[eR6홙23ayK٘*b)@PU$ML0[ mj$T$υPz=D% Ww o r!׈o~9c2wF NW* ^Aa8Y|0D!n8L6^p¶2za׵ppm0H9bcq'k Cn̻+xW}u(<}] oWxiSE -+ Բ.H:}&=a{{6w49>j1HǨQn{٩}#n{8诛§}ه; g,ROx8 ރHX0,ʆ1;gax7w-n qzP?Ҵ9׃okeL+}Ԁ@O&QA@큈s[< ʆj^7܃md8 ԾwN5Ծ;f -Z~Ά-t/u;t|J*bA˩Nffh\|.~o݃#{پ!YNݧom:S+ݷ^ 峚ocTAl{}""vTLs9m?GB%.=}QRҏRc\t"; C'8Ls;Ss%B0TZ-wXR()%F.%\є%XXz,Ac!cTK1ʚۋIX<"𫇬c&ol*Vl;RKZ:HƥD#njJwU":z!퇘YHbf_1TDVGϮ,S*."_; (X h'A7]*1 !WDÝjwX}Wϙ., "qث>꼂u;  U;|BrFW m.syF;\u'.0"xHhH=dfvT.>p$`pFƒwcD(83xhM*p"dE.^}֑JfoV^PVi%/D>(Fb@ m8Ɣ-65~,4kj$ OuB~r9&szRҢ\~[חuia+ᡒrK3eHQ@(+ϕ֡ TCRx-OA2!-(~\mOqz{eш>h|bEAha!HEG@\ x]96\S'h<$!F6*ixMk[I72rVy]N/cS"#0-ru󾪻+UǬYj)RWqq 'ix|!/?jܐ\ thY5].GS.pA88/prP\.g}7/u_֨U\vtsw{Q?i'f9LG{ҽw*JZx5{qpScnDrvWNXG]'&^J0 81QfT .I2Q\p/]EA>Rf9=WZ8g@[&z-w*`tp !HC0ځf)3ppQIΟLNƛOOn@᠞H.U*]%z}79jH>ovVeOlm&lC"J0Co؛ߟTM|~D7nOXO$'_d:WWO[:9߹WwnP}w}gWц|ȻsۣBStqz;|[#ZwrQvTS)1O D%s1V(U sAϓȔhܙws7/"J0V$-06*BjyR˳dztNwqr5=?ɔߡqs?O90 |l>l h r9L"3+/#:in+t'gE3T"A%z\ v$ Z4ZW6[3fՏ@#^/6@Iu`~jJjԇg,A3IB 2"3BrFmbZpKB4jb8.)D)shJaA%.f.20! V%QBMSXz%+ΥBhNeb.,ΡEI4,$ЋNm&k:9:W:%I,A۬eqhjPBHځ\a8bNd*B .Nsu>Z\,BMNdR$#Ww4oN2Cp0Ol-`"WXGnvU݀nA)r:Tu"}8~iq!_griﲒi;z? B?8nlIcaIl0K)Zrj-/,r|sUfr"dI+tӠ0@aqmh@uVC{[윒VP":*UG,pI{ۿo[quKoO:PL?@:DXO&,EmdQ&VF>Fɢ/׬+늪$/bW;*cc$Y]|#'"%Lުuk]Ԁ%Z .vIݬF}FH+>kRH\+$X`*Ex W䣵܄ P= >[rdcͭ݊#FVZ:!kAwΙqUޒ(6:ǵOFA6:Q]<[[p@|UCT c({Qx=SSʉMHBzbs|pFYvCMӫ/.%W5}i{%+[8ؼjd7Yr.K#=UYj,)IbS(B&#,PAe\,sD9٧.'A(%IHn\>#Rc|Ij>UTfC]%Axp*.|ty>p/S֡!`&G"MHdTr7ILP)5il߽M~޽YZbKT_ݼx*Lw7F102ƻ[?ƙs4)G~g':?ƼRӒck ~<9[L\29m1<,7nMtotii-&gz]sg@W`.U}>V*vՂJISqZc,SGh+ݑmQ!f/[[ JDmMm :ozڭ EtkBӍ(bd_WuCCL^M/~K! LUEui[@hk2-) d8z:yxW'+,!)jzmX`kƘNdP^@-uV1dt*"ȵ9rq#swF=B.iqK|5Ħmn?n[De/Z\vMo,uȶ^_d+ڄKE0c!W̰CkյOOkyxmk~)q UU"_|o\\MFWw9ʔLT5madI왆^0_UYxnyH:+6,0ݛ@ΉdnL~.ڭfh~g^m.m=,s ( NC#H&utNPgQe(%@$*Wi5OhPZ}{s܎ h2)zBJ8AEB[^$2dHSb{qԥT!ƝĈA6l'6rFL!G|R*W LG$M+)%48P"Opo.$362KޯKM )IdJSPFvTRdSI k"daZb)N ɹ m4]MN~O %1Nߖ27҈1nQA$;"8!5'KD'*9=WZ8g@[&z-[[tp !1Qщl 4*zJPUI[ð!byKrlMg1\8qex"[υamAXQPsWr- (f ZA7'}ǤcWp Irg5q6\S%UT1ٻ7n,W~`%rxkA,6$t+[$3"}%$,ErGحC\nɃz{I`ė;#ț.eL?R"=[W)52ȫc-S96PjFscDVE|K8o)h'#%Pn~^j6jxwh-v@ɗ!~noos\A'k9}X搸w M'c?m5Si-DB/J:DOqyxA5٢+!8s?7/2(yC`& In7o˻& 1ʁ!E4, RX i4x#lO/YES/rQ]KߛFۜ9:#C"➲ Na}"t ~v"̮.6+*%x#5uNǼ_Δ.2Δ%Ȩ$ywQYMP"+KԔ$%Ymź 3qcӁq)bh!rBRKZ?Pʈi}{n[(%T `1p;*W~\n(I.Kӣq9,~mxj-#L(ػ>XvR`hiX|m@n^0dۻ$~8+J~|\5iݾ (Í] \A1F' '[$Rh/:V"+ۣcI}T7=7oT?@A3MيzBA4:guΧ1 @55#gu@)`E +<Y#@w x"c. JJBPF ۣG 0Bun;M淿RdʴEnSed&^r̸2}:9 g@,6\y曮!T^/x0٩'O?ZM|kl"2;I䫤WI&_U5yՂN2*X:pC8wt"PX!!: c&FMѿR/8$_M ryE Ohy),ކ❣uz4)[` d0kbt˩S@=:8 b(Ee=áRf*>F !lä " N5J)^V -Fx89Lj{XD*FY:BeNp<2PJ\ T"T"̡hSed*0IBKĥj`FF[hR.}rsŦCʧrLԕ ijZ-Na3~\7_c7V~Ro {֕Fw,.>O. jp/-}B;BR啰dbԬs20:FT"q4+͵Higzs=iUrF[Ig!Jj"NN-Z B`SIOkeWqR-2 +|c RPDBh=c&N[Ô$g47{ Wc8T$cNpV:esoN2e(nWᾮ{R]5mF78\ `Jo]`( ”XI ZJ&-335{%0 RrHh8J95BAz<NzkphCPûg_U5Xqpe_>-z1:*ԄQmp3tJv!}} =^ &;vd;|~bp_ OPTT΋ jn7+ &xkp7D083 %.)auRoֽk9ʿ!嵝g7Ӧ>mm(8HË}3Zd*aDNPaёSj8[t)#h&}BT9(NP@c'i"t%rSt ##gPD\BH2 rGRɋ'UN&LYx~ӯgaYxP!de6Bσ{ԇ94eP?Ϗi&#p#KVǬi?cysIX%#3?URWI*PRRKRF+eP6DE ؒ hBQ[FБ@v&/j 6CU+QB&n(qܖVnU)IxGn ;ˆ?]7 00ͧW71͛qˎ[v0(M5>> ZKhH]J*m08.)%F c>J 5B)3 9"7>zA*!F3(>z)Kxk5gbs["(Cˁse⣛AD\]'yha[+p(3sPcRp1/am 7#b==q;j5u(p;ûXw-𮬇ww=z_xחgDx5aKdT-.rI;;rv,֋=nbipZX,`< '[%K*٣$@l;ȺU9ہB#ӿa//mH%rVcX'X34)AFYɈ' ֱ w!B!Py ?!pcP[cg^bcs&&6r.'raSC,ʁ ViOC/v!/rnk7s$˻ x_]Go>ߞݾ_~MzYv;ޭݲWwq}U__}X{[ 03OBjH,y"9Nt<|tU7/~ڛ7X^-/W>*Z\z~cnUCz Ռ!c:rWYܘP8{s>AC0Z{OJ%+F]8nn*&HbWtTdI\>Z)q#4_npZu,i%_15iLS8dKǣAr6u<`dP?[틽kDj.ǦŐ؉vp fƨmlvgAN c7jD҄XFYm쌭"Y/ I!Dp1+B@ծ+cN.`?!r㒷l78Mpq>rg򈹈{9I0gL)jG[gri!Qu;iD6ix/&[,h쑀vl,8QJ[5q׾]`k9\z]6LΖ#3 A6k@fp' Hïꛛ~_珿 83{AiF@/;#Fq Cyq wK9t`Ɖ4_=BĊ<+pdu;-ݨg5(E*apQ_RO+ >m!?{WbrNe2bqyg-cw"v BUg"wYmeρ T<ʹ e1uLɁbARgY5* ulСiVQETmtc TH"RC*]NcچBmB:){Ul} 4A!S"&hR4Q EJ[)吱η7)Q2}݂%vGY!$pߗOэZb^(~3gFDgcL aW_XsRB݄) scP`b&E3LjR[I+' 2)U 89/OR S@-lRJrjD)5;l$RS/6 O%K)Rrj%khYWҝnpW{Xw6k0pZZE# y옧:}پB~- xhHH7ǃ#E+Ea2 du;l}[]͟NL2?W_!@;pR:w77ÁuhF `s3|8+AW$ߢf+5u;Lm ixԏGj9u oz'$p?r|˅4{}M!W4%2YRJu2AHzBE~" # էA䀠TÞ Sqa?Tc!ex[խ>fu*NJ7w_gY/@<-(5EY⊓ws/Andz;+>NJ; V>[J^)ЈLo.ͳjX+ ~ltdt+D :=MTLŊp<tVQ&bqj<"8ϰdqi Q'[UY+v"!'y ;G'"v^` s|i9*h,UE{r@`4lŊB(5rj#ԩZgtR &rLb%*p*gR\@bٵڒ',eyZQ I.~fd%{U/|Q107#Z!s0hGRB$]:),RP+:UѲnr nWђ' WQB4y Ҧ%ɺ2螔dPyKyYx&5+(3^IyJ-3mKgǃC=?H$SVb9 $*vqTQfJ_9ɶ_I@ WŒ'S.ȃTcIJyu*d#5P 8 TT}^yLuW%l׷'f+ aSGttmۊX -@0]C|kTߒ5GèsmXdruMB R}K7&# C%kS܁Z%ajflL A:fuԈzN $ucJ"}1 FI~=:;euIcSQٻ޶r$W~Y`楊,vAkݶ㖝d&[Xd,v,Յ"YʎhlaxdtTR]%&ZI8ee Za+Qؑ"5/QeE6MvSYV;`UuR[$F,Sۨ7!z⟩^:Iȳ"TK 3^ O5͝Y. ;&wN$1d9dgV8-O%|NtKnJj$ݠ>MR՘+8WF1Es3dybSrODyBtK;9!}YZ(CL"sC}v%d]3LFa兏oon]Nח]fk^~:)ms'uuR2g X0qG˥n.Vs~Uܬ(&]ߪqZ׫ijazb9|xG̩齯9jݻt?>qT<瘵Op{ifjwʍ7B]o.YKwUm]`ӇwUEp+^s(i|V}LC,PӐU4'nWz9N{rr[Ip+#Vj=PK349yI=)if?hrt,H )Bx*׉H\\r.($n;-r︩6U۬w9+5ٸgiNgDMv}eEi CִJꅬy /hVn!"&#տ4i9in[T2ާLfa+*'+ٝ#?U9zǛe)~#zԳ۷WyBͅs7%Wx_]}u}{B\PD.D6I7$4x)'TQ)?<_KV9J^L}7uk͛^Ӫ뷱on?K׷䗏v{~nTJZ7֪>wzO%dq%_,ЪP zDkoaJV"JY2OYE'!M tXAVATJ!a9'5k TΒg u.9Wr1Z lv Yk{zTD/#mv$^D'^QvY&9YjZ,<'?;@"Iv>FQf tGIpZs>յ< ފucvݏ6jNۮ:ȹsgȶ b6[i5EjƉ#hjYĐfv Fh 8$k˲'^w9g7sA] B6 sT^[+=jHYx9P ٳK1&[Ҭː})J)Rkbu kDvA8xA&EhF΃ >L@m a8[-ieQHmEJ6c!@KP;g*R',(Ycg.*ePFJhۨM#jsƃqh9t(-8eeI72IIp>WuFa9lV 3y>QHNyk;5yzڿw0bװU?_5͂Gy-Kr m|ziࣜ XWgp+6t/nxŽ hLl8! mQ4=DQy)Z<]`Uï!oi=(]_O2XN%hd©%7S"}]ŋ)$ؾjDX .'yQԾŗN{̽ ;A£sdb= XR:B%M,NoyJmdSQxɜ svJ>SN]S886DkN'm8Rhd܆SKL_y~UVgcoVwNCm4N}ۃd(Z?ae -Hͷ{_.vhU.zMO<,:Dnq{4KxJugb a& /|n'QPH&teLGeY8Zx7#jy6ABy(ӓFe,Cv8X/="T14s_xtYLP J]a#hT"P}&̕DqFgz|JtO&'mܖk OF\ +3wmB][v3@:BEe@aq я0̞ Q"*ZmdxC<EAp޸GKOl3Fkܜ6]q!u QtD2G\coySG,l\Jfi#(C}mԠȝ^_ѫԍ^Z@^[n]s_f; mq6MU!ɛ ;ZtBcLJ 7Ov.͙FT$@βq R Dˁ,&%I 6x(\"yP[ vM !sA<"k2]):hr FvFKA2iSHr DX՟B ˠdaHj܇muA8UL E 5S)7 ωnTHB0Q/Dz^'[B8I& ჋X8kUJ*[+#6B%@h~Ke5?ب!Vghj%G S&;E,-g⑮5 cl ~.oSO]މ9vSYIUkT;hyl g5qcm'mx[)w-C5"&~Xvl'F8Sְ3aʑ08rEԻ7;VKg4[4(Bs]rz$g(y>=zQ-^z+ʿ-ǪUŽvj퓬F;"̗uoՌt=ܜrU zm3v`Fq\8k^懌5WZW540/`1^v& =8k?%N/UMݥȑZ{>g6@.n|y ^'{UKW7"?׫W[-v O9hue&!ӯ-)4~=ѣ)"D@>v&) ."A#MxoOeЗ ޖ^bUҾadY:n*5#?-sz!RS<~i b8Cٔ[ȦHP&F5yz faTx["l6SL8O#RgR8#f:W\)?rSƑE=="4৽M^l/w0sDȻ+ΌьENwZlC<-vX,}4,w> +h̀c yIh0Yd}Bd(WJ/GYʼnH|ayJw[Wv~vt<}˸ꦶV۟zx{ 0P_QYwNRfi/>z 1:,_-7/~PZZ|鳏 r{k99tw?n5_mOX%zԞ]p\Y&웳%zuEmk4AQbO{HQ {&TmYm$*qWbu)@9 7Nj\pU.(V9P!)JW!MɁZ@29J*F4*A31;rWM^0\r̋?U)\g %>&ί>ūݜIfY./yG.H.B&:YR i"R90S;1jkx`4j5 $Nq]XA.HD +-) \5- hq4MY..p&BK)"Z'q*Z8Z:cs Y`,v)Rz:'/GiaΫ=Qi ȅt nHDq1xIO\UJ*B d=!FF1-]:rS裸.&4fn\Qi'Z%@3@je5I-)g@Yi_eD9g'*yMc9 +'oH+lT B3}Ηu.1rhKZ[4(P֑\b0 ̆ ԁ`<=佑Hvqe/x$1iM.[t%31B1RWՆ95͙12KqMx}hJVE2V,N,fYdGʞS \24BeiGWQTjdwx_\nmF$_X$'aPNUP ݡ:@9ժsȹ^^6v]&Rk7FJQ؁as t 'l~b_GVFZ\NC5) f~:Qա3(C$9NP8!٬ANX8dOǺuWneEkYbºdr.-2lc25ҲmNvtͺH)Ӱʤf w ӤQjcc6i W&gX[5kxkpuΨowoR*[jV?*{ޑk朣7nVꁤN/_LB KH*/XO(,u 4ѣNXz6$V$Vʎ ֪g|Y Yq@w((5us9 U2BxCxVk g/7z:UVݳ]ZhϪl[YҞd5ƢL(v^n,l^cZ?<4;#\Aɩ^G01okvs KZ*f}К9҉;*_ k?gtѬL@8d+?ʖ;v>skˋ>tLn }ѻ#qw (Y\ߖCi6rXyB-h5S9=QˤڱhdL qi!J L1i=U{3ȡIkc^F?+Bq*40i=Zƒ滠Hu. y`V>;38f5@PN>|5TBS*LuZI'j7t ԭڗyC=2sT9r>cdxU^ށɑI;p52iQl`U"bLjUe‘/CGL;5圬{twb' ?92= k4jёI'M&U2if2-oumJpY5YS#3iYv`$X J$ ="-|̒˜D652ҏԣXқ߈mtyk O"s^OӒ)$lHqz@D% #[[kl"_(Ik.(Ì.Kcuqk JwHܪJB۞aS<\U7dP .%c> RF)`7~J Բ.ǠȷP a&9zSJTryv\}Wi婫ǼpU*W.ܞZx%_zzP7,UW_cj&J({}_zoRRjRsgH{u){7swMKN Pn0vRw`̣ AJ`}!JuV]"6Iˎ$; ;h/t+O*Z8PXy Ƅ_] ůOg_7d 4| LG)2jςIeX[!'qlD'bj<$RGmAA% l6Hy nN=/J2*5 c Ÿ:ӊdgv!3ځCTn̘v?y),%5k҂/(+Y7$[ro'έ4#+Y8$/@;V̮6^5g+ׁ~<,"OYۄuC1B[: qGFF0a)ZdkpDǣ) {K*Cr5t=UW7͛#-ٺG&?竄f ent* YאTLd1$klX^-}SXKvHG3V o 3aH;_Ej'Ƹ4/%.Ua HhA'w0`JIE<' &M=x;qTiNmYI1ZII<"?_-"J`tY?ܺug,<-ֿ=^2v"2c85`$$BsbnS0/@]f5??؅V嵠(4c(΄dE(,\q!TWKW|쾍`ujjyAtbKK*u~{1]:KYS5;>bӻ}& K4${OKC> A ALgu&0F' !4]ְ_g7I'JEd*Q R, iN5hxUb E,KP RR Q?k 7.Dl^, ́ZWϵu!Լ#FT*)c$Rl.)(7D/{TdD_)ε}"%I4b?e\Z=ʎ|inw{Is@Apazx})/ Bmp%xKں!9b(e@./Yʂ7 mDUKVS*k?Đ"W%RӺA4WdzKvoոDKZ}2<8jǕQ9-L% uYѲQӜU2Gy͡U$X/FOb&X@y!b49h%)K*ւE% eTȬ饌&j{Q{Qϥ\J-z᧕Z(\R;6R(vx.'4KcFxۧnfV.|^Ͽ- }ؔ|lwܓȈԨ a*r(imS{VA*H TE]pbO8B傠lh:_Z˞x3fW 'Xg?P.mhzEsa?`36XS%2p}8:nQIbnس] a;t}vekԞDŽ잛#A<ɁUhz=x¤G@,UJo^ԀM8ȩWa2ԭ>֢W8f1^1Ԩ-|V?]k?o%NɱrN .JpI6%IAml\K@_\+h]`y~KFBB6oFQ2mS]穷7uͤjM GpAL Jd0 eVЂe%08ޕ#TϐG)ӥGoX GGn_sn{zboPϔa.=mGD6i8e{ߪ3FCOz25z\h}QM1dI/02)) ab(P}9KBW,`;`MvpMz}x[J5pʺ],6V1RD'`k TR>(+S[i놂RX+s6K~t{2k4-h]Xi?3+jmj֧?쿟'?9)Omrj87(L^*CY](bE24VQ1&ͭ/ 7ov.fҙ°V£=}iF:_͘lQE?< ,9:|Kz@9}}DNt+ J -RVrN%UM,dIY :x!9H(%\DcwhDz`^ M )WƤxj*!`"~e!zWS(JBGO[Z+ J(T$`? R&L3D.6ir*r V, <ߺv7H|XORX؅u"þIJbjNc>cO$U`{pmUKN&hTŖ>0O?\䟥X.cê垮svs}]Y:?={rCL2,-JVวEuW\4 !;[x.JjҼ(W||ZJ?p(+ٛ:F1 ֟$h-@*l:ꮛd2M?+~[2͗_+|^uAhL ަrE|:EN1pXMt#3\?7.ݟT E 7{J NN@1hOḮɘ\z h)!4اp4'l:2L i4Vѫ+-0q-f6EYk?wJX|&Hg gt;kvK>uqoҞ@/[W_K@_>[v[ձogu}\bz4r?eWBnfՀ6?^~0Ucxiæjő/9!:۴*BJH)k7}li^m:6br~N#C)q>~̪.o91&8% `o~; |㾼7GZfKem16/6ndQ5I֒jxd{f݂rTh-\fDv։#DѦ:vNRwEC 3jXl1B}I1vo(RwE)Hnoԭ*) ٓuBJF>!91I%B#xKlÃe I K W$ [;^c^FĈJI(z>l)H,_p=_%EHVY;;MutI#ySGԼ$l[󒄍b[/X+SLi5fx-ٮ*=g gUnu]ZhO\=ΓO7?/GOyy\?\= HQh͑gWe&rZd(DfF&k7o'Y}Xd& Ս 2p"!G&봏9dFMndV$7@S4 ]j\B2aT]q@*T`qtY9 \d?Ow̴jjQ@>eq{l 7܆1lR#Gy7$;B16Rde0Vb 2* QQbpE3qx<;xoƨIy>ZJҿB c'XDԤ__J񶗚i Hp&SCFZhz@|~aӱ@Z&o6iA _Oj{ҥ.av|-qQ^q>/Eq]1z\oǎX< i cZM5jt$A(!);w꘽$fstx8J]D젴4kMUHQYj>"kUB!Q%Ve)֛nGgRU 9)-.vKr9#EY +Y_{.+踒-Я9BlJ9ۓ6B9[(Y%Xιb9C`A0 TuĞ1.ʻ9[@Qb︻q /HB{W~Ns{<(Ri EQI]rSFפ9!`ghU_^@uH'$TMOcʴm]|Kj8եͮF)B NͅTtASbrm?#],Ls \o TXc~F}5rQ@~}+IXfوؗ*Ӓ'SFJsx{Pf;s@ q>GSD\HݖhBщuiC5 U$݆ Q\7qIw2Bۃ ?yzgl9ퟰF !Ԩ}3bxo^t0c \8IQლ aLaQGМ ޫοnJ@{J@ єS$*Z}0 G.jpV\}uQ)6`B6+r7zyY?L5Ok{, ,bFm0r2]ת(9}U#|ABȡ}ggk:H7[DC=:nrvvOBX$B<uv/KQ@(5% il0VTO@׶ Uͨu&*FB"Gi]ʂF~CHZgnG37UVx3r+7~sa-*+󬰖ܭA~bwEB5{]?߱ĪvNY\^7n#s H"p[ɢmw2e4ʬz+ Wk;~}w4#/~~}w4w_{wO ^ CBHE;P"6*F/TR@N[\W~ہ(7K{R"Nh7'Laг㍼' uv6S|p*ڪO^s,jI;]gSYUNA-(9]՝N3nuBܛs5C%mS wPA&Mo#ܧGḈѵR))$< !HŘ8AqC ss¨ 滮`qg_ttAj*W=F!uY9`.B ui(5hRQZ^ R uj%"{_Fl @ YE"4iyz_}u}3U"JR5q֍95ޝ7ч\& B:ٗOZDQB2D`p>@Q ӷP)+`YmސfnDqԠ@EF0cބM7lig_s Rm꒕L#R+2gʚ_aecbѱn1y ERsFI}LoHţH`Qa7B#(>B7Q9VEjw|;A@Ԗ2R`9" wT%鍯+( oNhumɍM܈N :'w~%{lŁZsPItA 0Lf36Ëf/>< 狶:+ɨS5YN4SJwrW(` #p%*I*328RJP"2Vpb2I!U`vE*b鶫жg6Z9.LI1&b}*pZݍɪՁ{j>E_{\l}rZs۳0wǠ)ށhydՅ^Ik+YJׄvj{|WA=N=G. ѝKX{o)_z`߯=Dg&-B0nh48""ee슰?0'FץI{jYU݇ ]>p#~ s1OZҽGS 6W{Q);9;`}Ԗ|sRKN`jܒW[98.켵1}úHPL{5u\:+ISz E3=ӌfWb{0KFz3z1˄;N(0'[N+öB v=PZ^,%~o:'{0]o 41WΒ爪R.蜹eӍ#y#*Yi}nVhhn&Z6]G*6*B(Lq/ ξ8wo[s-0*Fƨv/²a0*jF~'&ԇ?Aح"E#՗ppF /2Eӕ;*IasFڽ9N)CG| @%WrF̿~\a\@}xMykAsψf $3^U!wR<]TEU&4svIod?٣M H;IIZ+&I2RcH-bAl~ }ϓMDs0 a˕Y}gӍF 2H_sM\j9UZ,k!gX=$w~M&0|Ꮻ!Z$?o:**uX -466/RQ\U֖krtBAƟ¸e2a'g eSKKxZ ­՘>h"@*ʪ(y|Z(|Lr)c~AQ) _0bo:fGJ{b!znm䥓.ޛ,6;!PIoC]TU}M4'T֡>9EsOh6A4\mj1M 0fjImqFټ{ڡK)Z8Pn%{k|*5}jȾЌ qr 2N &jPdf[I N&Ƹ3FYF9O>q2 6Z8ANqU ڤJxB=-8PiK2#ɨ?F0E@PՁ7 !4W~gM\(O. '.=p.}|z1g,icC䐰aю?x6ƓļǛߠ2PjX׃ۋ?֫Gr&7t͢NI Fp}3v6_>+tlh(!lT(Cl`ZF0~l^G6n'}Ir}W'Æ5 )eBfdi}`ŚǢ~YBRZ& 7K7tܧ,'L1,ii)š35gZ2h(l߆n<#ܓ'D !W27 W ixic#ٰ By3X dKYJLn>{? * (np< >""nX( 2pJor)Hfq\pr r-9hz3Fy^ 3 W@U Hv2jUCR:oP7wwwnjnC*t eǑ5Ԋmf_|_Ƴ/ 2ÈHaHHK.W=|8p %48\쌱5ki #%83OnFd"1Hqk\E2r,uM >:wԁhת! },jCG3PFb(G0)K0SYQ)pƁQ 3 U*LsqXk"J ޯc4C#J^:e;^j˪Aɉڲu(rlPKDrĊrTQɴ,VHl5&h" ɁVCšp Ys4ۼ|Dq#m$^HV݌;| ggsmƱo 3#88<h a #A!}K%oB7rt=)uo3D!Z ="[5K/Bb*HSd}o:.)+IC$_܏w_|r3[\+nL)dx!zag?lxt=)3iHd!~x38 >NpV=<.'f8 (Ahンw@г|:ifr0WJY{A~IwdEHE}8 \J[=hqy\?6q ha[ fv86*<:ȴABR'f$ 21ͿoQ4qd0GϞ!@7uBvf*1C AkPX9lZ&lqH5A`=ܐCd`pC.7B/qlԚc43 Y >KI&e9 J<Ă(!᤯'e6[DAϵU?@(=E7[3hϸI-@s,*g(\J0|z# z(JfD>d޸ .Y&v:*O$v+É'⋍Y(>ErV [Ƶ(2$ y AD3mӎz+L&R)PhD+3"h레,T]גh֙1h:E.>h!__'PQ!7|<.Ph QǶEo;)jz8N[΋vn{ UG8#Q%R:ƻiknʪ\{WQvM>.(jh]XvMі'/$q F~܁]qlbHq(eUq'w, #SMdp`3プLOZ⮬d<w~^{5秼l t Q ȶ1i%mqՏ]kAXkt-sj#uz$%MpsL>rcӐI@$>>do,ȉjkܗ4Wkm ><^'|;v?%7lI [b Rowv1v1J*([3_\s~DrPڅ;T 2d*r%vnw%U}f ! l3^‚ZDcE=,;0#5ےGk­J.`T4ci:G> .&Ju/񧣋)ܤ!\B<.5Hk!Q^|q{rrl`yzWMϾ vS%"rUݺ-871 0<|΍_QŠdk29狦%YD9Yg/UЩL7`406y0:T8+I\TdxFRTA Tu9HьZ펗J0ZvX|RFq51:Ԯ"cħBK]jL|кxG{C;KC1$\LH^PeR3#mrTdYp%PՆTSb[#Fγp=WݚrpU\KRrF= {4U(?)ylJ|KZ1b+%#X(d'HB/Ʈiv6f0m1EqDJ6HbTVeUE4"⋌"౧h*ENc&x,00FiFQlq>2VS2=6=R1 y}|̐]ࡕV҃aJ_~rcVM3+O?K2o L!1&ԭLELszKK[$m8Z_y4I[?;h0j)YSTM㐫pՏ!:сELyG 5\tݙCkA8mȊ$x=nߎ &uY)e@--;!0oHk,ֽYj91,Fc-9V;%VsvVH"Uo,`AL,b%#KK>0f(D9)d'vb0REe!X*5K*A %%,C785%s)[‚r5%*Rro6HF-ַ0=DZ9['y.SMYz%]Wn6HqoԉnW\w薭hjtkC޹hk+6rm:ߨݮ?LyيF6<䝻hcq6[Y%ַ7a}LtU4ojR-2n&XBXptDv9u Z`sd.~͍%2>E;Qe(0^0otatE(8D!w8JD_J'LE^,1a JAť篣ꙠUs4ǗYCQk*^dn|vO$umՃ}8x+Q%Wm]ŀX~9-_Qq>^ɧ4sOf0|cm0, }=r3~ ˇ=&rAI!5. ʉn>vXeR>eP#WNC'bI]fqahM%-)] RK$(|qKhc&u} 4H 4,plH[[ >%B>MDMyI%5DzFuf,N2O_߲;U!U+[Ӝ'&7lr7Tkӹ/~fȶonfp0M}f?[Io2sOorU6u\դSsb^Q&$PzFK.煩,~w`/ ;1<$t_Wg=;rO].R\=5)9.!jVAJ烶zDիioalwE(Jܘtj۰5:5wjZ 1q]\)daFs^*rJ4C:`:͔r^hF$ kBytqU˖zl"Bt]h2cqpٙ]rX"]#0( J5+ @i pyyX?{.5PsSRQcfzV$]jB5 i1$sMꁢ+4W'M8 kuLZk8U FH%K4pekyN%ݱ箆2UsNԻ͊ Xwn u)1{|㿮ve낇G.wwSez!/Fh5y;^:ٜzc8_M |E Q~X}/f~^cy. ?{ɒWթijo;n8 OE1ۺ /O6Cle%w-G;%((vꠖ~n;T㋂)D_ q(8mZGs;W=R9s{J&淏7IYvސ~RݻD[š&u8=.TL_ul|Y+z_ܻF_> {ٖ7~1gcUpSVD)lUÆ MԕXtm%l`| j-w-dm6.X 9ED m!\Vu Gm4`@|;r\aw'"krkRBW3h5ar^k(RL҉J1 ^]w)qDݦNb})J!,eʀ z(F6P 0gYjjٖQ#q܃yG?X"YRP^Q1H2l9XHlK~uq$i2hD|fXwrEr+I4Uk,8jVrL(%9}gwkE^4:ƬI7߯1[vkq~vJ5Q^𪕋?m5Hb$?w$*j}'W<DDHQ퇻Q<-NN0`S Aba@55lH/Qy#b5w 6yfϳS'6$SiK THvngx(Sǩ}7Io)/>$sc," a2 O$юdLm]ՎsˀXBD%gؖV.'%1L +MR8Rbd [7Z˲kDAsؚq `LN-]ŗEXX"R?iEP `L#̏ާ!*5"4Ht v 䨈R0ƺF{ 1>486,' qw&g f$OSm=D#1ׄ~-iUJ+z+79mI\2F1op,-4n ǒN-[J_=D5V^(`S(JD1ŌehlBBմMh2nI8Xdkji U$E7E×ߎNj ,1 NwHRآ*Q( B{$ҔBHΤD#Hկ+gPLذ7 L40ķ0!`YcIemɇџ'paJY n)cJ=>/ˏZ|yD?DLcR91GOvd3^˦xQ X~adE*b;󮯰^:\4N8Qwdٹ&LZS) RK$7: ,d"r$\7Ki0G1I1}!V[(l ;K  Jq)!{ 5UiRS5K![9gH;9C-,yxP"* lh&o-#LuG{s}J3~K vwA\ J;xRA^'!vB 9u\<{1k s)B]4X:K-("Y/N(_tfC 7v;6S7p<$oxݰzDH>?N0l5j/.˿ؼ,O0+d{-_@2N9NƖ^FM+|+ ssץOܹS mwjE"ZO,SPb:uVBk\.V$W۰.R첵QT澷[cXkEqV L'eLHa6KY (mr,LY =3ZqfK}YdA%ry =y)Q,e- }4Wm6(Ge% y Z^{I(5aFߜHB8]֨M7'K>h5mmH]q?nW %3'i9C6;;LE m+oÖ@UZݴC:{ Aűh8j ̤h#mzPquY۸a"C7O!]ݥStPͬ/a0cB3~h %*@i!冨PAG=L+ft,)V|7&U^;0rV@\(&.0R 5rꈄO1΢i>.Y>U\ҰVBRs=b?|쌌҂hؐO8׈T>*T`G)j񱳊yOTH?#>i %I7:d:ZFV 3:f+cGXnxlVQIδS C=n@OyWBSun)٦+Xp8 |KYJϹi02HYtEK+*FڑajӅzke2ƥ{Df4fK]O4v7ڌ>*34݃{8+xk^Q  )4 iWhq3A )X") nBf~W"2.RqD pofRxEm,L>.:d֛xkX#S݅Ob5Xmk|Q5]Etl.H +wǝ<.Znkr-v~/R(?_NLrHבuJzVZFcL,0U'RX\~-5G5KX Y< zVeeYz,*TE< zLjIyy4/>b)= zLjYG|> #KϘ*q jWHD<)T]lp4ʁd4MhvzqI7BߛJgSuʌ ٟ~OEfZ,-X7Jq\"~\-noK%RΎw=3E!=`y& SYp1GDNَ6+T}%>1z| .I@ to (Ay#6ap##$ S:]As'SLM[I1R(" ./.9:oô5"ڹ`hG ؑ`;D'U5Q4Ҟ;eDW\H@: )Aiк9a3@F SI1~sѹi5(ԣiSS0X{&ɵ #ԃ& p̝>/Dh/ P+ÌԪ.NZG*5XI=K7aGr%2VF`>gu7JB,=GbtHz,IZ)5D,=cij\JDX_35#ϛX <2;&KI=K,=k2RFV75I4R!ϛi,xSC'R4Rk%YeK,\t%܃Yzv,e$-|n <]wt Aj]{qu7EP,ŭhݯO8׍C45z ¶Un _<`օzeo'WLЬTՉ [E ʹxF6ha;cqe0 a S'yf"S88] Kie(\TMgҲ8%G-Z(eᦂ(\n*҇]m3ai̾ΆA Pa&u%$hǻ|q4h㕏fT0l[ M>?p2r2r2r6saLa 7\L0>Ǒx᮰BZ *L~X"d7vZۆNV&H0 ˏx1ҕiY9Yw nsۺt~O c_&?8pTqN#Nm\=u|Z~W0Kp1%&U?H("Y(f~!%Z[vn8hhG U;I2E&"8ZJh=a[| 3;# s 0/¯BZNvjlbڤWA_JĭE>(y荞nv }Nr-TR#s$YйRgSMr->Xz /6HA6vRڮCM=X#l ɕG8;EK΢x;m>Κ6ZgKzxWּe{lyMl{Gr]?lBcN7zS̀QYЌͮOll Eܒ΃z5΋UO^6Π3ތÚY]E8͑M KܢkMR=MC&ZS!q $7ZmNcH` ]f&<77;Iks; &ܵHއ˳JK}% MRg%֓62-f5 "ߟ4 ے6W^HН'§$XWiAc<7Jmn位q^'=tDAB, Dafkt~,Ȉ^WkU )j*/˧g`Z Ve"Q5_ u\ r 4#qzhf,d O2s:W2{|b\gT\Iv|ekvLó`2JiѸm'F&qʛ*i?=nh{߆֡5K?)[ eyW6+L~VNc}WU=$hM%8.~ە]J#<+󾴹/]}{d"2 TE~{I>]==8 q&OV;CO+gr\dECíN NԞƉSz ,=qb6bЩAuw`Zi(f%PW殯c(C4#6k) T )!'Ty$yy]{͸TF۵֮r,$012 ɔP@A RO&I- LZIPtY]YZ!mYee oXօ5tjV*ԯZAl uںz ;uNKSRkiB B e8*uQ76L̍+{9HZ'Is:in>W0ًK*核t˖6벥uRd~_rE{ )W]Kӭ翮$ݗ, ,=}nՇE2"o!I' {^$L@x@2uEQrR!؅!X{.- W[lo3͌\H8Ё6~C2 դW"dl J>J[:d<9 =z`pr .J?MsG_tMϨ$2dg늁 0!_ NU6 +A0 Vw_OiHƞꬎv C8n!B8^PCBxaHb_x1Z c>00xF17`=A7VcG#@ &!0ń'q^l[nq#ZUW;DkhC(WA7nXv &n\l ]qk dvYr\bj9niu'qm<xw0S.cRƽְ)D!p1NygЌrTT%v%пLx :lu?1a[2)CY~/3coUޮP,淁ޞx\lԖpG KQR- 5:tNOQK4ޒ*C3iIL~'\ڇl-O˲4ZV֍NlDViHv)>mKWv-YahOjŁWfOxm<׹Vās+Ll-(}nz£ C83{&C:%M X .cZ ؏ Ck _/ _=2$Pξ{bJ]ͯc"#4?x"8r_?^]wg(*oO3EAxB!/_>r!$b%K-%x1]77]Qm Jmy:|a]KNΥרjcjvc'Ԫlwuxbf.JVVեF!ƚRxKZDKK=KKό,=kwRr$'Ҵk)+9KKc)c+QItԳLT*,=GLc)RU9|>&KsLZMJiFvsPJHAS(4ai:m= ](cCť6R KsQJOLa%ѷ$i1>7Jڗ (-: |\f" .Ձ&,_{ U.&ZD`&*}ņi'F^N2n4/B/4bT1U@qq{@X;9C2D:~>@9Y 33%ˬB ) -Vm<dʉpJsA\.x9b0l`?}JHTPIteܗ&-ޜp2eΏRQTz}^2FB&̲BWX!n慱an#2gٜ&ڑN%G;d^fW?xSp_O$k "n(,jBPXOa9cفB` pE\}/yR^9lrFj]̺ #mu4׸s:W*/nf04)M82:K#*zk-Nx8剪)?{Wƍ Knwt_lm%딳ɗ˖ `lR/vW߯1$1,:4h1rYpW׵`"qDXZ 'h[~k[HPIZ~iDv"XwcoBi0>}f ̵FJbh!0)ib 'ZyϽ5FaO1b^ի1-\]R/,:PKى+ÂX*%DXrmja sBSJeFMsC&+*xߵ#]!(\*c'f$ "N5_^#R=^և,Ieu0Ty?{π}nĨ5TSٮޟZ\3+X{E؟KQ}^bj~q^hkwU/O9O z~u4}:z!A^:\sd!ɺ;=}XD.b4r[Q;u-1Z+{y=, x4̥КQ:bsLvw0]^M}b=}I٥HR^E']ԧgW5;gwg2p^I5#J,9aȒ> Y=3:إ'K.]"( W5ȸ72 [|O>odqDm^9OJV;7UͫpC(U|O""@uy] ' s-.[EIQ2|Ut}C܏@6Qcu!qs,ħ 2xU(ik5YYkBSFJW-TSrjzf(Lm$ӌr?~Sw;B)}/ܤ0zNyn?s0OKf%7+{w}YT-t!T~:rk|4Pw?F  e b=eG'=<^j?o8w2V{@cYv7&N1ܧHhAPԏ臟WiXՑXiچAyykާ 4}W5d/lL})4KNm=lfYRK^~,*;&kk&TT% XadLeRjx:hQ!`U*"Hm_`[Q[cgErTr[Wz8l7$7VwhcZ0M:d9o:w|/19kkB3&殕oyW$ 1ҽsybZbAݾrjCrO[*Deק{25kODc2iUx7]<-=:%bߩHeVl x-҉ yq)&IuN"ֽUNGrHU1=Sѯ3''h(G;}N~;*Fc[#i$3N!XYr( LT!-0;wTU#PY&= "*(gFP1͔DRf lצV[iX<݇#G~>5!yr>ؙ~Y]zz.E30C9J?a-DW$ԘYĻ?ËDDHR!dlxO7ט!Z]f< cƃ{l?{d0=Ma{1 Y>IIZm])^W´q{`%jx?HiR0me`%Nɳ-g10b jJ\H.T53T`\7ݦ քF91/0:qpef2<ć)'6L0C5ֺ$5 S^د aZ d /ôa# Ά(aϟA>l~8(/ ^$ರ@ӊ{{RV[pME͚gZBE@HۇQzi>%qL ?| yދxK "0dtGEL2BaυJvZ_oQB-'$DØG~05v RRMeVط!;^?YwVŠpd@wU 26*1#keCḬCӦ._m,^*hZ,rϹB·B~* 2qʝe, =j:i :j1$)p؀=AW  8+E3@w,6a!~A͊S`JB%$g!]`&ÄxHR`!JwP(ʡPQHmȒ hA]Y06PcEX8e"SY&BB(OXNyPh0=E  a-ZZ#!C)!( SD*=Fq2bؔPIqi1\Rf*\Hl~x\$K Y{y.$X؄|X1!|5r&rRKraFHcʹ&jO Ҝg".|#,i=9Qq9{U ˭8 ρ/7jtWX[}_b8.2(b݉egnn_PDiI?y V( Ʋ1>]<|l"Qn g?|ɳsWI H'sr УwUOt8 1ɹN'껝O'B~W3/^-VPk3_kP()%'=<]jٷT+ehsZW9tvRk!+K-Hn&,5)V9?X֎]mVaptz{5}r/!مU௓2%9MtYdݬj~AH-Һ@wl܄Ӊ忂lޟ`!`M|L(TbjjDnx~|wYjR” T.Ґ7tӻ֭x[] bT;X3S;nc"[UNBԾ8wĔ Fu꾣u;;ns"[US*R.YY/nz1-쭬oByc2f^u5Y)ey%IʒNdN \Py!ZrF]wbRTw.ZuH!ӻ{3R(㤉.0Ev10.˓L(j񱉏F Kt5z|HH‚h%0fMa&%ɁwRK#^D*wQry}HL^CZye`* zWbmYGe.sy<+EޘURzֈsBUž.V( "Ըwv|$ G[q$o5\,l)X8BcJƠ-bLqV :*4q-G€Uq qd6=fC%*XZ1hlLf/Rc^()$V"vL"'R+:2aͰp~5teZQP!) A&Q}C>賩l^uɗȭJȌ,PD]~OJbH1PIF̅if)`KS-,i䵄7hlDcFP]#YwN ]0"o \֍AS匟وT:]*NŊ4.ܹ=_Fp3ɖ1uu0zUBZi-Nvʤ3ť6RX4ʐ fWnD]pQ~3 o? 㰾pikqV:kPOBc@d.BR7dZJ}.+Ahr)um@/Bj.,wa Y<~l{re$r:Cwa}^-[-ywm|洔 ~JCC6C6(,&)CVX_jȮkOt8u4ƹT ,cKOaKO7%)fԞd(CUͰW{:ǝkBNYg13W{z#! aP$ZjpʑûQ)3^4RhsI=VG$clH WK{??mr*@3߾lVW7a kkY p[u*\ͅpŻGs%n`*<7ČJ-}=fSsmsXRϦ "?MBf#,cD2˿vtBEiaKZJ3\NKFvAicՑ3f7(X^Rb^DZ^0Hzf:^] }UZwEis{ ݁\m/COW!Zw^0oy C@!O_YC 56U3vFu4OT4*uʊ> EiXshf0Ae'GSw3p,Z4.Yw. d>AC:Bo Y1 [uxݼ-g]!>cwὃ[gq5mUc@4u-leufmy;qSs;%9[I։+QdSTrχjy֙,쳀g}\w_aS"{ z+ϻ~:Żi] uexhW'וad3J)[mDZ9EjX*%H0^#$N`oHk?{ ^;87 n&(^D6W΍EjeQ84vrC*mOm&ǺoXQ1 wW[߭mL w4MA 0<MS*8bB((12jr4q(=u!H~Ƭ]*~fIJP[** `š >lx0AbW?Rs$`G=n\AX8eiVXxi>?\j{|-'RRp*;E1"}<"ce®w,]dh SQʲm{?KŴd:vyr"W۴qSh2G3)QYdICs )BN~~b,j[(>*~g&^Ӻա!߹'q>n[(>*혳D%hͺ7׺ա!߹gy:i`N]:}%C'LMx]SQS)6kG+]R-=@},2H=]Y䵤br"Wj(8[anY?[q7O褸N:kɨDX)Z0ZnJ8:%+]>) "Ec҈ J0yH)MxJ%z u 0^AjKDeJl~ }#7؀cz[sf*(9ч-kbim<^ށVQ2vASF(i !J$2yꗭ0gcцe-B{tE<Ϝ"`X}EΆlϙ gVlC8+_8g5z?c=ՌUcIbgR QJZ~Ĝ;$4 }G;-rSN)k-Ns U4HJ:RWn˓ȗꄾuk'(IijںyZ:4;W :<]~kS&JGu_"wTu}Y:wp UtMN!/ԓ/oyWD'l̗ȭ~g%>rQwy^ޏ`b&y],/&Z?$" lb#2R^/YsPrsٸ@"rp-tl.h:yuۣ>\O&kxFr]w:Lw1ѣ;r=YѵRM:͜7Q\̃b^^_:nFnT|Yh# mC~"T]ro8l23 :D&NLO%CwP#9N Q9'l#; XKz<5V^Y9r 2Vձ8$1S@$Z\D[Wϑ`׬N2|nK(0MeZ>[0'4@4뭅ZvqwqrCii9; BgOujޔΟ-V!{9OR*/Em< O&)j_`rD|-Q.l9HkZuy]n .3GX=2 Ѱ 9OWw@yZ X%˖;'VSbu)DL$=$%,p;[O>i)NSF]B\%Mb1FôJrqiRxJSThS@PO6t*Txsu %S9A:Vj:F:1ٌ#$)f`ժUҹfJz"aYץ+$HhXB~x8 ˀI+EJ_9W./XW bR #f] $"0 d4V cUC`nceMPR(̙8"HgnFrCYX*{xu>uTCJ7lTЬNP@ 7F,埀n-rL7L?j )FS#dbbSbwB6 >1X]~EjtL2V Cs6Yb*3gy"Vbc`Xq8 * R5p6\ !(QLb"҂(x%22Ii  "(KKHB(VXױyX^&w.ȌL[f"1s .O~\2__:Ng>l6ܰީ+my@vȇizA `'e _mNCw^OGÇmUȍY]_ºف_6El--,2qs.|9L!سEچa&L;׻\Iڷ͙B܍  ,5H~X|V9UPɾwC~SW}+vrwc/ۡ-?w7W㉻]v^5/1f7~3Nc2e_t7lϩk=ί߮Ջ|!Pwe)̾J%oǛT2s励\v{i5k<ۛ͠{edX<3.nYsp͓HlUE,&Q 2YtmMT.V(kDt-uo8MΣ/|Sv?8 gn{R̽ȬZV8>uB5]O\Lz\YjܖW-KoяW|݌AB!|b>[l9WΫz7u/C+Ɍk;w¬vk5R,qO_JgjoW2~4I7ۑ%+6aA "@qprc/sܵ;KY}S7|X6ӪW cdCӓK >dekmF_Ř?n/i,Va 2"{_+ٙݯ-٭dQ2]*փU__`+4~-zձ؛CO׳K,|stsb-T?s'zoWAoO_l/VNy߹"Ֆ5a.ٛW)5T멯=gJL I'׼ ϧB(U;!dQ2>|' RF3Ni:>x8,6W' ~sw [ 4R `x g}LT׭W=Aw7.Y0If펡49K/Z`SXՎ)oo4BMa.; m x"s =ڀQVT#L=>),RǯQyk_7y&J~\Fjb-ck۰Q}֦ {۰IŃ#y2i V3C¤F f) Rr**׹ ؤ PQ#w[x86ήē%:v~tūא`ZSorD 96 cws@Dv忠Їkʡ%MV)8uC| X _gmIY8%TDuqAK{S}֦-飖JR"\0f?TbVTsq\KmډiPyOQLau*XBQfnCc%l-%$hq.1TG[zZ"OU5n4m8\9Hyڳ6<{˟6QJw3*J{vD՝L! _iP/}ǧ7A_ὟǫPK[q+{="\{Wܺmcԗvb+J4]5Jkv^e_؎B}Ly`h/~ʴ,=s" 4V*KGHd;kf}- l3=ʨ4Qo=ʦlHٸi9 9avl v UDN&lJ2+T(TLgiVj*9 gȕN47B;$ &T$ġDDqcuWBySy6֕ /HHK1'[@$(Ro(F p ζQ$il_;&[?R:b8_`Sl?Qыb8Ӹ(F[R1F1J7jlut(FƜDއ'>kS !^KCX/!.=DC/". SY 5999=!魕La2pv @pV25`:n) " Ec@ZQ8-(yº)Tx7A+Q?IddٙZ ae gN*R!Qr.!Cbܓ~{O/?}gyZ1T#Z:ݣnV:.`bjXg==1g窥[ t}gԪw;lo7ϛ?s6ӓ{5f|>?Yt;yuJZ]ϯNk}u,FF ]j@o0.;F1]f20_]ǙAAHvN΍WkC# 6W(a;.Zʩm0:Ѹ (.{u^ JZGN :sZYO<>Yi=w<_Cn!{Qj;zY0RgkôgoCXֿjgrjb!-4z9?S&ackv`qj,jv`~݌jMZڊƥ.ԭenV޺ד1JSN P➭(KMWg]~6Jpե^g&83Vàdb0s<CP׷}D?8tDߑν#ϖ%Ɔu!,!T;%^–SFb_EXjJ똮}-[TS$4HqNm E 5<1@U ~﫳֪Ŏ)`:jiZ8tz}TZ ?_5I8ʃNۇ߼9;>γŵUqH4N44ƮUP;ihZ ruN-=yu @탺~e};UtNu#]ZRdcךj'xG-*cFjĭ=7?} ǩ=kSK1{fZOxn~w=$N_-OHDƾz"?5"i$^%ABQ;]'Lw;u';*|PL’m,/|Ht?:BfLm!EN:&iW+DҤWDŽ~giZcMoLG|:9Rr|b?Y1 &@Kv|بMrd:?]1SZ#)<$,qH9PKPSlc^Ms R5̞QFd%NACd$D{gDtvU`N23:WJe24^d2S~^ga{̐HKC/?*j"4Q QXf:'JPSZ؂ ,‘) E^ryZ]VݡM5(F9Hv^XO;)1 r2k0eCJ!ٴ0&Ѣ:tHڋgnn !- D`y; qZs҃R*&{!^h)QVTspCRu*0KAKM58b[K j {qqC_5fےZReVZ!"NԸ ~T|D 2ץ0J0+17?".lb;V.tZj`$04jGv6K;џ0I@C&j 6A,B4*hUu9w5 Nco[D0Tt<նdSgşwZ>VޚHK\3[rClg a䔚0x]\'֍_S1/A;%#b'[(&#%PqJ77x^yxȆQcCYąDJK(C?s bEoAKJ%}]vů~&>/ɢ~G} mu}~fݏKssh]Kf> }4ثe&!ZFAJ-qL d @]"-䋿j~rL~;^lz¤"ANj$ADkNPNYt0ge[A#U%d@, M:X9 &E[@a iHE4QM hS!B)JJ]jnr7[/Z[ly~< /it$g35n&zQlrߪf֏'5+Eȍ={eG;_sh|yrK–MpO74n7t'GςOw΋B߹2/_a)l s…ߩ.lwSJ("Y1{F0QnAIwMeP݊=dX1_i_7Ǔ*fQ;);0~v/6Ǯ;|,h,!w,T1viw:Kwږrym)w]ܻJ$r(.\=$/cD Y͝bܽjr>1mTf~stӷbUJD*'L94fh̐*$/;x|.h9jM6u9ZߕĽGN: ,A+Jzsf!BK2ûbt)E ׉CGNwnTcZR!JJRU܀Ua+Gtr@)1 SUioJ4ZsUkōtiZ r^!2"ZaL9r]I+))4sv j!WIݘ|^&g_]A%SK,{$$Og1t$`\f}r?nw(S+Z~;:/BqXu::k#z䐊g{7OmM6~rtˀ`F=H{]|9B4H bE9?W<(+?;-^౵oUF8}ĄFm7|IDi6@nzO Uk V Q2zK]?uXw/_~Nr%Сj;>J4dl6ŭ\Ĉ[-pWP 5 1qGŒuڻߣwڟVH+o,R&z3^i9d,U4QjzB*9D?Hd73y8r$Vz+:فr4܍Rzy4iӦ^6 @ZPK,XIBԌVTH#%(F܇Vf5r|2GX\ 22L|(nk$Jg@ntTskIK%NUȲJxxЪ"H%lkQd  ^Ļ(VHaA¿*HOͫV)sLRSJv(#fǿ?O]Z qKE" ╸h 5,8<\)˕gŴ!Gr3:!R3"W|L(pee%_ WVKM\rVJU`IQˎg<9jN6b13*ۑ1&ٛz9{ܛlFPxYI]I&ٷ,,rllMS]ǎɘw*ПViG m` Tơ ,@ |+ĠA3$,幢 8 )6.ҟ +(qluq,f$I,88LŤ/K 2Cszԓ)!Q/D&J\\^-10BU=2$ p-5'W<sH o∧8g@9&Єc<&xbHbHf27,d$dJnE& 2Eɒ;f )M%LcG e%#,8‚.e@Aq0BJV"ÐLcX L CnC!SP~1T m B@[c L5eD&& +J~~Sٌv |p9I3]-/iΫm&󄵘QƋPi({uC@U4KZ7[.bT')mWic-!Һ !\EstJ6 uк5'9hry:Hn3"!t:떏Ni݆АW-,duQ7adcE1w(db[&oeeIL֠[heiPЫ^N1 {KBӄMQw&O'u}ta8" (!G^$RXE 8-'&fjw y/v~rCnjsd[Z#[ۛ;޲PBP$~%]4N8T1eqT`W̕ʲ0.]UBΘT%N_e,k*+IEY1U6|ӒHZT; NcJx+j(/+'Ik),Mcձ["V[PZOW;u^^}(]||ۋ7 9&6[hI.l9vUa|#\5{5]d4˔$< %#f$wz!bX],,4`&ʐ$9s p-Iq9;4H%%g z(pK%w1/!Kˢ4tY](rNN66b`+ 3%g."Dvг9 W3hrƛgI$6N4m g.P7^uN8UzfrZ˄e 9#@^P% Xf1֊Ƃ)JfteB3d5HcPlJ-)e{ ﶽ),@ޡՈ8RBMc/m,-eig҄R5TRF4Zo4tEIkĿGc c'_ )?8NW+Wkz)J`6VJ! y׾^1Js"rp@=_1epܝz{Sq[yn죵SL=$-)pt4otӱH @zg O@0}xfu&Vptӏ;zlS}s7teȶM=hddBqhd63B(r߈AVpg N~qj xuDiD7Gӛ'e >&Q10޿L8 tQTl8t'TNOGdnaQJT;i,`4:žCCgx,VHj}Ϊ $Po9gLԁՄJ%z"޳ ?j!?x[jFb˃gLt({H!4䕫hN:xOкFk&03:+ȴ5-u y*Sq>nALAhg$MetCb !\E[:%#uJJqcZ*8W:#mGo݇=>ISJ9lHܩ^kÛ%q2q&&aH%s ,Q( ҇=bi%8P$7fF4p& 1Nt{X Lm~.@g`oAY j%e) )Xxm~6*c$Z:y z(sQb鬁A'za Rp8r`y ^:Ja~WfBG*{88Us8v<Rln8hG&8Ѭaqr*c=V9 BpRUKi JmuEaC Z vW'q;[2Ÿ'O,"y~ |DӠ2js)U"J[R9nUU$*ҲsbQi-w}mH2w[KhsyjRSI7N쓔If@6EK6HP(%lt7F3JPiYKQ>E硧@P=lYڲn,J-\"7o9=@NjJUx&9Lb.#7?cRy! ue^I^TBiCB kW|BL(_5Lx/+B^40_~a Sps JǢ|1DD2C_;|mtA_hX-"UOA1PNfhD DE8rM{7w[?خ2=Zs'dV}1{HϷ|u!N٨WVrJ#ZC߸rVQeDgշǃr-nnV\H |Abp'θUT[Yk^|FYtcy],9n61U= \s%G}>˨|qM/ِ42< D s"!H21:qɲк`BhK̻dZj]19S AZPaH>{E(Q (s[*"ozBN#-Rp;-Tqy0Jj݊̅ (;µ0RƷj165۱5E\|~,ƙh<3 2.d՞]]upݫG1CR7TA h[ީnGWIٱVKE2F(!h~9:3!$Opv;w??g'e4L]?Ak3֨c1wd I@Z W:bى96ZN蛍33SoY?aqNݘEЬ+U0b.[G%L0/#hb$N>C/fo(Ӂ` ;ٓ`Ip g"N)ILY yrf%1=3N qxO#1u'[^Wս+h j莐uGꦫ&H:Kk@V\}Bhͧh9ݔw Sva@xQ5:*~zt$\u= 4#iI~ H \-` -Jӝ}XW-@N`: ߒB#5$ej$Z0a b)a(7ZRMejF\i!nQXTĨN;X)U-k9z^^!_]l Z6/쏻9h~KW(_0@WIO~wXg K[IPQ|. *}IlwWZĴc d%_8$JY1YhQ*$QQQ*e)(T=r`{5g}[8-*0iGOQ}٦Ku>V [?xue_oFx6RCht Jwr^U'[w;#=2RQ#&YYOϟ"~x8( AS`JSx,(J$[]%D1)D D@ɣC=tP1P7,~kJ !>ӥDV4y6쀚39)Q$DLWb<> W8y/D{loϭ/9UF$>]E*L ~L;DU9*2G2纔qj]JMG4FsQ?pX,^.j<͉ D s@|OMNt^oƈP9׶!>Ln+VUrG7uk"w kGTK&YKOZKEdQ Ö>E#UJ@JcҧlSMbTB]*L=RKAiiM5>8-r\,M=gZ0}ʳ^I& ^qv9Ä6S;&^3 X9"FZqnEnsUUTڔJh( xy膊U1׸nWW/<(=(Uua ށ:/P)ʢ,Pqf F$+ !lFsn8geQ8cr)>IU*BJZ`?9.P1T&Q ҤCFM`pWo#qG+jM^{N,@4l;C>'g87P ߏ.eNNL R:Oc6Fyp>Lj ^G'@7q/;-ᮓ3LSX ksT34HVϼS`b1jϸ#F] )m(Rs2@}DH- 6(nq k*(4x4?3eA;@$+GGyTQ|Օ{cg WXdsn/ț;$[]H׎`*b>e&g0}L\l{(:J1AY.ծC|#d-JCce^ܑ'6X#q;k$}X#~ _~tdf<Ę$1fRM 1$ k#yd믵eѥ&` $8F tfd挲? %3a"3JMuMw`^5AvokӞG޻UCK4;W$2t;#IuKAꤾutdjƳnRm U4Ezl[7yf FuRźMExJ=uKnChw :E˩:K\j6Tx?z# ^F =?3Np?)9yQ5zdOY՗mk:p FUĺV|OywԸgz;`f?,߷ FPCjq$Z&p_+LŐy Ehzo!!`JJBIhy4bIT(LFdũMi1MTnC? ;( W]rH/BBғu%SC+M6^7>D2Uт>Dqx<7Щgշ4 ;(T*BtQ`θ6f:O^Ϝsl" tFUBJiYn.0goHy]&b g7͌@LeK6.9sH@bE|+j\cE'+H-9BL~$z࿵>085\-i~^\Wooj)\\p]œ>nۂaswEW6ZJ1Bo'1+\H>bha RXō֝")]4)rJXCsA tZ1rdDAZ3 :MUjaGDq;a='X}!#@$\)1bn&F讟TwƙMugñ5k2$Q)utx ;'NL 'g4k>#ͮj_Ñd:$.IC:TDNEQ2#pZO+eU=r=;OG\Փ`IkS @J򌜭2 NfTL\vj&PGTs;H-5s3 U0Ұ2ϵVVU%49Jn+$g2s4`uSѢٻ8$W,f Fb˻]b(I:v0}#buv҃DvvU%I\X4g]TS':jߢ2Jiq3!82J dGm<1OŌOÆ`˶{ETKұT3e݄`찅+ )w;6$Z $،pɷ P-"+E@󭛧eYϩwY `7. P4+)k91&F#5tNGsDn <馝ngQd~<{W10/)G]fJM~UCt Og]Gq"x$ZVd"um$ɜ밳k3aVT3/9˅r.\:,rĕ%4P,1yn:[rK@$i޼18Xڠ(dSX)Wmi\yEa ,bJlö"8Muvgq% ͈rCJtVS:M3",l_P{!&w-GBiJ!$Fi5ChTgi!H8JW,*tO'a(٩3w5x"s|Jcl!1nmiр`Ș4!ԝ`ʼnbS8M:OT$Z(빏ͅ*hr0a+ XcKy>6v $汚Nn֩T-v.͘@iT09PRj 3%TSp,L*XXH[&nҢM9 ߢ8ͷ 9>֑KLҖNIt[['J9bcRxKH1~{xfzyx ڑHRɐb ,@gVZːak8]K C-FH4QV6$"{P'5H,w,ZY :;jE`c|rǕw`B!NQאp-8\f3b&zZD: +*_Xڈh#+w\?wgLJ<5&[晆?38Xi ̉l;k&P9My]r!Rv:1wVa TtyJT. #ѯ*UϫhWU"+(9k@ xǜ3rjX{l43~;DE軾f8Ѡ Q4aeЯwRbVnBW)&D9zSO:$iԠ^ 5 y"<7i"`K,|x1l9bm3m*Ԕj]Y#(X)DԖ2N0b:eiVԬ!a<^i{!&י.Gui3MZg$:F#J{3Yԙ°8n 3NK#!Vw((]' Oo:Mg0TWdmـ`*cՙn%X$+$$,TQNʄmFMӐq)+RO$Fq0H[ M y``bxU@PBO^iQj`l@}_<޽_*UV.ZiR!nÕ△F⦲!p@5 w{`QH:M,7Nȕd3қō?,d0\hgcLf/(YdsKgp$1L#o8_VgRʱ<1 1 f60g ?9JsP~sGۈvpAՆc^({Ż/#c:zaܴꅁ5g31V')jyFf@0)&"'p1㰂nkY}SzB~v ?ܴWG Q/k(&<)/` =1Ur&R0S(;ovJa/^L~1 #?ɸvqi>|Up4 9L4r:R3@ 4/kA~tXB<=zўOןCV$y,Bz}yVbfLPE7 Q&H4`DqL`uNۙkA̪ 85tK(rG<h5Ժ Vp[7^Fi|ٱ4x[˫wuint 8Ðw4gv4 g^\3*UB*HrNv&ܞjН GIܚu6河F_ &8-a;0ըNlhzJ ~ =z~zhTku%IӅC0bvhM5}ӤN[RK qX*jt T g%no`(AqY?-lnݲ:",>~,/3txiw~^ -tnmEr|i `"L#_EEm,7N6%PPѭ RL;D%Qlֽ,1a!߸vJT3ƐBt*&퐫*eCGѭ MM\Հ\$-5'W&(A=C-ZJyF%`J'2gz"wW̳k󚦒%82'jQ?9/*%iK5L+6(ETn8RURYQ%=!BG:>c@mǽLuzsꃝ~~8|, :bE:-3ms9M%Q*l;V[X)Ly޾!K+N%)jkIC%)gC1"$:EQO aQSk6ffGmʞuc}q~z``\H' R:E&C/qiF@ટR̪vc#j7aYP 1eyf" Ӳ]e{6dg4EJ$r/dyq@He1cf AA(gV |TtM4QtuBW"F!4:̩Z'\l3.pRm0LI<廓%ʹ"^Ţ4]h2Y 6o52xyNALƌ2BF:'- 87 -wF"9cI/2HZf9I:Kn6"Qfb"u:HxxwC̍50a i><Ctͪþ˳I R"""߽[,./. ŗJjuZe%&=[؇# ?T? =<ͣCUv2sᄐ dMyO.DK555A˱ɔ1,Pe :'|P5uA% )g5|Fْ2οt?_җ)A1KxjZ/mx n*w>;"'C_=y%l8g-ѳ?iAk2w~o{i}~eJ] ңfK)㭯}jtw#Mo. y)K!:q_s\>aͯ=Yѧ>iYOk)0$bc"WY|lp#mH pc?I@5Ժ|pv1k]:{_ :͆{AQTr }5񕶉]2۳ bZ{ke!q[QlrA\(k)p 5Vq MfҢwR* `]ҿ_|j^ʹ+wr:u*r5CX+4 yzIDAa"Kǃ .N+,k@!ƺZ~`7-Of\"RiA6- Ě-_6[lTD(&Uu/"`L ccP;i>##.KM9CyJs3ng(Ļ0CIi݉{qG w˧6P7#^p>>gpLιZs][o+ x)ބ8lټ#[ޑsN[izzz,3,MY*ȪӕW!X#Ǘ_ʊ48b Ɗ_|j6vk-JZl{*F{Q=>t;?ǠяcJ݉P{[rX"4%IMrUjf7W3ƹvmqDpWrW;T>Ts(*Snk wyQƈtU[F/oM@s5=&%de1Ph4b#n؊2Bۤ^6e*czge\e]ne޾j 4,6j(P}+t=1!4 LAya?+*txhӻK$SE7z ang{ _]^l&fo`K]L5X3fЄ gl Bk!S€9I> {#ٟ!'$-O=-HN۩I+H =M0Tl>̔*yzv)1fJDcb,!g:;WS]]_}vF/eukʫdsؔ ̙ RP>ɐ;VTj-czy,0L~zcWdkGTVɽnnSqRՆruG6313ťf0IO}C/#|2蕭:,(8jE۠.jKPZ>j֮W7ib^Xr+贠fǻ5>;}xݿsmṼxso7.4(cW}_}-tUQ&*6UG܌9N\b8>F{vuCKK4;WҜN}`k'n#z3zY5|+\m6/Uѷ]jkab~g6pYDB'Wޔ\n#&⤐ԞĚZVZՂjOYý SZ*) |k9qE 1 - &<dPZEw_SWWa#;E"Ims+*9;v}z^WKw/adQP\O{o52قȝF'΢`Fo辸#j\RD8$jRb J >>dNw~In hjxU-A*'Yoɡwѻ}-?7e^?wjqofsW 0"8:| .-Ưk)\+ü1Ji "* >=Ųt+p41feKT}bS[)K#[a V(l*.j+T]j=`Ymb 2rh hT ҕۘ ق^-ŭ qN DduFq (3+gMK 8Hty#8>~,% w^((_ VB!KwvB%must'С'ѝ↳?{7YF$KЀSՄ]e ~ Ԫ;=[n^t8ӎm^X4"1|;oߋh4EGx险_CTuFK29 {}+'~pЯ4NBBBBn }T99Xt%dPIAC Y׮&!l28kuZ0TtT:`[LY:UΡHšO9)eH %%-Q,z3֌9^# 6<>o~wYQNhLd\ȑr V.YMjVJ-uڲHyCo[I\SU'=`X`(cD0EZ |E.ND.eho"\- C۳e;7Oj珛Y: mڽC(<^9 [wIw;,s +IhUbޡ#ZzC4Q҈6?Von п.~lsqAzeD&oܢtxF[anz73{~C!hc3?5s# , E5ǵ|8G@Q95@N>XrN7+ +h`A#'(X*$n tXԂ1d"v \֭WVY0J_3N=LDUAS8~%_]PSF:DŽ(5SR_1RǵĄ*\u[`S$('K$OVwG8kVZ2xV`1\IU)vRljDf_bMUhd:^Px .%mrp*|& \OMoѰ?&BğHȆ=/EcXB`]i [18# 9Q{u t%Aŕ5oݾw2UODŋD!QѼKr)|VҎ+r''enɇ_~$& P~oy|ŮkX\\ B.|\alqz]ELf_c$*ZM %MV1q@BP76a,QjQ}mkGXDY񍬌^dϐ$ {#+ȐH@Y1&w#+1DIrs(+eť{(t,N1`$_I|J2 e,|f+y$jgaՑlN:ӤN n-Kϧ`+SW5Iq k̃lȣӋ:'UMdyӗ&{W0P2uVM"o>|OڗsS>3ތ)>"'Ng"1s'4%9yS d$l\1JID1Vf̌^1SwJҘC bY>gLYlQZW&.d=Qd&y9'hHt,drI8ni'"-=c[nmO3~~ϒm|S>v- _U J0l"c'y,\"q(KgٰbP 3aLI MNѳəED L)JM:;xWP 3 ~:Qc!ZM[pE:mjnE79kG[ЦY6?)O1o.o.֟4h,xf#o9t/o7;KhQ:ng/wF JL(SH@kƯwفyݟ(`=dw%M>HhNl8q>y݌qJX9瓚ݽ^8 m$99 M.x~_9?C^~x/VpI[ |ttlw/?4ٽ7k|MMW>֕" o]B&(d;bh'kTR^IJm}BZ@Κ )C:De x h%kG!]%yx2 ԩwh+iwfVcK@ዯ"QXD x]{ES쨁 (riU@ݼyz$Q-c(`)i9Qꨁ'ﻻW@ B@Q'5-{pJFwo53CsIZ)9n|?PSÞpU@I%}ݩO1ޯۻ؏#~j^ !2 V}֜HFGH aIh-\a!K7V5DR^&׍5~a~L_Kd2^, L@*o4Fr TD$HGOpF0t-}V)%c1fȟ6HIU1S"Rb!)9ƹՒ{j 2KI4ZJ7C/Zhe?`gb_]4z5})Fd$埗#KR d1#&FY)aت⮒WrLHyPs*Q gT7N䆍݄8tRU(54$Mgq]ܰ !s \^΀4EI};OT.)+=8c͛xH~be`\gl5Y6y"Ӵ,stx#s\!ϞxÃ^URւ\[&ztqKl*ϫFhT q*!e+5O WY47/"S|RFdX7t2sRZիR:y4*%6IRZ(^ݾ2҆~ =7ESvד%I*N"$X.uO0l) 7U<Á!Z- q/\/Vfx`1)"PeG]8 Cs؃z^T05D<2A<&YDʤ8J1tw%NursvjV{'ZB6HPPT\pqz")H:(j,9ǑMHex|~9̱(FAR!#kI8%pX7f =&L8,H.[]K:SDXAdOK>g7۾ozFs=d2)| l:$%AhAUB $Ȋ^6"GR}en@0-9/~Y旅Ҍ>9#0+t346\T\6$\ ~Z-Z_T9%Ty7|ze!@~OaqVR|rZLxn讗3e9{#bۉn?~ }Zټ8ã 8 Uӱ|>W ̗]-w|[@ 6p^l.EC9p' T3,&}%Sv$D&^3 6MlZB\tKWe >+a9reɎĂZK$FHGK!wZ) 8r AGxsH W#+t 蔕qL4eF!،.z̋?LuYCj`K[4>YM|T)MT)d9*!{hdsJ|6:CUV%cwՔ0^wWD ]=SyNnZVfoO/&iyQur=|sj[?٫$c<1%T_?}r5nF\5jQs./8):pI$]`u%[|D4s"bYX˔#=%<']QX=4!۔]=M4q7s@ ;F@dp@|`Dm}+QA[#\TF3zNOwe%FVWb0(vU{]d`>-׋Ki)ţ:ۑ众d%<9Is4kbG Z.2cИ,=, ѯSsZy@{nG$(P GB&cMF6G82Bgbs%~+MZKr便iۛ@ O"+[>`{E@rfϑ iq79.#kC!UCMm{t*!Qb*$2V}k&Wr k$ˏwv RG "D" NV<&j3iGJ^ew%n QhewɐMh";)OB6Gw­AJj@qCC<.0(Y܎CNHfqd;S׃F3K=T,yÇ%bSc`Q1k=23v&941jtXٚaeKcrE>8Ca=\J?MɌlxtDZO[*{q|]C8Wvqc\%օa'(+WKt8Ό1TW\su/!XlV|E(RdU\ W$C4!8Ɋ\ 9l}"Z||e~z9>>V"[]qM7O_.b+߿#h)- w+XiV_sr.~|)ׄr<7.ϟa/v=HUfuNI^rIoK_\7tl5(Sm]Lby?[Q5l#\OXW" 4OS(~]:d5WJe退QӕO*<-y%&dG9#;q$%qT6-0ǜfZ`|P+.U=㺚V Bk>-?~`\_aYTdXïF˞fwk >TXy,yb<$j^`Hhiv+>h;)po0s=@y4DNZk{a]!q,`[H]}33RR N X K1d*CbMy&Ȭ`c]CU۾)PD|N=6B4ҳ!ϛ3*!Z{XE#x!KWDe/# [UU2̠gLrO7#kKB Œ3wmH %ފd5$؇dj-۲$/[|9vd:R#A"Gb]HV\u}iB|nlESE%VyVב#@_ wy_BBMEe4("E23mа dT_ucI!aQ h}љG]&!S Z9$frGq.ᖡEo7}i/SJ`cF'BY-%zC֟Vؾ& z"Lk Y?+Yt5M>x,|tuI{ Cm_m;"B"+O^s'"*¸7 w(/h|f髹 Ia†H_dvAEg˘:8U8ؗBY^|Xi|=eBϡ˂V\Mj1̴,I9)Td|A+֢4њ+.4lF[΍%݃(b燏?9Y'.ݿ,g\ź7. ۧ=dy.oKx2_}WO(mb̓hRԊb'^B̢ JTjVuӇ#Ny~=!Nr1XER=xH[BTV`ix'xGNQh|ɑ~^=9/bu} /2'u$Z+/u & Rrrl.HMJrQJ .dֺumje}ވL!k< &*I'QS`Z\C8&F`ݰk dp\1 d zTG:6Kb hmKCZs<=P{GLd C!XAgЪ3ɅQ#h@`,BH&7em+r~xWi-:6=r jރJ&DةMCv2D:w<0E1T_uC'{X4F@G` ύ3Vt V E˥/kS;ECCz@Ԩ=ks Ń/ ~BZ EsAb7'-ft%2w6RгE$Dm,DR*^-l^Btrg :&۠Y?V֞oT@aB_0q{!/R-洑dD<2LU5B`f̹/BOO6^DhHG/M0)3%pKKp[Up6Q4D,\ύ Nj\mS0mu= $~݈~yPJ<Ԋ(4/Ac\DZH\l SYz:00Ń4Z9)mmk>_ݼṢ\+p$F$]y7t} |M%ЇpnsgU\MU& !F/gP+1E9wpo=ooQ'isb)d"1ߑ^Q<8vTv}Gi[蜎}v$Q5ƕ 4+-j[B=-4&fi]d+D"Pel6Rpwc 8~.6iRe﶑XWL!҉Z@G\_>/l_}LnW7's4ܶlS^7+,EH}Rݷحӹ{ Rǚ|RNoNޑ"skoOwV9=1y%[/ V16f {7jldLZւ9<Q4j%ƁӤk/N,Z7[ qD%Y'c 9⎺bnwߥmsD-n-G{jc;Wdb\y7#;3CC-wхl--7εbh5HZLWF|WA/ä́wճ{wVV ~&7mmAꥐi>._Eֺ4K)b*2F5Ebd BrICH!Idx(&+ݮԓH=0嚽 ڛs\y"!=Z$/K&Zдy0=׵I]JrϴIm=еM&*ߓe_'Y uw;]1‘mΤJ@ii=vhwz ?կQx`#Rz\?coL]I`Zހy\zwm: #9wv/v`-lޣW1n>.>ujGm@BJ;q䘻 ˜us1#3#usy>mۖnmۖN?׶㛚@oԲ+Yr{gљ~UsX@Wk,/RjJ9s>:\%bo]on_c #@(PJrEeUR?wEeGX_qQ߉B >Z1h ELQH1q}əBTGS>KUNޛg_fm}WoZaˎ$  L1z̳]ա9p&W8 8HzLKY1/SxFk[n $Vd*c-yzghBH Iz()h /t'%3b`gy@VuaYK3ҶY^xV8\7z&%y$?RoOg%zMaqTA.z,{q*Au o~9l6r-l4xD3@ۯRe-:sfPhylGJ,ap制.xXAS{q݁yDZ1bg !rj]Syi) cQzVʃ2ç YJ">,;?{Wƍ_vGh讶I%yJJbYTH*N|u5=( Chj@7~dx}}q^rK$¦M:?]yzE&bi4t)V6V yV %Z)Vhl"¸X`u2-|ֳNl\餠N'-.V{"&`x-oa AK Xdz»:ۜ b~_Ѧ3ц2Ȗ| J@sud|a% 4Sj L;hX(yٸVUȜ iU!aGm#PjpSGl%xyiNރ $7Ѹ/Թc$&WbZGL^t!qR[Z(#K+o~>wjN6:"6esjœQ.EtԸZ[[Wf 8kϚѨ@';DiXr!rĸC1ȧ}>O\ fWwL -ژ@"?qs5/5^D}V6z^x3u]Z)qz{*lYx1)չ>Vˢg *%Ph8⏻2B-vu1"_o/o/ϵ" ]y,=xpH2BudRm>?.mɣ  bYX`\JLah06ҕ 1 M-d=K6%A +TKX08肁g4 i,3J/A ؠn(ѿUd{@uvdR3|v{|8gB0Jh^W\U{࣓KhYt(.>v^0BQ7Lo8 )RZ'#3't.#cSwם Z )U&H :jY۞ U ՗lgZ\XsqN35FU *8eRx~oE`Y=m[ub^d`V}-ܲm*-zҿ^qH LcV ^j6Sˆ)fR6k}hI#nU*8dFZMauځa!S T݇ /lj0B/RefJјmjRc%.M>gzg&>&.ZA5o< ?)< 7&Qz)0Y{%{?5|%F+ B*]H cAeȢ88'+W4{09^;cf$vӈ#0 {V΃)&8028KQ%K mjPQ7-u[1m R; a@=4 QGG'{Ph`ۃN(=MzJF` =mU41s1VDeC@fTta,s(%D< FʀRj')zQY{nFgi_'Zv2|0,1D㵮%7Ҍ:)n0x(tUR(wF2<cfH eFHǘ" ʲt,m)+.+F`K/K ?̙0 ,93Z88fZ)6yy>zۉaz{wvdʎ3JAB;#ܔ2JK!d0X˼(a65[V Fu]c s3~L1gR~b]/cikzLqYn9C2I\pflb<Z>x x ,eg |<3 }hZ)^G-uLV,)Sd}j"V d\{07unep׳?{@ i+Cʼ׏k&_2O/əgXI/m + k ~ӻ<Z8wۻٻ3ZoODM+N޼U)>nNk]T\6 iV^%9fLt=r*X ߍS-Jr\`PTT6GrNEdYB*. Z-8 #zq?d:vӪTyZkPZ|E3};"u)u5J=QtWvmϿn>@[^M} UloMdE=jbzdI n6$,L6zd3#\y_Yywգ.n????}mmo\xJ_үFVף\εJW$1 ME4"=moU&^MH9;6Bxy!<&oU *˦kʽ5 \s}KE[4 YRo/@k :B{mel-VUmpdްѹis}ڂ^hw >ㇳoic7;oL`KVVe*p4;fwӓ8N_GBy|W]g s$/1d+ru"xtρVaP弈֓rdS0Hi[WNwn˾Nؙw4Wn]lJnvq>np>[WNwnEh+μ[۸ޭ y&bS:Ӧă}ހ9.Yxr{Y_oꤜBA>E*JY8t,<+1B4 DThm*'ZI+i5+d4ĎVzxV o˱\~u'ƅN z|R0L~Ŏ=[V6K=oԴ2B)ww#TP,Q=]"*{X'0G}h7ہ (8:šh7%nfȼ:dn秅29m|8I,IQM1 -g(^8axQϟ7Z*Itc`GQ9g͂ܗC#+{![rU%%r-Aso\yZm }Auö -|FR3#&Tf1`n5$x:&s0ꄔR:k`Zq/tz a6TQrlcrO5d"l1,iTn46pA^&c# UqUB,eD_ [d]oh)Ŭ&{946fK[uk|&Fq,K[败ӛ8wI[z86miM(iKI LJBu-IG* ^=90A`˚&?zBA0A󁇩Sȡڠ8֐)a:QP♊ z=0J$ArɄGwp<# {1T`s' =pҩ973:]f:fKf 3s$0 Y X4䅫p vK:UԂZ9< 0]q~zE5(y*hywRi\2~Z*ev*?iSR8-jsZ;TVR4qZm] MGhtEuOZTEB}cP&-V/vDPzLDu'pU.T2^QYa̟=tYJ,+\tL/atT'G:{K-^)]fk{ЌqZ9lХ>8ӓN۫oEX`jScnu oԠn3l$N iO2$`sL}0d cCx$xC^2`Q~;[ Е3?֓y_a+tY]}+Xn08 yCnkƼ J:0 KUqsnQ*i@’j߮k|%rx&|lANKry*F.HTY\{M+t5Q2M+smK~tmI_I&ŴNw<}h0gV~wT`S߿ZO@jfWОaaaDѵcOYsy}s?V%~Z@2|هF1(|ή'J&$ϔ5ϻûډ_c@Z6bԫ/խ{CKٻTN݌T?h#SAoDGzldt3L҄ss "h%4V@YNFuÙ]~zSg7EIQ.WEϕLud&1!9x7 )LS8E;ei;bFNXNd$̥0W;HۄyTB0H$ƒ<Y'1gp|&̫7<2tO`OS"ODGaU{{ INƄ`b ]U-L8gXҎ4Si ;%"͡P t+2Uhxtk# 8 d< RX$oD,M5+CLկk V!,h@JSIdQ.A5柀0JOA)rʾ7{E2ZRTWM!4+6P1޺{;'n1ZV0.|je}_%{Pq\}hÞnF۟] qӹNH60SrɻI&.IJ$ O*] R8T$|BD ? y@\ #,!Dsxù2?+6gw=|Vj'Q\]Yػ {U˷^mmozyT0>=F6f dl{_)^Lq%77gKlUI@o6ת_6Pi>٥|2i*']0 Fs2+UR0Ib SZsI ^NLcpz!b=o'DԍlfB)pJ&w%p1v43"8s*V`U+_7Bu#gӌ,iBU*n6cR-{-Z P=}6HҘyEA-L_UˆM*lg@1 bZ?[b;1JvB. Z v\;& KnQ2 @$y鰳atRAZIR kDaJ [׫{&euľW[$ڹ:>Bf|㻺}S )l΁G@dlPrf6|C8b`GIc&n95~V˛?V(+;tG}ILЁyeHL׃q6Sml6xq5VS0}]88kx  C{g_Z vc#ru d1Y%é& 3Y0c;Q͙dGѸ;1}CsHimMme zֶOdk14䅫hNi}~z׺u{p,i2y:c$YbVFu[X6p)DA X7ޅEYк FurHnmzZ*5u˯nch WўNEf"DaNx^8AGϕ!jEP\:HW>g{'sP3ՠܵ6+}Q S#A^Z$/ ; 1< 0zx =.JD{XS,"T̈́Yvj1%0TCX|C'_ 셮t,~/x Xr'_$-EY69(i |M:Fҋԅ\Q~^k%i9cTLB7 #_3Uíy-+B*m*>4ʇEŻҕgA=M"pǔc^Iieа ,>2 `SJW7kgAUMv:LƟZڷ8o6NiqVTʐ4ו`[+}6 F!J٨eOQKc(zFmt54jKIIܰ)˶3~0. {E [Eak7J"x%h$SL!Z G ӌZkdZHİ(إy,2c޽/\F*d iİD`FȰ `40âgP"'2r6а)f&ӔEoKw`{wݓ†=5~ e;øaI\79*@ȍeS jawa]^k`6DP! ^`iLBj d}k s֗!l./;|mA'uOVqJ>yX>Foj \sQ4giZ.d{ۤZ?6,X[]\ OqI+*T0Rcnړ[>90}c aj$:9RIFd$1412I6d\4muHys-" NɘBpWy-Bx&RYH2޳ֶX|}e(Ipd(nljű6Cɘ0" `T`I A %/uM蜨%;@ {ξA$i%m-L {bn/MwY.nh8Oo*NL$غ5 IԨX%. $洀:<-3zD|r&XGsg!3s9,%ZWo~Z$,9^b@U$$6ol / ,¦ #+cpsS2eǺ?pws[\9\X|pӗCH[fk-l=TǻZ`r"?c ;c^2CF&ȭ]`DqmdZ8I6%ROw*jM`  [[(Q~øΦ;\Ƃn6~%c<]!I&L7z'IDTVN&,Qv^#ߥd&PM}}L YȫkJR.&Pk8NSJ&2⮮3 ֓+xU\.5'weWct@H.F鈚%!PoKM G>(|ie}>+xDTۇ*ɎxBM ހƍ["#_^n0+$(NwX7,j@Jl=bjLf%ANFdJ`r3OSIBF_ǃ )ɓK]B+XF7x4ە'۾yOuҕM4Z$ޟ7zhwcw "2)NhFO4[7x?"X lr_!n FEu z͚Oy@ 79|+CjЂhAgUF5PoK`fW<-1X慭 .ύF Wu=+u%Akm}eX%ʀV[:ʾ w Ww]3l=:փx􈣵s~$U%zv=tkfY[6kV=Yys(mPka~ ^<}ta 8 ;xu팺zt$$ANjGcaSn4Ҁ6VѴD4J ^>\wO̪6%Ҡ!L;z0㯄>TGKl8=-UӻXnNZl[N)<3L g1H$m `J"gS [40֬9SVUc[k ?7B~BͰy, R}4.3ߴ..Ѿ$#&#kxZSh2.9oGbF&i`f!8.d%׊ARUXQ'#hoE*'4zx#")+B39"% Ƣ~&ݍW'E)ޔ\!VI?6V()<Itp)){!i=qԋZCOpؿDS KWiUXf>F{=v*開ꂺVeYG` uQWz=(8|U8q[]`=NJ_X%H-@=12z,Z6LJN>ҮYaKJJZqxD]JVtvGgѰ=SڧߴνS'EENQ\l2xI).Kaԋ "`}F4 ^X֜F֖/]Xhz_ّ  ?כU %P"f({7 Y/*e~\0w=7Ҳod `L.@" #)XW2 .xv5 Wl2a?0000 a:%!HQڹ ܆QEVnt 'J!#WI0{3~GGu:mq}ѻ/PfD~ZByſ͛wɭ~?&*R1i^fGhJpEM\eTZe*n5റ(w"8n^Wmg z_Jއ~VWG'U3}`Z꫇U1ɡaJSu3 6A*N`Y~ >$yѪwpu^"Au/%YU)IiLW=|R.a㔒t뎹.h a:x)dew3CsXn%bGqŚo}ې!{][vېu+KtЩW j_RP6pN݁\N [wŤM Mv״l(gz8Mۑzsd7J AߗH?zRoX|A}P2l59E`rT z!W:V,zQ:Vx BχeNA`&h?fC(w*cx[&EoD *jjGW+ 菣Orvmٚ>܅KW4s2nt9E]:L6Ũ_XU$AoX s{E޼9}KУ7@!AI%w%+OSP"Y*bP2ܝsRbZ.4dޛ0K|'QMr9xa$r?D5"Ar%lwY]CT 4*64n( 3Tsr e]ȎȴUȟ~5B! v5*|$SƁヒ; V]9OlW9]Ϳ8sС\4NXA<$?c@/mwyѧ45ur]P9#r. ?2K~|ToNv }yt`$kFtFWn6R?mYA,\P. \ix[ŘڰAPv-FWT蒸5Tr-yTO7Ra :#@ҝ8YrQ @iڃ9_Uht>RDJ* i`%W6*'ʳ(ZnRe )۬@2HR. n$OH0Z4 -U O4BH 5hR۰AK8MȻiQC`mD{o/so{݋i/MD)? Oibo{Cw9ZGg$5mlux= 1!0s窿[Gnw1Yy?rd%1c`2`O=*P|8ضbC{\RL$4#6MkBl3Uk"҅>8tVCt[֦L^.moMyݜ hEcd+1`Jq$Vs" Ml%偋t^✎%(f[;^"/FKEe6I/F% }!j:H\j T@""*8}!Sٺ.uX|V W >$)-"%VGMRt[{yJoD V+5SBQ[94!$J&M *AЀH1$O6]/vx&1]~)C83O&w^(% :&bWZQKy{ShhN2 w, |(J4EF852hBʀ!-C+MLֻr$4<2W1k*qv}1h!Qrp@op ᬄѣ!3!zK@bNRD>_f*`L)0&hE}sVwM 44/FGՙr] =L= &q RkR*"}+~LLQ@O  &4^]Kany'E:Ye~RMh]?hke&w" N&tTaĂhehG㛷o .9O7W 5@]OpaGM&@Ħ$eu2ܧWM#A>>p՞qB0kAijWmÇ;g4o/XApT:S'6?_._>GРEc޵6bsʪ )'A%4jkVLe%w[Ne$ow?f#BD-I_8%9G'Fڜ+#J1Gv׋vu1g֒z:{{m`)AԷ.-( d{D `S8c[)ٌxwK޴ IЂ G$j $7'M 7=P PƖ6ub:פfjo䡲D8ls Ȥb2)"mcGb à¼l&Fxb\}= FSR2lv?w f9OSH{s!R( blV[-ڜnyY/kC*O1y xn*.g|P|qͦ~ڮ;%i6D1wJ!VSXx" 0jt<߭rɯ}'kTHQ ,լ"'SK!DJL?>)?..'gFNXDțWE /&Gbtuy`Gy%HLG9gNavjԓ8ɏ@E*I`x$ڷC.ȈMڱoO ,b&4S p&Kǔ͗ ytBKVc'*|z(.S8Ze/XhɒxwQdfA$Ph"w/ϡR(jvsVìbX1:k"v!+8JUD'Mb=/ @j!ڑZS8h$,^P)Eag7 e*92&ʸ9]:|v !([999&ir]'~7yTŢb *K2wŻWuNo~ t_m5|S;kș@3gu<;b#"tJ_vޤ2ĵ'Mu6ވDtr4Os"uYdwg8tE^L>+9N1r7zwc9~ljQ_0Ւ)|-{vxS Tw %2,pŎ4dUøSuD#*ɘ̌#1>!7Y`|vP\J 4#'6n!o<4u'_pfCdR$MN3yP@N>vNLEXrh M:'m$NVy}dc(>"b*&P@#&̓`FR -L"Єi  KalΜ9h8K]mTF$v2+6׌xR83( @q|rZiPKLDrHvk%qT\q3N4}:D PXH S >6/Q ӼED"(Ϊa d rDHe\w]FCHE;`aAɢMto}ԡ2QJ>2\9^^)ؗW&!SfZ$f2E BJ_1`d\OsFs9X3e[}X1iN"FfD4:{KyBVŊJq,ʉDxD@Ñ w -ˆ|ǍmWKwnIt`TxICMrQg ׹L^0NG+$8WdIU;t.Ţ"]z- I'>z0wizT.k8#\mI!;+XY=iQc׫ Ώ(y3q|"w\ՄN*{VZm޲6Y׼vL~=j d>j=ÒR[-V~XN<,c-5f8vP|jmծh: Z%oClNE542lkf! ̒WA֊9H<Ycz[{7?JX N!,B/r4 E2!kOh+֢xdVڱ%zJ7+FWMkO**(x7 )RJT6g?ye7^t#.Ci5,˪*Ez AIaMAy&U9TK֊QnI oAq f6vFop/ըt@A$6(b䧎+Vj{nPy q]LEzNdpOG|iiZ_ͻpS73|Y]1ufu=EW:ȫ&NӼzK_Xļ7rs:m8ۓKP!r# XH߫[,>]mMm*vP[ŧ>JqV}f ݫE;ys1Y|1FjcFQ~,9Vsuc0^|)1җvgf_8u_>X6Jd p{>=zP~"@1b.~V?A`A$Ŝ2h=Q+D<یҨPd.s/}Ĭ) pmO+S֚F)YTHo5Wo(A"|h V1"64צ"v%blLU9Ҳb4]A` N:d YX"XHM?}ԯ~ؼ(b|t+pm/2J]bϛ)"}Yz>>q_~}Z}%xk<-j'@ٸSWx;~RLr뻶?LsĹo^YQ{v \T>)IC~rSusiʠ ^QS=u uAC~rSΞq~g?GX2:CuQIsj6hW6*n;)EEY5vq?MUTZubjuxӸ?ւ,߶~eȨ.W39,ܵaZ-ԼbSM+6y 64Xgsu5n^*FY G|<$)sȰ5/bV'i-Y-/z^H= ca.#LX*9nSY 2% uxm$lHA[^1DW2EDtؤⓨ.1&vA[7IWu{A Fs={>Z ˻`1(^ yG!?[׃%a@7Z-Q[KNRTb ?*=<]_ѿ&Ptlz(YUd>9zL.qS7m} uP,}cE'C#1"KT=JKPc7yՙ5:[0|j 7؋gDc5Bri\cXR 4?mcuRtu#4$bzN0`'G$:OZ,h :&c#=g"^Ku/q2r8mi<>d $["s,?&$l(EWIKBԖ(8%AdB/G} G'k :206ʔH(3 (9r"GecW8~2&H.=k$1 A4yM_yH0| dfr:;Md1__n $']!e6["v,͢o>7|B1c+ن`^fEʵdjqpI=| y2Q!Y:Ԥ7_CS%AIydB%H1 _ĵ#7>Jh'nxe8y\B 1c0?WQo1~aOMr@%yt= lhus1m{w 9qI6E%& t>1X0̨txul2xzI6b06eH7#GExR s7jڃ i.{~C`IH=[WYK0X~**2y}(`k9#A"{.Z:MoN)3d%o8%ٛ-mt>58 y&8 Dw\=7azwX%ʞLOfw>߹;"$*2sh|=څ) i׃hs?$B_򈼄n`3zzK7<%ZaHRQ"*UQd6bUȈ<^(" 2{z}pٍ9IdPU&c38h&FYtfӯ]Pm(1>^R/"Q"7X,f;L^b-Dd"`X+-QLo$2I'Jm V. ߇5FqaC0Ds(_ZOd+9,).mJ|2!L>0"aOTaO#4(6#ZkV\!ޙ1hrMF_cdX_4QK lkjR) m/Є ``,G&x<z999D+YJR?uf}Q~d߸hbwDSy}# ,D O) $N89p*P|1<ڷFpkj-z;u}w3xeQѓG4jD)h0#Ym01k*rUfh7Ԥ7Ꝼ4,G*ŕ˾{S*Lsj%r梈_ӴA)qtxȸS OMOvXmNeoeCl8I6>J.2˛6BqyU^+mL -.F*)w\e̦=uvU]HZO|biirQ6ΈGJV=v2MG9peBeѶ8K ]_ p%cOy,a-W_'LE*d.,68]4|`7`?mQEY2a`"1\ Zu2 HT(/ P (2uV/LkWQﭛk87v&+{p {X&%0 (Sg!mZC.r߱I/mpQrfFg藂M|bFVvaȘ&e px.k#Z p@(Ac*?plg͡^0cPeQWMD:loۭ nZ.HmFW!S7*#6ߎix[H_BeNM2?{o=nRv-ͨT֭J%6JM86 EVtM}KEY9?FDt%;.B;=@αV{|P}n:<[U+B(ez +lwh_+U!K1AdcЌ EZV)Th6?6C̯uy5rK?b+wG~p8JøUz~ >1{Z"#"qD~kpKw*OlXfnBY躏o5q*ȵLazjdIŃc_Td+ }oĀJ(.^g\ j_^FnH `KBPIN/{߻rRHKsRj7=!}}Tn ڣ7 ~@+?mJ^jtj.VB9ʞ`]b}L4Zo.|B7@~| #ВYbs`$]fOpbJ+v}v;x"O{96䘃 %1*%jg3B4Xoi_e1ZfdẍW"%@! [)ۅO.;E_Gg,ˬ A<{bɄ2% Jo~@[-$ɦ!DzpBl˄h4ZA‚V!"SIDfB2VdKH3珃S"\Ҕ%[C ySF3[R3Ć&a  jg ElMa0@BߟF`5P #ɢP\ 0D2:xNQɔ!ZH`أ(I zcm*.2{wq5Z)ߐv8h@O@]ggW pjBd5cH09W ZZ =^{4CkWݭ}>ZIoeo=c Y *Y þYa\&EGȂ9œp)ʜkQY`D!&YH6ʅJ %ުU-Pˬ}j<[@ViF__Efi">w:@ _{G~`;AT^W(}J+1|2P: :R=zb` leQj߽S[V; OGDl|qJ 4z~0[7J2 j{.n b~#`zߌ!L|+,+Zy[}JX*wٛ7>wLHI bZ!y"rv,PL;`9؎zl j5ܹ&\s!@Rsk}S?\kȩI(9V__y+jV4rukފfJ*Bz57km,,|K' jp i|u>*?C>G3:}!5+#ڛˌhM^%jP(mۥEՃAQ xxA4l1@yirVd2uJ 6Y*l~ԛOy\]dm|0^K4{g'0~x};a+^牦pW8t|[^3FR[EKU$GWFKcVg}/p\1/@˷D,C[Oe dwo! Rw*b5 7R#c4rAr>JEFktN eel!F#)=XwwZs^5˟IQ~졷m06l+ޮ)CWBIi^ƘqB`8KJG!=eT,JQS|1b,D3gxuYogsԴ)cC0b4lA;UlR㥯oBn'gKԗi_nF~^y$f#Id{xck.&Kb܋k[n5T9XGdQZv՗pMRXieP.}e34 -`Z*ap:C zcMڰmg":;p<.bf9Bcm_$BË9ٻ6$W|ؙm~8R؞yCQ]L" 8Yhl4n4@a0Ш2++S!x XaSM$1ooaT9)144Nd_T]栜OݵOAɣXxIt+"@NWOA ]4k)_c"#Q k)BsG Ok%Ȟƴ~:ČfXh䊉 )a֘*Zr%LV=H{Yュ"%+Uk9ۍw_=yx}[}T#y's%CKUr>7KϨ+'k [e|Rlp)b_.*g`-x?nРԚu ` XƊ~;E?[B+~k@5l܁Ut}C;Ղ):v;MyMU' U4K(:zͺ⥎b:hcLkdžn֭ U4K=zyݺKGGn$N;ZYv=QκeΥZ4WцN%Ix=UrCu^Nonx?;.QԨ%)#\.d>1i&2ѴE_8},qTA@ҳr\rQ4:r eOYKOOKYb ||n%[} 3siw}( SA5+M`_V,}I-!׾ϝWaOÙMEҬ#E$Y ZxY%ڪ4< ^+>

˨ֈ{ĞeM6Z9&?'2F_ Wp_r"i):NJɼƁ)iKJzH0%QT'x# JQ幱Tf">4GXk979 Ih3.UhkO(_K0mIȃZ9Ķ7uhυG,D6BZW!tV0֙`e7m-mf@ebrp44.kLNRDaSMM!0S58{.@ϽRa3^D%yh'e1'/HF豱nz*,֢cJ:, SA1f9[N

g9vI< xJc@cʺ{m *d#D%8DQDK8J &QS Zjcέ]T_Q-:ϭ[4qRӴtA5묥'Li)S [CKwQ}դ1|SNYKyb0/1+n_,.my[zxm\9≹gK$D{2Q9!Hd&AuMvv (܌9ѝ1 did2lYhJOjK_gS=6 4^WG:="Kkk2 L(]ZBzIxRޓwlZ,mV0Q{9YT'45~bc{{(w&O>Mr_䨒2/bQLD%ȞFs*Փ}5-_o1޳xS0@WĥG.i뇎'2ceu/Y)ZariYpҡ,3X^jLJ!,p+e@r3业Lz]^!R"7cKe R\ɘb)gi/H%~-MĖNDa4lk:{q,IОXڂ~}rtJc<!B/iom+O-gΌxuW'i̪>Ќ?7[񾻼A^: {!csEK/Ay1 a^c-"-bo}O0}~>Zh ߈mbpN!w)jP1{c~L?OgO1jׂ ed0q8Q> M/8^).]cˀir &=)F Nnii aS;Tq/Y]/z IQmDr}'-1zIPo w~<#I>؍ٌh߇|*SmrӺq:떋A侣uq:em#֭ UN1Cbdm=/6а^.@ľj}HG|Y@]?YIi++)gsV"iEeG&Քiu[88X79]T_5fDɳr\D[q]T_5fvZ*و 4R}Z(7_j*Hv6Fv=E'rBjm;?5 j+cߚuP@MOZZ% =硾ξ`T+%j AƑHW@/q`B(9M /.n^X3yQm O `j*×,%W2ܓ(z6 ֘j:Ϳu*ԃ?@Q(`B!/ Lt 7l*;(YKLyC 0rnguWn {M'H%bGD&8}I' 4{it"):0U9--SjgVF8)`!"-gKepkZ(i$H\ +Q#!- |U[:ƕW20+1X`#%%Fb*Jee)2p|;sC(o3jW+.˧'2흑-==J8H~\}b Uuޛ_|xADu#W7p={/EqQ97`k$i^~=_9zHcq11)(ޛC7 S׽[*y8'%iGД-҅Dfn}:3 t˦Gfk}rj,&{mOtAXmѻc QRͣቬi40Tr9#n@wHz;{4ӸLKz8z"vL}tɃkSV@|F0G~˩u>'{EW?!X{hewHH$XLRJT:* /UC$_&XlOZmj)@RvS}դrrNX\ ciN;1aQ[ZZ*}['=hG(RwS}դ ZzZtE!b(4-]P O9e-!ܢ#RD>k)j)iZ2SKwQ}դJt^O\KE S'21_Z =P0>߮!&I,*=? Ww6oV[TKc HBΦOS[*HP!w!J ZFTZ ,j>ŦU H1s`5(crңI"W".Me ﻏ'â\Z5H[,*18[TVq6a-eSy^uSnsS %9^כ Urv^'.͋McܤAXIW`SKiEZ5ιKdH$B:+ǥVPAHl l+h kHD的ݾ< Now,JksSy]4=_~ S's _GWۜ6j9xrYB*1̚2%'pb%$+]a|`%eeKa1co~!WsR^@4jN&V5-NӖw a)*:$  ֔)qti'sCpvlZ5Äy7,'ߞ5"p>=-$;2LOqC`6S9$4B)!_\ Aj\Jjͤ$躖+F^"Ḥ O`$1^\%ՌrT25GQi#>PKn M(cQF)ՎV0TjV_U$gg$!\Iib w&ZP&*$JN9̈́i0VBk.o|r `0c Q˪9ݰU&$bHb .諘Ԁ!$@0gɒ`mY L J ` Zj+i`ڙ %"_;޶{3tMhq@poWq[$I;dw%E;y\ϢZyǓ#si9@0,7[ $ᩬ$UJj}%XɆHS#61CfEaNT T|ͦ`'rCDqVepmFl6fI l憓 u>,S@ fl>GCޓ/FZicaahbm_uޢF@ Ԧ@Psy.U j2-J_啲pSȱt 8l"eBPԬ)C3Nmvş4_ &2UOFR)nIdQZ 6"cfZX2I6H'h#ΒgKgĎ^A$X7j,D:l nF?VuXg˛Zasf||^. Y绫z,%"Tw%߯]%)^C්}GZ6:3mwl P34 zTYy`ބשflH|hs>Qvg6= pe`#C7{,)*o)I4^ژ|vMgn\dx3boVfޑ77oz6S3'&.,sHX Q4dU0uK 9AƮ;r ^cETr2W{oAf-gv)p= V"|?RW݉{b)A zݧ}s"emM"YO`$fQr`;0gyy\< Չ ,BA%F6s, vuY64$q+ge0"\aࡅfN2™x5H.3PWz~E]1ޕ6O#O.̃yjE]N#߰.T)4Ů5 d R6NNO@I$q rŮ¹%mk౫"<=:O*vQ,;=cWqqUVȈшx",yjT<(q\t=A`clUR $ǗpU(5\kD4z(B t[CQv:߻ӱ;dδ?ymUG"j %rϨDM+%7j[~?Y\ζyY'OhŶ[~v,B^ӻٿ5_9[*77w:ӓ_=Lv懾ؕ_^.z2:x/m fv:0r=X&wCVoc8{?ݗ~by "K>n[_Noh^5O[#чnijh/vvּKj4߈nmV z n $ ("#&l :?UsM]h0$<CE@8F}V?%T@e{:0h/ĥoD3\LQ)irޠhR;A9_tD] aANǼFIM^wsW-7z1w圾$IZι%Ck_Qϧ}[>qכeQ|Ɵ=Y垥k=o*Nf(qo7U~mS#r1$"Lɖ$N0E ɔQ8VR,I{RTkLI{R1Q`E\}[ xY'˴3$2|:+)%UWS;[+( =3' *зB1>y ?m߳_ Lm3;/8u~՛Mrgn4(aP`wl R.@nIZbf =Gk'\z0gJ\nuq%UEEXY;\&.K΍,7f*s'wI\aM?yԍGwlM'c3/qqcfZbldH$j9Td^ ^:ItSR٢4 #"HG$/C@WL5*L:YFXWXZm92 [ o푉%OrsLQMC9DL Q!\x۝5evD <ШC؀<<,1 $8d,Nr%- byo ޜ3=s"{zf'J>g>ZDlt X5G_ |. ҏ9ι/%Od [?1m=:, [1IttIveG]sd^gfW? wovar/home/core/zuul-output/logs/kubelet.log0000644000000000000000003371116715150260674017714 0ustar rootrootFeb 27 07:49:00 crc systemd[1]: Starting Kubernetes Kubelet... Feb 27 07:49:01 crc restorecon[4566]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 27 07:49:01 crc restorecon[4566]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Feb 27 07:49:01 crc restorecon[4566]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Feb 27 07:49:02 crc kubenswrapper[4612]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Feb 27 07:49:02 crc kubenswrapper[4612]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Feb 27 07:49:02 crc kubenswrapper[4612]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Feb 27 07:49:02 crc kubenswrapper[4612]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Feb 27 07:49:02 crc kubenswrapper[4612]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Feb 27 07:49:02 crc kubenswrapper[4612]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.605166 4612 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.610388 4612 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.610412 4612 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.610417 4612 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.610421 4612 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.610431 4612 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.610436 4612 feature_gate.go:330] unrecognized feature gate: Example Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.610441 4612 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.610447 4612 feature_gate.go:330] unrecognized feature gate: PinnedImages Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.610453 4612 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.610459 4612 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.610462 4612 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.610467 4612 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.610472 4612 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.610476 4612 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.610481 4612 feature_gate.go:330] unrecognized feature gate: SignatureStores Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.610485 4612 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.610489 4612 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.610492 4612 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.610497 4612 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.610501 4612 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.610504 4612 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.610509 4612 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.610513 4612 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.610517 4612 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.610522 4612 feature_gate.go:330] unrecognized feature gate: NewOLM Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.610526 4612 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.610531 4612 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.610535 4612 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.610539 4612 feature_gate.go:330] unrecognized feature gate: InsightsConfig Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.610544 4612 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.610550 4612 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.610555 4612 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.610560 4612 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.610564 4612 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.610569 4612 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.610574 4612 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.610578 4612 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.610583 4612 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.610588 4612 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.610592 4612 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.610597 4612 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.610603 4612 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.610608 4612 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.610612 4612 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.610617 4612 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.610622 4612 feature_gate.go:330] unrecognized feature gate: GatewayAPI Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.610626 4612 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.610630 4612 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.610634 4612 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.610638 4612 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.610642 4612 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.610646 4612 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.610650 4612 feature_gate.go:330] unrecognized feature gate: PlatformOperators Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.610654 4612 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.610658 4612 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.610662 4612 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.610666 4612 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.610671 4612 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.610674 4612 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.610681 4612 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.610686 4612 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.610709 4612 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.610714 4612 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.610719 4612 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.610724 4612 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.610729 4612 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.610756 4612 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.610763 4612 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.610769 4612 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.610777 4612 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.610783 4612 feature_gate.go:330] unrecognized feature gate: OVNObservability Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.612051 4612 flags.go:64] FLAG: --address="0.0.0.0" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.612069 4612 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.612081 4612 flags.go:64] FLAG: --anonymous-auth="true" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.612089 4612 flags.go:64] FLAG: --application-metrics-count-limit="100" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.612098 4612 flags.go:64] FLAG: --authentication-token-webhook="false" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.612105 4612 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.612113 4612 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.612122 4612 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.612128 4612 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.612134 4612 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.612140 4612 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.612146 4612 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.612151 4612 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.612156 4612 flags.go:64] FLAG: --cgroup-root="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.612161 4612 flags.go:64] FLAG: --cgroups-per-qos="true" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.612165 4612 flags.go:64] FLAG: --client-ca-file="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.612170 4612 flags.go:64] FLAG: --cloud-config="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.612175 4612 flags.go:64] FLAG: --cloud-provider="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.612180 4612 flags.go:64] FLAG: --cluster-dns="[]" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.612185 4612 flags.go:64] FLAG: --cluster-domain="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.612190 4612 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.612195 4612 flags.go:64] FLAG: --config-dir="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.612200 4612 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.612205 4612 flags.go:64] FLAG: --container-log-max-files="5" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.612212 4612 flags.go:64] FLAG: --container-log-max-size="10Mi" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.612217 4612 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.612223 4612 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.612228 4612 flags.go:64] FLAG: --containerd-namespace="k8s.io" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.612233 4612 flags.go:64] FLAG: --contention-profiling="false" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.612239 4612 flags.go:64] FLAG: --cpu-cfs-quota="true" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.612245 4612 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.612250 4612 flags.go:64] FLAG: --cpu-manager-policy="none" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.612255 4612 flags.go:64] FLAG: --cpu-manager-policy-options="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.612262 4612 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.612266 4612 flags.go:64] FLAG: --enable-controller-attach-detach="true" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.612271 4612 flags.go:64] FLAG: --enable-debugging-handlers="true" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.612276 4612 flags.go:64] FLAG: --enable-load-reader="false" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.612281 4612 flags.go:64] FLAG: --enable-server="true" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.612286 4612 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.612292 4612 flags.go:64] FLAG: --event-burst="100" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.612297 4612 flags.go:64] FLAG: --event-qps="50" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.612302 4612 flags.go:64] FLAG: --event-storage-age-limit="default=0" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.612307 4612 flags.go:64] FLAG: --event-storage-event-limit="default=0" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.612312 4612 flags.go:64] FLAG: --eviction-hard="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.612318 4612 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.612323 4612 flags.go:64] FLAG: --eviction-minimum-reclaim="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.612328 4612 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.612333 4612 flags.go:64] FLAG: --eviction-soft="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.612338 4612 flags.go:64] FLAG: --eviction-soft-grace-period="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.612342 4612 flags.go:64] FLAG: --exit-on-lock-contention="false" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.612347 4612 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.612351 4612 flags.go:64] FLAG: --experimental-mounter-path="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.612356 4612 flags.go:64] FLAG: --fail-cgroupv1="false" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.612361 4612 flags.go:64] FLAG: --fail-swap-on="true" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.612365 4612 flags.go:64] FLAG: --feature-gates="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.612371 4612 flags.go:64] FLAG: --file-check-frequency="20s" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.612376 4612 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.612381 4612 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.612385 4612 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.612390 4612 flags.go:64] FLAG: --healthz-port="10248" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.612395 4612 flags.go:64] FLAG: --help="false" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.612402 4612 flags.go:64] FLAG: --hostname-override="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.612408 4612 flags.go:64] FLAG: --housekeeping-interval="10s" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.612413 4612 flags.go:64] FLAG: --http-check-frequency="20s" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.612418 4612 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.612422 4612 flags.go:64] FLAG: --image-credential-provider-config="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.612427 4612 flags.go:64] FLAG: --image-gc-high-threshold="85" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.612431 4612 flags.go:64] FLAG: --image-gc-low-threshold="80" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.612436 4612 flags.go:64] FLAG: --image-service-endpoint="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.612440 4612 flags.go:64] FLAG: --kernel-memcg-notification="false" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.612445 4612 flags.go:64] FLAG: --kube-api-burst="100" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.612450 4612 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.612455 4612 flags.go:64] FLAG: --kube-api-qps="50" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.612459 4612 flags.go:64] FLAG: --kube-reserved="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.612464 4612 flags.go:64] FLAG: --kube-reserved-cgroup="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.612468 4612 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.612473 4612 flags.go:64] FLAG: --kubelet-cgroups="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.612477 4612 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.612482 4612 flags.go:64] FLAG: --lock-file="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.612487 4612 flags.go:64] FLAG: --log-cadvisor-usage="false" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.612492 4612 flags.go:64] FLAG: --log-flush-frequency="5s" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.612497 4612 flags.go:64] FLAG: --log-json-info-buffer-size="0" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.612505 4612 flags.go:64] FLAG: --log-json-split-stream="false" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.612510 4612 flags.go:64] FLAG: --log-text-info-buffer-size="0" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.612515 4612 flags.go:64] FLAG: --log-text-split-stream="false" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.612520 4612 flags.go:64] FLAG: --logging-format="text" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.612524 4612 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.612529 4612 flags.go:64] FLAG: --make-iptables-util-chains="true" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.612534 4612 flags.go:64] FLAG: --manifest-url="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.612538 4612 flags.go:64] FLAG: --manifest-url-header="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.612545 4612 flags.go:64] FLAG: --max-housekeeping-interval="15s" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.612549 4612 flags.go:64] FLAG: --max-open-files="1000000" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.612556 4612 flags.go:64] FLAG: --max-pods="110" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.612561 4612 flags.go:64] FLAG: --maximum-dead-containers="-1" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.612567 4612 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.612572 4612 flags.go:64] FLAG: --memory-manager-policy="None" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.612577 4612 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.612581 4612 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.612587 4612 flags.go:64] FLAG: --node-ip="192.168.126.11" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.612592 4612 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.612609 4612 flags.go:64] FLAG: --node-status-max-images="50" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.612614 4612 flags.go:64] FLAG: --node-status-update-frequency="10s" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.612619 4612 flags.go:64] FLAG: --oom-score-adj="-999" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.612624 4612 flags.go:64] FLAG: --pod-cidr="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.612628 4612 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.612637 4612 flags.go:64] FLAG: --pod-manifest-path="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.612641 4612 flags.go:64] FLAG: --pod-max-pids="-1" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.612646 4612 flags.go:64] FLAG: --pods-per-core="0" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.612650 4612 flags.go:64] FLAG: --port="10250" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.612655 4612 flags.go:64] FLAG: --protect-kernel-defaults="false" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.612660 4612 flags.go:64] FLAG: --provider-id="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.612665 4612 flags.go:64] FLAG: --qos-reserved="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.612669 4612 flags.go:64] FLAG: --read-only-port="10255" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.612674 4612 flags.go:64] FLAG: --register-node="true" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.612679 4612 flags.go:64] FLAG: --register-schedulable="true" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.612684 4612 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.612710 4612 flags.go:64] FLAG: --registry-burst="10" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.612716 4612 flags.go:64] FLAG: --registry-qps="5" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.612721 4612 flags.go:64] FLAG: --reserved-cpus="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.612726 4612 flags.go:64] FLAG: --reserved-memory="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.612734 4612 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.612739 4612 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.612744 4612 flags.go:64] FLAG: --rotate-certificates="false" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.612749 4612 flags.go:64] FLAG: --rotate-server-certificates="false" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.612754 4612 flags.go:64] FLAG: --runonce="false" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.612759 4612 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.612764 4612 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.612777 4612 flags.go:64] FLAG: --seccomp-default="false" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.612782 4612 flags.go:64] FLAG: --serialize-image-pulls="true" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.612786 4612 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.612792 4612 flags.go:64] FLAG: --storage-driver-db="cadvisor" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.612796 4612 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.612801 4612 flags.go:64] FLAG: --storage-driver-password="root" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.612806 4612 flags.go:64] FLAG: --storage-driver-secure="false" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.612812 4612 flags.go:64] FLAG: --storage-driver-table="stats" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.612817 4612 flags.go:64] FLAG: --storage-driver-user="root" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.612821 4612 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.612826 4612 flags.go:64] FLAG: --sync-frequency="1m0s" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.612831 4612 flags.go:64] FLAG: --system-cgroups="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.612835 4612 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.612844 4612 flags.go:64] FLAG: --system-reserved-cgroup="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.612848 4612 flags.go:64] FLAG: --tls-cert-file="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.612853 4612 flags.go:64] FLAG: --tls-cipher-suites="[]" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.612858 4612 flags.go:64] FLAG: --tls-min-version="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.612863 4612 flags.go:64] FLAG: --tls-private-key-file="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.612867 4612 flags.go:64] FLAG: --topology-manager-policy="none" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.612872 4612 flags.go:64] FLAG: --topology-manager-policy-options="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.612877 4612 flags.go:64] FLAG: --topology-manager-scope="container" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.612882 4612 flags.go:64] FLAG: --v="2" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.612888 4612 flags.go:64] FLAG: --version="false" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.612895 4612 flags.go:64] FLAG: --vmodule="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.612902 4612 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.612908 4612 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.613045 4612 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.613050 4612 feature_gate.go:330] unrecognized feature gate: SignatureStores Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.613055 4612 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.613059 4612 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.613064 4612 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.613069 4612 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.613077 4612 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.613081 4612 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.613085 4612 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.613089 4612 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.613094 4612 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.613098 4612 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.613102 4612 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.613106 4612 feature_gate.go:330] unrecognized feature gate: PlatformOperators Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.613110 4612 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.613115 4612 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.613120 4612 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.613125 4612 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.613130 4612 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.613134 4612 feature_gate.go:330] unrecognized feature gate: InsightsConfig Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.613141 4612 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.613146 4612 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.613151 4612 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.613155 4612 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.613160 4612 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.613164 4612 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.613168 4612 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.613172 4612 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.613176 4612 feature_gate.go:330] unrecognized feature gate: GatewayAPI Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.613180 4612 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.613184 4612 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.613188 4612 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.613192 4612 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.613196 4612 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.613200 4612 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.613204 4612 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.613209 4612 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.613212 4612 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.613219 4612 feature_gate.go:330] unrecognized feature gate: Example Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.613223 4612 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.613229 4612 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.613234 4612 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.613239 4612 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.613244 4612 feature_gate.go:330] unrecognized feature gate: PinnedImages Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.613248 4612 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.613253 4612 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.613257 4612 feature_gate.go:330] unrecognized feature gate: OVNObservability Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.613261 4612 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.613265 4612 feature_gate.go:330] unrecognized feature gate: NewOLM Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.613269 4612 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.613273 4612 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.613277 4612 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.613282 4612 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.613287 4612 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.613292 4612 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.613296 4612 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.613300 4612 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.613304 4612 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.613308 4612 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.613312 4612 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.613316 4612 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.613321 4612 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.613325 4612 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.613329 4612 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.613333 4612 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.613337 4612 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.613341 4612 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.613346 4612 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.613350 4612 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.613354 4612 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.613361 4612 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.613376 4612 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.626931 4612 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.626981 4612 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.627102 4612 feature_gate.go:330] unrecognized feature gate: GatewayAPI Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.627115 4612 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.627123 4612 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.627130 4612 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.627138 4612 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.627145 4612 feature_gate.go:330] unrecognized feature gate: InsightsConfig Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.627151 4612 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.627158 4612 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.627164 4612 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.627171 4612 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.627178 4612 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.627185 4612 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.627192 4612 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.627199 4612 feature_gate.go:330] unrecognized feature gate: Example Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.627205 4612 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.627212 4612 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.627218 4612 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.627224 4612 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.627230 4612 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.627239 4612 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.627247 4612 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.627254 4612 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.627260 4612 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.627267 4612 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.627273 4612 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.627280 4612 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.627286 4612 feature_gate.go:330] unrecognized feature gate: NewOLM Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.627299 4612 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.627306 4612 feature_gate.go:330] unrecognized feature gate: SignatureStores Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.627315 4612 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.627325 4612 feature_gate.go:330] unrecognized feature gate: OVNObservability Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.627333 4612 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.627342 4612 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.627349 4612 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.627355 4612 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.627362 4612 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.627369 4612 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.627376 4612 feature_gate.go:330] unrecognized feature gate: PinnedImages Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.627383 4612 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.627389 4612 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.627395 4612 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.627402 4612 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.627408 4612 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.627414 4612 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.627420 4612 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.627427 4612 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.627434 4612 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.627441 4612 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.627447 4612 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.627453 4612 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.627460 4612 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.627466 4612 feature_gate.go:330] unrecognized feature gate: PlatformOperators Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.627473 4612 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.627480 4612 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.627486 4612 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.627493 4612 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.627500 4612 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.627506 4612 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.627515 4612 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.627523 4612 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.627532 4612 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.627539 4612 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.627545 4612 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.627552 4612 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.627562 4612 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.627572 4612 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.627579 4612 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.627586 4612 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.627592 4612 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.627600 4612 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.627607 4612 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.627620 4612 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.627863 4612 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.627878 4612 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.627886 4612 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.627893 4612 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.627901 4612 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.627908 4612 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.627915 4612 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.627922 4612 feature_gate.go:330] unrecognized feature gate: SignatureStores Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.627928 4612 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.627935 4612 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.627941 4612 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.627947 4612 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.627954 4612 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.627960 4612 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.627967 4612 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.627973 4612 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.627980 4612 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.627986 4612 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.627992 4612 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.627999 4612 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.628006 4612 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.628013 4612 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.628020 4612 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.628028 4612 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.628038 4612 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.628047 4612 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.628053 4612 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.628061 4612 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.628068 4612 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.628076 4612 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.628083 4612 feature_gate.go:330] unrecognized feature gate: NewOLM Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.628090 4612 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.628096 4612 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.628103 4612 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.628109 4612 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.628115 4612 feature_gate.go:330] unrecognized feature gate: PinnedImages Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.628121 4612 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.628129 4612 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.628136 4612 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.628143 4612 feature_gate.go:330] unrecognized feature gate: InsightsConfig Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.628149 4612 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.628156 4612 feature_gate.go:330] unrecognized feature gate: GatewayAPI Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.628162 4612 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.628168 4612 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.628174 4612 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.628180 4612 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.628187 4612 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.628193 4612 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.628199 4612 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.628206 4612 feature_gate.go:330] unrecognized feature gate: Example Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.628212 4612 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.628218 4612 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.628226 4612 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.628233 4612 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.628240 4612 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.628246 4612 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.628252 4612 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.628259 4612 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.628268 4612 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.628276 4612 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.628283 4612 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.628291 4612 feature_gate.go:330] unrecognized feature gate: OVNObservability Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.628299 4612 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.628306 4612 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.628313 4612 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.628319 4612 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.628326 4612 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.628333 4612 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.628339 4612 feature_gate.go:330] unrecognized feature gate: PlatformOperators Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.628344 4612 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.628349 4612 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.628358 4612 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.628605 4612 server.go:940] "Client rotation is on, will bootstrap in background" Feb 27 07:49:02 crc kubenswrapper[4612]: E0227 07:49:02.633147 4612 bootstrap.go:266] "Unhandled Error" err="part of the existing bootstrap client certificate in /var/lib/kubelet/kubeconfig is expired: 2026-02-24 05:52:08 +0000 UTC" logger="UnhandledError" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.637020 4612 bootstrap.go:101] "Use the bootstrap credentials to request a cert, and set kubeconfig to point to the certificate dir" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.637164 4612 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.640211 4612 server.go:997] "Starting client certificate rotation" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.640251 4612 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.640458 4612 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.674239 4612 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.676122 4612 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Feb 27 07:49:02 crc kubenswrapper[4612]: E0227 07:49:02.677815 4612 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 38.102.83.158:6443: connect: connection refused" logger="UnhandledError" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.689309 4612 log.go:25] "Validated CRI v1 runtime API" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.728903 4612 log.go:25] "Validated CRI v1 image API" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.730773 4612 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.736224 4612 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2026-02-27-07-43-41-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.736372 4612 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:43 fsType:tmpfs blockSize:0}] Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.750203 4612 manager.go:217] Machine: {Timestamp:2026-02-27 07:49:02.748040728 +0000 UTC m=+0.601970726 CPUVendorID:AuthenticAMD NumCores:8 NumPhysicalCores:1 NumSockets:8 CpuFrequency:2800000 MemoryCapacity:25199472640 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:6e98ef2c-c447-4ff8-86d3-9b32f302ee4b BootID:a8a53c94-4eff-42ba-8a96-6c53b380c5c4 Filesystems:[{Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:43 Capacity:1073741824 Type:vfs Inodes:3076107 HasInodes:true} {Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:12599734272 Type:vfs Inodes:3076107 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:5039894528 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:12599738368 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:2519945216 Type:vfs Inodes:615221 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:429496729600 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:27:d4:8c Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:27:d4:8c Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:d9:62:af Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:65:5b:61 Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:5d:74:0d Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:cd:c3:43 Speed:-1 Mtu:1496} {Name:eth10 MacAddress:02:07:ca:da:79:f9 Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:32:5d:12:94:0f:e1 Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:25199472640 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.750463 4612 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.750732 4612 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.751343 4612 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.751626 4612 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.751674 4612 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.753183 4612 topology_manager.go:138] "Creating topology manager with none policy" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.753212 4612 container_manager_linux.go:303] "Creating device plugin manager" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.754127 4612 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.754160 4612 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.754752 4612 state_mem.go:36] "Initialized new in-memory state store" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.755320 4612 server.go:1245] "Using root directory" path="/var/lib/kubelet" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.759803 4612 kubelet.go:418] "Attempting to sync node with API server" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.759834 4612 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.759860 4612 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.759879 4612 kubelet.go:324] "Adding apiserver pod source" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.759897 4612 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.764371 4612 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.765527 4612 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.766971 4612 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.158:6443: connect: connection refused Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.766993 4612 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.158:6443: connect: connection refused Feb 27 07:49:02 crc kubenswrapper[4612]: E0227 07:49:02.767080 4612 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.158:6443: connect: connection refused" logger="UnhandledError" Feb 27 07:49:02 crc kubenswrapper[4612]: E0227 07:49:02.767119 4612 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.158:6443: connect: connection refused" logger="UnhandledError" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.768672 4612 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.770436 4612 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.770480 4612 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.770496 4612 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.770510 4612 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.770543 4612 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.770558 4612 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.770573 4612 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.770596 4612 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.770645 4612 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.770664 4612 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.770686 4612 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.770726 4612 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.771611 4612 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.772455 4612 server.go:1280] "Started kubelet" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.772437 4612 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.158:6443: connect: connection refused Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.774725 4612 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.774941 4612 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.774994 4612 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Feb 27 07:49:02 crc systemd[1]: Started Kubernetes Kubelet. Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.775645 4612 volume_manager.go:287] "The desired_state_of_world populator starts" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.775664 4612 volume_manager.go:289] "Starting Kubelet Volume Manager" Feb 27 07:49:02 crc kubenswrapper[4612]: E0227 07:49:02.775823 4612 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.774776 4612 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.776620 4612 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.776909 4612 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.158:6443: connect: connection refused Feb 27 07:49:02 crc kubenswrapper[4612]: E0227 07:49:02.776979 4612 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.158:6443: connect: connection refused" logger="UnhandledError" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.777283 4612 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.778536 4612 server.go:460] "Adding debug handlers to kubelet server" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.781800 4612 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.781829 4612 factory.go:55] Registering systemd factory Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.781845 4612 factory.go:221] Registration of the systemd container factory successfully Feb 27 07:49:02 crc kubenswrapper[4612]: E0227 07:49:02.782046 4612 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.158:6443: connect: connection refused" interval="200ms" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.782168 4612 factory.go:153] Registering CRI-O factory Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.782186 4612 factory.go:221] Registration of the crio container factory successfully Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.782225 4612 factory.go:103] Registering Raw factory Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.782250 4612 manager.go:1196] Started watching for new ooms in manager Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.783877 4612 manager.go:319] Starting recovery of all containers Feb 27 07:49:02 crc kubenswrapper[4612]: E0227 07:49:02.789405 4612 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.158:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.18980afec884cc0b default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 07:49:02.772407307 +0000 UTC m=+0.626337345,LastTimestamp:2026-02-27 07:49:02.772407307 +0000 UTC m=+0.626337345,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.810129 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.810198 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.810211 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.810223 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.810233 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.810250 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.810275 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.810293 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.810328 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.810345 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.810357 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.810371 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.810380 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.810394 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.810406 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.810454 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.810463 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.810472 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.810480 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.810491 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.810503 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.810513 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.810532 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.810576 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.810733 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.810747 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.810761 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.810774 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.810788 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.810803 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.810817 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.810832 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.810846 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.810872 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.810883 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.810892 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.810904 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.810916 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.810928 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.810940 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.810957 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.810973 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.810985 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.810996 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.811008 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.811020 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.811033 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.811045 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.811057 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.811069 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.811081 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.811090 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.811119 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.811132 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.811149 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.811159 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.811170 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.811183 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.811194 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.811205 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.811222 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.811238 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.811251 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.811262 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.811274 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.811286 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.811296 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.811306 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.811318 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.811330 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.811373 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.811385 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.811397 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.811408 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.811419 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.811430 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.811441 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.811454 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.811469 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.811481 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.811493 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.811504 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.811516 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.811529 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.811541 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.811552 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.811564 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.811578 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.811594 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.811607 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.811619 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.811629 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.811641 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.811653 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.811666 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.811681 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.811747 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.811758 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.811769 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.811800 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.811822 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.811840 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.811875 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.811886 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.811937 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.811955 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.811972 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.811988 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.812001 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.812013 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.812029 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.812040 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.812052 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.812064 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.812077 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.812089 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.812102 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.812114 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.812128 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.812139 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.812149 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.812162 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.812176 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.812190 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.812201 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.812214 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.812228 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.812241 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.812253 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.812264 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.812276 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.812287 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.812300 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.812311 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.812324 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.812336 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.812348 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.812359 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.812369 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.812381 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.812394 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.812409 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.812428 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.812439 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.812452 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.812463 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.812475 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.812487 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.812501 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.812512 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.812542 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.812556 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.812568 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.812589 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.812606 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.812619 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.812634 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.812646 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.812659 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.812674 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.812705 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.812717 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.812728 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.812741 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.812754 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.812822 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.812837 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.812850 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.812871 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.812893 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.812909 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.812923 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.812936 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.812947 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.812962 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.812973 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.812985 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.812998 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.813011 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.813024 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.813037 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.813048 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.813063 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.813075 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.813087 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.813098 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.814632 4612 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.814662 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.814680 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.814713 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.814727 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.814742 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.814756 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.814770 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.814783 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.814795 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.814809 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.814823 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.814840 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.814854 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.814865 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.814876 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.814889 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.814900 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.814910 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.814921 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.814934 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.814948 4612 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.814958 4612 reconstruct.go:97] "Volume reconstruction finished" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.814966 4612 reconciler.go:26] "Reconciler: start to sync state" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.828560 4612 manager.go:324] Recovery completed Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.839296 4612 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.841532 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.841611 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.841627 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.842878 4612 cpu_manager.go:225] "Starting CPU manager" policy="none" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.842909 4612 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.842965 4612 state_mem.go:36] "Initialized new in-memory state store" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.848528 4612 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.851609 4612 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.851678 4612 status_manager.go:217] "Starting to sync pod status with apiserver" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.851733 4612 kubelet.go:2335] "Starting kubelet main sync loop" Feb 27 07:49:02 crc kubenswrapper[4612]: E0227 07:49:02.851800 4612 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Feb 27 07:49:02 crc kubenswrapper[4612]: W0227 07:49:02.853510 4612 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.158:6443: connect: connection refused Feb 27 07:49:02 crc kubenswrapper[4612]: E0227 07:49:02.853587 4612 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.158:6443: connect: connection refused" logger="UnhandledError" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.862060 4612 policy_none.go:49] "None policy: Start" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.863163 4612 memory_manager.go:170] "Starting memorymanager" policy="None" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.863273 4612 state_mem.go:35] "Initializing new in-memory state store" Feb 27 07:49:02 crc kubenswrapper[4612]: E0227 07:49:02.876194 4612 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.912198 4612 manager.go:334] "Starting Device Plugin manager" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.912630 4612 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.912672 4612 server.go:79] "Starting device plugin registration server" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.913949 4612 eviction_manager.go:189] "Eviction manager: starting control loop" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.913973 4612 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.914297 4612 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.914417 4612 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.914427 4612 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Feb 27 07:49:02 crc kubenswrapper[4612]: E0227 07:49:02.921861 4612 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.952603 4612 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc"] Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.952774 4612 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.954113 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.954146 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.954154 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.954273 4612 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.955196 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.955219 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.955227 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.955556 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.955637 4612 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.955759 4612 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.955779 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.955948 4612 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.957101 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.957145 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.957156 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.957173 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.957239 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.957261 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.957355 4612 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.957632 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.957770 4612 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.958343 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.958389 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.958399 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.959201 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.959227 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.959235 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.959204 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.959321 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.959348 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.959541 4612 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.959790 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.960005 4612 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.960473 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.960499 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.960508 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.960677 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.960740 4612 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.961019 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.961052 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.961059 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.961923 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.961945 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 07:49:02 crc kubenswrapper[4612]: I0227 07:49:02.961953 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 07:49:02 crc kubenswrapper[4612]: E0227 07:49:02.983622 4612 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.158:6443: connect: connection refused" interval="400ms" Feb 27 07:49:03 crc kubenswrapper[4612]: I0227 07:49:03.014195 4612 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 07:49:03 crc kubenswrapper[4612]: I0227 07:49:03.015838 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 07:49:03 crc kubenswrapper[4612]: I0227 07:49:03.015905 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 07:49:03 crc kubenswrapper[4612]: I0227 07:49:03.015919 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 07:49:03 crc kubenswrapper[4612]: I0227 07:49:03.015978 4612 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 27 07:49:03 crc kubenswrapper[4612]: E0227 07:49:03.016979 4612 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.158:6443: connect: connection refused" node="crc" Feb 27 07:49:03 crc kubenswrapper[4612]: I0227 07:49:03.018621 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 27 07:49:03 crc kubenswrapper[4612]: I0227 07:49:03.018662 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 27 07:49:03 crc kubenswrapper[4612]: I0227 07:49:03.018709 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 27 07:49:03 crc kubenswrapper[4612]: I0227 07:49:03.018734 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Feb 27 07:49:03 crc kubenswrapper[4612]: I0227 07:49:03.018759 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 27 07:49:03 crc kubenswrapper[4612]: I0227 07:49:03.018801 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 27 07:49:03 crc kubenswrapper[4612]: I0227 07:49:03.018869 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 27 07:49:03 crc kubenswrapper[4612]: I0227 07:49:03.018921 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 27 07:49:03 crc kubenswrapper[4612]: I0227 07:49:03.018974 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 27 07:49:03 crc kubenswrapper[4612]: I0227 07:49:03.019004 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 27 07:49:03 crc kubenswrapper[4612]: I0227 07:49:03.019253 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 27 07:49:03 crc kubenswrapper[4612]: I0227 07:49:03.019278 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Feb 27 07:49:03 crc kubenswrapper[4612]: I0227 07:49:03.019301 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 27 07:49:03 crc kubenswrapper[4612]: I0227 07:49:03.019337 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 27 07:49:03 crc kubenswrapper[4612]: I0227 07:49:03.019383 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 27 07:49:03 crc kubenswrapper[4612]: I0227 07:49:03.121179 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 27 07:49:03 crc kubenswrapper[4612]: I0227 07:49:03.121486 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 27 07:49:03 crc kubenswrapper[4612]: I0227 07:49:03.121568 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 27 07:49:03 crc kubenswrapper[4612]: I0227 07:49:03.121408 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 27 07:49:03 crc kubenswrapper[4612]: I0227 07:49:03.121614 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 27 07:49:03 crc kubenswrapper[4612]: I0227 07:49:03.121634 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 27 07:49:03 crc kubenswrapper[4612]: I0227 07:49:03.121543 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 27 07:49:03 crc kubenswrapper[4612]: I0227 07:49:03.121737 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 27 07:49:03 crc kubenswrapper[4612]: I0227 07:49:03.121813 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 27 07:49:03 crc kubenswrapper[4612]: I0227 07:49:03.121816 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 27 07:49:03 crc kubenswrapper[4612]: I0227 07:49:03.121835 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Feb 27 07:49:03 crc kubenswrapper[4612]: I0227 07:49:03.121877 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 27 07:49:03 crc kubenswrapper[4612]: I0227 07:49:03.121901 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 27 07:49:03 crc kubenswrapper[4612]: I0227 07:49:03.121927 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 27 07:49:03 crc kubenswrapper[4612]: I0227 07:49:03.121953 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 27 07:49:03 crc kubenswrapper[4612]: I0227 07:49:03.121955 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 27 07:49:03 crc kubenswrapper[4612]: I0227 07:49:03.121979 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 27 07:49:03 crc kubenswrapper[4612]: I0227 07:49:03.122023 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 27 07:49:03 crc kubenswrapper[4612]: I0227 07:49:03.121996 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 27 07:49:03 crc kubenswrapper[4612]: I0227 07:49:03.122006 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 27 07:49:03 crc kubenswrapper[4612]: I0227 07:49:03.121944 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 27 07:49:03 crc kubenswrapper[4612]: I0227 07:49:03.122069 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 27 07:49:03 crc kubenswrapper[4612]: I0227 07:49:03.122094 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 27 07:49:03 crc kubenswrapper[4612]: I0227 07:49:03.122121 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Feb 27 07:49:03 crc kubenswrapper[4612]: I0227 07:49:03.122153 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 27 07:49:03 crc kubenswrapper[4612]: I0227 07:49:03.122206 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 27 07:49:03 crc kubenswrapper[4612]: I0227 07:49:03.122822 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Feb 27 07:49:03 crc kubenswrapper[4612]: I0227 07:49:03.122882 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Feb 27 07:49:03 crc kubenswrapper[4612]: I0227 07:49:03.122910 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 27 07:49:03 crc kubenswrapper[4612]: I0227 07:49:03.122980 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Feb 27 07:49:03 crc kubenswrapper[4612]: I0227 07:49:03.218167 4612 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 07:49:03 crc kubenswrapper[4612]: I0227 07:49:03.220426 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 07:49:03 crc kubenswrapper[4612]: I0227 07:49:03.220557 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 07:49:03 crc kubenswrapper[4612]: I0227 07:49:03.220662 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 07:49:03 crc kubenswrapper[4612]: I0227 07:49:03.220746 4612 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 27 07:49:03 crc kubenswrapper[4612]: E0227 07:49:03.221681 4612 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.158:6443: connect: connection refused" node="crc" Feb 27 07:49:03 crc kubenswrapper[4612]: I0227 07:49:03.291893 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Feb 27 07:49:03 crc kubenswrapper[4612]: I0227 07:49:03.309144 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Feb 27 07:49:03 crc kubenswrapper[4612]: I0227 07:49:03.322838 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 27 07:49:03 crc kubenswrapper[4612]: I0227 07:49:03.345981 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 27 07:49:03 crc kubenswrapper[4612]: I0227 07:49:03.350835 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 27 07:49:03 crc kubenswrapper[4612]: W0227 07:49:03.354554 4612 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-91f2c917356e2d443d5ab787d6dba7a980c95c410f9d612e2dc7edaa044c92f5 WatchSource:0}: Error finding container 91f2c917356e2d443d5ab787d6dba7a980c95c410f9d612e2dc7edaa044c92f5: Status 404 returned error can't find the container with id 91f2c917356e2d443d5ab787d6dba7a980c95c410f9d612e2dc7edaa044c92f5 Feb 27 07:49:03 crc kubenswrapper[4612]: W0227 07:49:03.355667 4612 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-c58124dd3ca6aa0b1c95090b80cc10ecadbc18271b517170f4314fc40caefdb8 WatchSource:0}: Error finding container c58124dd3ca6aa0b1c95090b80cc10ecadbc18271b517170f4314fc40caefdb8: Status 404 returned error can't find the container with id c58124dd3ca6aa0b1c95090b80cc10ecadbc18271b517170f4314fc40caefdb8 Feb 27 07:49:03 crc kubenswrapper[4612]: W0227 07:49:03.364312 4612 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-7854b0778b75ba54343571a1bf710f9d92f8e109de7221dd54b11fa85b58ce14 WatchSource:0}: Error finding container 7854b0778b75ba54343571a1bf710f9d92f8e109de7221dd54b11fa85b58ce14: Status 404 returned error can't find the container with id 7854b0778b75ba54343571a1bf710f9d92f8e109de7221dd54b11fa85b58ce14 Feb 27 07:49:03 crc kubenswrapper[4612]: W0227 07:49:03.372880 4612 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-807c638c9942df5ae4944303b6f5b59d05264c97fd77e75b2682e2698a62e3c7 WatchSource:0}: Error finding container 807c638c9942df5ae4944303b6f5b59d05264c97fd77e75b2682e2698a62e3c7: Status 404 returned error can't find the container with id 807c638c9942df5ae4944303b6f5b59d05264c97fd77e75b2682e2698a62e3c7 Feb 27 07:49:03 crc kubenswrapper[4612]: W0227 07:49:03.380345 4612 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-0a39e69a8762ec922417a7223f3438df7e653b5277df931a2d6b872f26561419 WatchSource:0}: Error finding container 0a39e69a8762ec922417a7223f3438df7e653b5277df931a2d6b872f26561419: Status 404 returned error can't find the container with id 0a39e69a8762ec922417a7223f3438df7e653b5277df931a2d6b872f26561419 Feb 27 07:49:03 crc kubenswrapper[4612]: E0227 07:49:03.384664 4612 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.158:6443: connect: connection refused" interval="800ms" Feb 27 07:49:03 crc kubenswrapper[4612]: I0227 07:49:03.622856 4612 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 07:49:03 crc kubenswrapper[4612]: I0227 07:49:03.624379 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 07:49:03 crc kubenswrapper[4612]: I0227 07:49:03.624430 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 07:49:03 crc kubenswrapper[4612]: I0227 07:49:03.624452 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 07:49:03 crc kubenswrapper[4612]: I0227 07:49:03.624486 4612 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 27 07:49:03 crc kubenswrapper[4612]: E0227 07:49:03.625024 4612 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.158:6443: connect: connection refused" node="crc" Feb 27 07:49:03 crc kubenswrapper[4612]: I0227 07:49:03.774008 4612 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.158:6443: connect: connection refused Feb 27 07:49:03 crc kubenswrapper[4612]: W0227 07:49:03.812874 4612 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.158:6443: connect: connection refused Feb 27 07:49:03 crc kubenswrapper[4612]: E0227 07:49:03.813001 4612 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.158:6443: connect: connection refused" logger="UnhandledError" Feb 27 07:49:03 crc kubenswrapper[4612]: W0227 07:49:03.855229 4612 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.158:6443: connect: connection refused Feb 27 07:49:03 crc kubenswrapper[4612]: E0227 07:49:03.855335 4612 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.158:6443: connect: connection refused" logger="UnhandledError" Feb 27 07:49:03 crc kubenswrapper[4612]: W0227 07:49:03.856406 4612 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.158:6443: connect: connection refused Feb 27 07:49:03 crc kubenswrapper[4612]: E0227 07:49:03.856479 4612 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.158:6443: connect: connection refused" logger="UnhandledError" Feb 27 07:49:03 crc kubenswrapper[4612]: I0227 07:49:03.861103 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"91f2c917356e2d443d5ab787d6dba7a980c95c410f9d612e2dc7edaa044c92f5"} Feb 27 07:49:03 crc kubenswrapper[4612]: I0227 07:49:03.862880 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"0a39e69a8762ec922417a7223f3438df7e653b5277df931a2d6b872f26561419"} Feb 27 07:49:03 crc kubenswrapper[4612]: I0227 07:49:03.863947 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"807c638c9942df5ae4944303b6f5b59d05264c97fd77e75b2682e2698a62e3c7"} Feb 27 07:49:03 crc kubenswrapper[4612]: I0227 07:49:03.865325 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"7854b0778b75ba54343571a1bf710f9d92f8e109de7221dd54b11fa85b58ce14"} Feb 27 07:49:03 crc kubenswrapper[4612]: I0227 07:49:03.866940 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"c58124dd3ca6aa0b1c95090b80cc10ecadbc18271b517170f4314fc40caefdb8"} Feb 27 07:49:04 crc kubenswrapper[4612]: E0227 07:49:04.185873 4612 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.158:6443: connect: connection refused" interval="1.6s" Feb 27 07:49:04 crc kubenswrapper[4612]: W0227 07:49:04.309725 4612 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.158:6443: connect: connection refused Feb 27 07:49:04 crc kubenswrapper[4612]: E0227 07:49:04.309853 4612 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.158:6443: connect: connection refused" logger="UnhandledError" Feb 27 07:49:04 crc kubenswrapper[4612]: I0227 07:49:04.425804 4612 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 07:49:04 crc kubenswrapper[4612]: I0227 07:49:04.427672 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 07:49:04 crc kubenswrapper[4612]: I0227 07:49:04.427732 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 07:49:04 crc kubenswrapper[4612]: I0227 07:49:04.427746 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 07:49:04 crc kubenswrapper[4612]: I0227 07:49:04.427795 4612 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 27 07:49:04 crc kubenswrapper[4612]: E0227 07:49:04.428466 4612 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.158:6443: connect: connection refused" node="crc" Feb 27 07:49:04 crc kubenswrapper[4612]: I0227 07:49:04.764062 4612 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Feb 27 07:49:04 crc kubenswrapper[4612]: E0227 07:49:04.765057 4612 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 38.102.83.158:6443: connect: connection refused" logger="UnhandledError" Feb 27 07:49:04 crc kubenswrapper[4612]: I0227 07:49:04.773838 4612 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.158:6443: connect: connection refused Feb 27 07:49:04 crc kubenswrapper[4612]: I0227 07:49:04.871411 4612 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="bbae713c2a04c3aa252bc4ff1f22aac3d9469ed108c19b0f9486a5e73fc297c1" exitCode=0 Feb 27 07:49:04 crc kubenswrapper[4612]: I0227 07:49:04.871464 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"bbae713c2a04c3aa252bc4ff1f22aac3d9469ed108c19b0f9486a5e73fc297c1"} Feb 27 07:49:04 crc kubenswrapper[4612]: I0227 07:49:04.871510 4612 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 07:49:04 crc kubenswrapper[4612]: I0227 07:49:04.872433 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 07:49:04 crc kubenswrapper[4612]: I0227 07:49:04.872465 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 07:49:04 crc kubenswrapper[4612]: I0227 07:49:04.872475 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 07:49:04 crc kubenswrapper[4612]: I0227 07:49:04.872865 4612 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="d668c29bc9c37606836d548a6f61a7796e07c57559ea0726b328c62d61c0de36" exitCode=0 Feb 27 07:49:04 crc kubenswrapper[4612]: I0227 07:49:04.872927 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"d668c29bc9c37606836d548a6f61a7796e07c57559ea0726b328c62d61c0de36"} Feb 27 07:49:04 crc kubenswrapper[4612]: I0227 07:49:04.873147 4612 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 07:49:04 crc kubenswrapper[4612]: I0227 07:49:04.874412 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 07:49:04 crc kubenswrapper[4612]: I0227 07:49:04.874444 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 07:49:04 crc kubenswrapper[4612]: I0227 07:49:04.874457 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 07:49:04 crc kubenswrapper[4612]: I0227 07:49:04.874842 4612 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="2a61276e225aa5b0954a16d042ad7eb9a030c90c7cf8aeab772dfdfd00fbeff9" exitCode=0 Feb 27 07:49:04 crc kubenswrapper[4612]: I0227 07:49:04.874921 4612 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 07:49:04 crc kubenswrapper[4612]: I0227 07:49:04.874931 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"2a61276e225aa5b0954a16d042ad7eb9a030c90c7cf8aeab772dfdfd00fbeff9"} Feb 27 07:49:04 crc kubenswrapper[4612]: I0227 07:49:04.875911 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 07:49:04 crc kubenswrapper[4612]: I0227 07:49:04.875929 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 07:49:04 crc kubenswrapper[4612]: I0227 07:49:04.875938 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 07:49:04 crc kubenswrapper[4612]: I0227 07:49:04.880541 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"a968b1d57cf615a7995c051e012dfeb6bdc44886c27254fdda06414c7c01fc8f"} Feb 27 07:49:04 crc kubenswrapper[4612]: I0227 07:49:04.880597 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"01dd4b7697503da78d456884e291b17881c7f2a3027944b7871c0423a1d5771f"} Feb 27 07:49:04 crc kubenswrapper[4612]: I0227 07:49:04.880615 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"0a55081c1229c10dee891581d85f34e62567e1fd7f76a9c96d601a2e30bec08f"} Feb 27 07:49:04 crc kubenswrapper[4612]: I0227 07:49:04.880631 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"81b42fcb5b5a6dd8e9bbf8836e9a20cf3f603ac4913e90d4907df9338209f1be"} Feb 27 07:49:04 crc kubenswrapper[4612]: I0227 07:49:04.880750 4612 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 07:49:04 crc kubenswrapper[4612]: I0227 07:49:04.881597 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 07:49:04 crc kubenswrapper[4612]: I0227 07:49:04.881648 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 07:49:04 crc kubenswrapper[4612]: I0227 07:49:04.881663 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 07:49:04 crc kubenswrapper[4612]: I0227 07:49:04.885332 4612 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="61e0e770b36396d456078614494ef06c46bd43529d791281df3a621652328e21" exitCode=0 Feb 27 07:49:04 crc kubenswrapper[4612]: I0227 07:49:04.885384 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"61e0e770b36396d456078614494ef06c46bd43529d791281df3a621652328e21"} Feb 27 07:49:04 crc kubenswrapper[4612]: I0227 07:49:04.885446 4612 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 07:49:04 crc kubenswrapper[4612]: I0227 07:49:04.886422 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 07:49:04 crc kubenswrapper[4612]: I0227 07:49:04.886449 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 07:49:04 crc kubenswrapper[4612]: I0227 07:49:04.886461 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 07:49:04 crc kubenswrapper[4612]: I0227 07:49:04.889819 4612 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 07:49:04 crc kubenswrapper[4612]: I0227 07:49:04.890859 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 07:49:04 crc kubenswrapper[4612]: I0227 07:49:04.890895 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 07:49:04 crc kubenswrapper[4612]: I0227 07:49:04.890912 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 07:49:05 crc kubenswrapper[4612]: I0227 07:49:05.774577 4612 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.158:6443: connect: connection refused Feb 27 07:49:05 crc kubenswrapper[4612]: E0227 07:49:05.787351 4612 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.158:6443: connect: connection refused" interval="3.2s" Feb 27 07:49:05 crc kubenswrapper[4612]: I0227 07:49:05.890521 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"ed16a770df070ce00e9f07f63f64e7408967f7d05df4770ce69319a7b866676f"} Feb 27 07:49:05 crc kubenswrapper[4612]: I0227 07:49:05.890596 4612 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 07:49:05 crc kubenswrapper[4612]: I0227 07:49:05.891765 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 07:49:05 crc kubenswrapper[4612]: I0227 07:49:05.891804 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 07:49:05 crc kubenswrapper[4612]: I0227 07:49:05.891816 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 07:49:05 crc kubenswrapper[4612]: I0227 07:49:05.892596 4612 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="9d02841306e154f7b7ab85ea3030b388af7183f39c612305464fc95c18fa9d04" exitCode=0 Feb 27 07:49:05 crc kubenswrapper[4612]: I0227 07:49:05.892663 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"9d02841306e154f7b7ab85ea3030b388af7183f39c612305464fc95c18fa9d04"} Feb 27 07:49:05 crc kubenswrapper[4612]: I0227 07:49:05.892828 4612 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 07:49:05 crc kubenswrapper[4612]: I0227 07:49:05.894081 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 07:49:05 crc kubenswrapper[4612]: I0227 07:49:05.894106 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 07:49:05 crc kubenswrapper[4612]: I0227 07:49:05.894118 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 07:49:05 crc kubenswrapper[4612]: I0227 07:49:05.897660 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"c42dcb3dc7d1c476d588e82b497ab7811415ab49cdd8360a952cdbd21ed5831e"} Feb 27 07:49:05 crc kubenswrapper[4612]: I0227 07:49:05.897706 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"42c30fa5b88a7867ac23ff961b45f7e2ac03cc1c9ed855e527a190396be63f7e"} Feb 27 07:49:05 crc kubenswrapper[4612]: I0227 07:49:05.897724 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"25c772b9b9d21ca401274d7842e2456e3235bd7658942197fc3a83bbbb0ca9fe"} Feb 27 07:49:05 crc kubenswrapper[4612]: I0227 07:49:05.897797 4612 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 07:49:05 crc kubenswrapper[4612]: I0227 07:49:05.898827 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 07:49:05 crc kubenswrapper[4612]: I0227 07:49:05.898853 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 07:49:05 crc kubenswrapper[4612]: I0227 07:49:05.898861 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 07:49:05 crc kubenswrapper[4612]: I0227 07:49:05.901395 4612 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 07:49:05 crc kubenswrapper[4612]: I0227 07:49:05.901840 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"4fc98ba294403eadeb9ad6861f1a1932dc5298dfa790e50d15148773ad290e65"} Feb 27 07:49:05 crc kubenswrapper[4612]: I0227 07:49:05.901867 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"cc3a5e13e8220c8fa63de382ecc288979c3fcd96a0ee16c496af5e7a272b7355"} Feb 27 07:49:05 crc kubenswrapper[4612]: I0227 07:49:05.901879 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"a86d4fd23cb02a252fef6ed720b0f3eee818373ba88fe9c5586ced6566913221"} Feb 27 07:49:05 crc kubenswrapper[4612]: I0227 07:49:05.901887 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"ff9c7f0695586759696c95557a43826bf1198542aa46d532b4725d7d4fbc11f0"} Feb 27 07:49:05 crc kubenswrapper[4612]: I0227 07:49:05.902180 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 07:49:05 crc kubenswrapper[4612]: I0227 07:49:05.902200 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 07:49:05 crc kubenswrapper[4612]: I0227 07:49:05.902209 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 07:49:06 crc kubenswrapper[4612]: I0227 07:49:06.029394 4612 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 07:49:06 crc kubenswrapper[4612]: I0227 07:49:06.032893 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 07:49:06 crc kubenswrapper[4612]: I0227 07:49:06.032937 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 07:49:06 crc kubenswrapper[4612]: I0227 07:49:06.032948 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 07:49:06 crc kubenswrapper[4612]: I0227 07:49:06.032973 4612 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 27 07:49:06 crc kubenswrapper[4612]: E0227 07:49:06.033476 4612 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.158:6443: connect: connection refused" node="crc" Feb 27 07:49:06 crc kubenswrapper[4612]: W0227 07:49:06.126874 4612 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.158:6443: connect: connection refused Feb 27 07:49:06 crc kubenswrapper[4612]: E0227 07:49:06.126954 4612 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.158:6443: connect: connection refused" logger="UnhandledError" Feb 27 07:49:06 crc kubenswrapper[4612]: I0227 07:49:06.519984 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 27 07:49:06 crc kubenswrapper[4612]: W0227 07:49:06.521732 4612 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.158:6443: connect: connection refused Feb 27 07:49:06 crc kubenswrapper[4612]: E0227 07:49:06.521814 4612 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.158:6443: connect: connection refused" logger="UnhandledError" Feb 27 07:49:06 crc kubenswrapper[4612]: I0227 07:49:06.908899 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"28f728bc67c04c9dedfd4ec5ddc8fe577f3f5f59c9acaf1a1c68477b10a34f41"} Feb 27 07:49:06 crc kubenswrapper[4612]: I0227 07:49:06.909093 4612 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 07:49:06 crc kubenswrapper[4612]: I0227 07:49:06.911067 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 07:49:06 crc kubenswrapper[4612]: I0227 07:49:06.911135 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 07:49:06 crc kubenswrapper[4612]: I0227 07:49:06.911158 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 07:49:06 crc kubenswrapper[4612]: I0227 07:49:06.912359 4612 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="980bf6f6cf5e0af495ca38fc16c0bdca0a23a7b8c3270e69aebf33759ff32a7b" exitCode=0 Feb 27 07:49:06 crc kubenswrapper[4612]: I0227 07:49:06.912458 4612 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 07:49:06 crc kubenswrapper[4612]: I0227 07:49:06.912470 4612 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 07:49:06 crc kubenswrapper[4612]: I0227 07:49:06.912474 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"980bf6f6cf5e0af495ca38fc16c0bdca0a23a7b8c3270e69aebf33759ff32a7b"} Feb 27 07:49:06 crc kubenswrapper[4612]: I0227 07:49:06.912584 4612 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 07:49:06 crc kubenswrapper[4612]: I0227 07:49:06.913521 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 07:49:06 crc kubenswrapper[4612]: I0227 07:49:06.913540 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 07:49:06 crc kubenswrapper[4612]: I0227 07:49:06.913589 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 07:49:06 crc kubenswrapper[4612]: I0227 07:49:06.913603 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 07:49:06 crc kubenswrapper[4612]: I0227 07:49:06.913601 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 07:49:06 crc kubenswrapper[4612]: I0227 07:49:06.913633 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 07:49:06 crc kubenswrapper[4612]: I0227 07:49:06.913653 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 07:49:06 crc kubenswrapper[4612]: I0227 07:49:06.913564 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 07:49:06 crc kubenswrapper[4612]: I0227 07:49:06.913715 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 07:49:07 crc kubenswrapper[4612]: I0227 07:49:07.577427 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 27 07:49:07 crc kubenswrapper[4612]: I0227 07:49:07.577667 4612 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 07:49:07 crc kubenswrapper[4612]: I0227 07:49:07.579113 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 07:49:07 crc kubenswrapper[4612]: I0227 07:49:07.579201 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 07:49:07 crc kubenswrapper[4612]: I0227 07:49:07.579224 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 07:49:07 crc kubenswrapper[4612]: I0227 07:49:07.920371 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"92f331629ead8b0c7247e6336f55695e52c510a1a860184bc44c2146a053e89f"} Feb 27 07:49:07 crc kubenswrapper[4612]: I0227 07:49:07.920422 4612 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 27 07:49:07 crc kubenswrapper[4612]: I0227 07:49:07.920442 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"d620f04a7a1a0aef4b96baad46cfc1fa86eb4a081c37b897df170f2e5e22538a"} Feb 27 07:49:07 crc kubenswrapper[4612]: I0227 07:49:07.920461 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"77a75b084a41bdbcdcb2d37f552cf2b3f6ee8cadb3002690f12fca81da04831e"} Feb 27 07:49:07 crc kubenswrapper[4612]: I0227 07:49:07.920474 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"85667209d019e8fae802d7105b83afdd74da07ea36dcb9da0dd9778756ce1f18"} Feb 27 07:49:07 crc kubenswrapper[4612]: I0227 07:49:07.920478 4612 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 07:49:07 crc kubenswrapper[4612]: I0227 07:49:07.920621 4612 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 07:49:07 crc kubenswrapper[4612]: I0227 07:49:07.921883 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 07:49:07 crc kubenswrapper[4612]: I0227 07:49:07.921925 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 07:49:07 crc kubenswrapper[4612]: I0227 07:49:07.921940 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 07:49:07 crc kubenswrapper[4612]: I0227 07:49:07.922251 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 07:49:07 crc kubenswrapper[4612]: I0227 07:49:07.922305 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 07:49:07 crc kubenswrapper[4612]: I0227 07:49:07.922323 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 07:49:08 crc kubenswrapper[4612]: I0227 07:49:08.927837 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"5eb6e155b425e9c3dd7c266515bd468291d43ed2d20e8bb86184f250fa3a8874"} Feb 27 07:49:08 crc kubenswrapper[4612]: I0227 07:49:08.928041 4612 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 07:49:08 crc kubenswrapper[4612]: I0227 07:49:08.928904 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 07:49:08 crc kubenswrapper[4612]: I0227 07:49:08.928940 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 07:49:08 crc kubenswrapper[4612]: I0227 07:49:08.928956 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 07:49:08 crc kubenswrapper[4612]: I0227 07:49:08.969595 4612 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Feb 27 07:49:09 crc kubenswrapper[4612]: I0227 07:49:09.233607 4612 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 07:49:09 crc kubenswrapper[4612]: I0227 07:49:09.235013 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 07:49:09 crc kubenswrapper[4612]: I0227 07:49:09.235052 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 07:49:09 crc kubenswrapper[4612]: I0227 07:49:09.235064 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 07:49:09 crc kubenswrapper[4612]: I0227 07:49:09.235090 4612 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 27 07:49:09 crc kubenswrapper[4612]: I0227 07:49:09.270193 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 27 07:49:09 crc kubenswrapper[4612]: I0227 07:49:09.270993 4612 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 07:49:09 crc kubenswrapper[4612]: I0227 07:49:09.273215 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 07:49:09 crc kubenswrapper[4612]: I0227 07:49:09.273328 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 07:49:09 crc kubenswrapper[4612]: I0227 07:49:09.273361 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 07:49:09 crc kubenswrapper[4612]: I0227 07:49:09.930595 4612 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 07:49:09 crc kubenswrapper[4612]: I0227 07:49:09.932171 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 07:49:09 crc kubenswrapper[4612]: I0227 07:49:09.932269 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 07:49:09 crc kubenswrapper[4612]: I0227 07:49:09.932329 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 07:49:10 crc kubenswrapper[4612]: I0227 07:49:10.215781 4612 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 27 07:49:10 crc kubenswrapper[4612]: I0227 07:49:10.216025 4612 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 07:49:10 crc kubenswrapper[4612]: I0227 07:49:10.217519 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 07:49:10 crc kubenswrapper[4612]: I0227 07:49:10.217584 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 07:49:10 crc kubenswrapper[4612]: I0227 07:49:10.217599 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 07:49:11 crc kubenswrapper[4612]: I0227 07:49:11.701322 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Feb 27 07:49:11 crc kubenswrapper[4612]: I0227 07:49:11.701559 4612 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 07:49:11 crc kubenswrapper[4612]: I0227 07:49:11.703166 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 07:49:11 crc kubenswrapper[4612]: I0227 07:49:11.703212 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 07:49:11 crc kubenswrapper[4612]: I0227 07:49:11.703234 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 07:49:11 crc kubenswrapper[4612]: I0227 07:49:11.748475 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 27 07:49:11 crc kubenswrapper[4612]: I0227 07:49:11.748773 4612 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 07:49:11 crc kubenswrapper[4612]: I0227 07:49:11.750352 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 07:49:11 crc kubenswrapper[4612]: I0227 07:49:11.750409 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 07:49:11 crc kubenswrapper[4612]: I0227 07:49:11.750430 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 07:49:11 crc kubenswrapper[4612]: I0227 07:49:11.773516 4612 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 27 07:49:11 crc kubenswrapper[4612]: I0227 07:49:11.773678 4612 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 07:49:11 crc kubenswrapper[4612]: I0227 07:49:11.775039 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 07:49:11 crc kubenswrapper[4612]: I0227 07:49:11.775083 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 07:49:11 crc kubenswrapper[4612]: I0227 07:49:11.775102 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 07:49:11 crc kubenswrapper[4612]: I0227 07:49:11.782322 4612 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 27 07:49:11 crc kubenswrapper[4612]: I0227 07:49:11.937605 4612 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 07:49:11 crc kubenswrapper[4612]: I0227 07:49:11.937766 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 27 07:49:11 crc kubenswrapper[4612]: I0227 07:49:11.939454 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 07:49:11 crc kubenswrapper[4612]: I0227 07:49:11.939517 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 07:49:11 crc kubenswrapper[4612]: I0227 07:49:11.939541 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 07:49:12 crc kubenswrapper[4612]: I0227 07:49:12.062128 4612 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 27 07:49:12 crc kubenswrapper[4612]: E0227 07:49:12.921991 4612 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Feb 27 07:49:12 crc kubenswrapper[4612]: I0227 07:49:12.940507 4612 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 07:49:12 crc kubenswrapper[4612]: I0227 07:49:12.941811 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 07:49:12 crc kubenswrapper[4612]: I0227 07:49:12.941872 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 07:49:12 crc kubenswrapper[4612]: I0227 07:49:12.941890 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 07:49:13 crc kubenswrapper[4612]: I0227 07:49:13.942613 4612 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 07:49:13 crc kubenswrapper[4612]: I0227 07:49:13.943849 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 07:49:13 crc kubenswrapper[4612]: I0227 07:49:13.943883 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 07:49:13 crc kubenswrapper[4612]: I0227 07:49:13.943895 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 07:49:15 crc kubenswrapper[4612]: I0227 07:49:15.062236 4612 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Feb 27 07:49:15 crc kubenswrapper[4612]: I0227 07:49:15.062334 4612 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Feb 27 07:49:15 crc kubenswrapper[4612]: I0227 07:49:15.265823 4612 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Feb 27 07:49:15 crc kubenswrapper[4612]: I0227 07:49:15.266085 4612 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 07:49:15 crc kubenswrapper[4612]: I0227 07:49:15.267595 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 07:49:15 crc kubenswrapper[4612]: I0227 07:49:15.267654 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 07:49:15 crc kubenswrapper[4612]: I0227 07:49:15.267675 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 07:49:16 crc kubenswrapper[4612]: I0227 07:49:16.775100 4612 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Feb 27 07:49:16 crc kubenswrapper[4612]: W0227 07:49:16.784895 4612 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": net/http: TLS handshake timeout Feb 27 07:49:16 crc kubenswrapper[4612]: I0227 07:49:16.785036 4612 trace.go:236] Trace[1751911800]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (27-Feb-2026 07:49:06.783) (total time: 10001ms): Feb 27 07:49:16 crc kubenswrapper[4612]: Trace[1751911800]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (07:49:16.784) Feb 27 07:49:16 crc kubenswrapper[4612]: Trace[1751911800]: [10.001874779s] [10.001874779s] END Feb 27 07:49:16 crc kubenswrapper[4612]: E0227 07:49:16.785135 4612 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Feb 27 07:49:16 crc kubenswrapper[4612]: W0227 07:49:16.885762 4612 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": net/http: TLS handshake timeout Feb 27 07:49:16 crc kubenswrapper[4612]: I0227 07:49:16.885969 4612 trace.go:236] Trace[1407229729]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (27-Feb-2026 07:49:06.884) (total time: 10001ms): Feb 27 07:49:16 crc kubenswrapper[4612]: Trace[1407229729]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (07:49:16.885) Feb 27 07:49:16 crc kubenswrapper[4612]: Trace[1407229729]: [10.001867129s] [10.001867129s] END Feb 27 07:49:16 crc kubenswrapper[4612]: E0227 07:49:16.886020 4612 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Feb 27 07:49:16 crc kubenswrapper[4612]: I0227 07:49:16.954411 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Feb 27 07:49:16 crc kubenswrapper[4612]: I0227 07:49:16.957664 4612 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="28f728bc67c04c9dedfd4ec5ddc8fe577f3f5f59c9acaf1a1c68477b10a34f41" exitCode=255 Feb 27 07:49:16 crc kubenswrapper[4612]: I0227 07:49:16.957750 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"28f728bc67c04c9dedfd4ec5ddc8fe577f3f5f59c9acaf1a1c68477b10a34f41"} Feb 27 07:49:16 crc kubenswrapper[4612]: I0227 07:49:16.958049 4612 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 07:49:16 crc kubenswrapper[4612]: I0227 07:49:16.959350 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 07:49:16 crc kubenswrapper[4612]: I0227 07:49:16.959423 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 07:49:16 crc kubenswrapper[4612]: I0227 07:49:16.959454 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 07:49:16 crc kubenswrapper[4612]: I0227 07:49:16.960534 4612 scope.go:117] "RemoveContainer" containerID="28f728bc67c04c9dedfd4ec5ddc8fe577f3f5f59c9acaf1a1c68477b10a34f41" Feb 27 07:49:17 crc kubenswrapper[4612]: I0227 07:49:17.964180 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Feb 27 07:49:17 crc kubenswrapper[4612]: I0227 07:49:17.971880 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"8cf2ba693080e1b29495be0244dbb644aad0b6d082341df3fd4ce54c741c5b23"} Feb 27 07:49:17 crc kubenswrapper[4612]: I0227 07:49:17.972119 4612 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 07:49:17 crc kubenswrapper[4612]: I0227 07:49:17.973268 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 07:49:17 crc kubenswrapper[4612]: I0227 07:49:17.973332 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 07:49:17 crc kubenswrapper[4612]: I0227 07:49:17.973353 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 07:49:18 crc kubenswrapper[4612]: E0227 07:49:18.207820 4612 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T07:49:18Z is after 2026-02-23T05:33:13Z" node="crc" Feb 27 07:49:18 crc kubenswrapper[4612]: W0227 07:49:18.214003 4612 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T07:49:18Z is after 2026-02-23T05:33:13Z Feb 27 07:49:18 crc kubenswrapper[4612]: E0227 07:49:18.214104 4612 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T07:49:18Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Feb 27 07:49:18 crc kubenswrapper[4612]: W0227 07:49:18.215484 4612 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T07:49:18Z is after 2026-02-23T05:33:13Z Feb 27 07:49:18 crc kubenswrapper[4612]: E0227 07:49:18.215556 4612 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T07:49:18Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Feb 27 07:49:18 crc kubenswrapper[4612]: I0227 07:49:18.217271 4612 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Feb 27 07:49:18 crc kubenswrapper[4612]: I0227 07:49:18.217313 4612 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Feb 27 07:49:18 crc kubenswrapper[4612]: I0227 07:49:18.219470 4612 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T07:49:18Z is after 2026-02-23T05:33:13Z Feb 27 07:49:18 crc kubenswrapper[4612]: E0227 07:49:18.222225 4612 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T07:49:18Z is after 2026-02-23T05:33:13Z" event="&Event{ObjectMeta:{crc.18980afec884cc0b default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 07:49:02.772407307 +0000 UTC m=+0.626337345,LastTimestamp:2026-02-27 07:49:02.772407307 +0000 UTC m=+0.626337345,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 07:49:18 crc kubenswrapper[4612]: I0227 07:49:18.222826 4612 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Feb 27 07:49:18 crc kubenswrapper[4612]: I0227 07:49:18.222924 4612 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Feb 27 07:49:18 crc kubenswrapper[4612]: E0227 07:49:18.223858 4612 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T07:49:18Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Feb 27 07:49:18 crc kubenswrapper[4612]: E0227 07:49:18.226001 4612 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T07:49:18Z is after 2026-02-23T05:33:13Z" interval="6.4s" Feb 27 07:49:18 crc kubenswrapper[4612]: I0227 07:49:18.776540 4612 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T07:49:18Z is after 2026-02-23T05:33:13Z Feb 27 07:49:18 crc kubenswrapper[4612]: I0227 07:49:18.977938 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Feb 27 07:49:18 crc kubenswrapper[4612]: I0227 07:49:18.980020 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Feb 27 07:49:18 crc kubenswrapper[4612]: I0227 07:49:18.982854 4612 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="8cf2ba693080e1b29495be0244dbb644aad0b6d082341df3fd4ce54c741c5b23" exitCode=255 Feb 27 07:49:18 crc kubenswrapper[4612]: I0227 07:49:18.982986 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"8cf2ba693080e1b29495be0244dbb644aad0b6d082341df3fd4ce54c741c5b23"} Feb 27 07:49:18 crc kubenswrapper[4612]: I0227 07:49:18.983310 4612 scope.go:117] "RemoveContainer" containerID="28f728bc67c04c9dedfd4ec5ddc8fe577f3f5f59c9acaf1a1c68477b10a34f41" Feb 27 07:49:18 crc kubenswrapper[4612]: I0227 07:49:18.983443 4612 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 07:49:18 crc kubenswrapper[4612]: I0227 07:49:18.984763 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 07:49:18 crc kubenswrapper[4612]: I0227 07:49:18.984962 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 07:49:18 crc kubenswrapper[4612]: I0227 07:49:18.985116 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 07:49:18 crc kubenswrapper[4612]: I0227 07:49:18.986794 4612 scope.go:117] "RemoveContainer" containerID="8cf2ba693080e1b29495be0244dbb644aad0b6d082341df3fd4ce54c741c5b23" Feb 27 07:49:18 crc kubenswrapper[4612]: E0227 07:49:18.987224 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Feb 27 07:49:19 crc kubenswrapper[4612]: I0227 07:49:19.271150 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 27 07:49:19 crc kubenswrapper[4612]: I0227 07:49:19.779449 4612 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T07:49:19Z is after 2026-02-23T05:33:13Z Feb 27 07:49:19 crc kubenswrapper[4612]: I0227 07:49:19.989517 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Feb 27 07:49:19 crc kubenswrapper[4612]: I0227 07:49:19.993948 4612 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 07:49:19 crc kubenswrapper[4612]: I0227 07:49:19.995411 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 07:49:19 crc kubenswrapper[4612]: I0227 07:49:19.995485 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 07:49:19 crc kubenswrapper[4612]: I0227 07:49:19.995514 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 07:49:19 crc kubenswrapper[4612]: I0227 07:49:19.996591 4612 scope.go:117] "RemoveContainer" containerID="8cf2ba693080e1b29495be0244dbb644aad0b6d082341df3fd4ce54c741c5b23" Feb 27 07:49:19 crc kubenswrapper[4612]: E0227 07:49:19.996981 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Feb 27 07:49:20 crc kubenswrapper[4612]: I0227 07:49:20.030794 4612 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 27 07:49:20 crc kubenswrapper[4612]: I0227 07:49:20.227042 4612 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 27 07:49:20 crc kubenswrapper[4612]: W0227 07:49:20.300424 4612 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T07:49:20Z is after 2026-02-23T05:33:13Z Feb 27 07:49:20 crc kubenswrapper[4612]: E0227 07:49:20.300541 4612 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T07:49:20Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Feb 27 07:49:20 crc kubenswrapper[4612]: I0227 07:49:20.780240 4612 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T07:49:20Z is after 2026-02-23T05:33:13Z Feb 27 07:49:20 crc kubenswrapper[4612]: I0227 07:49:20.996519 4612 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 07:49:20 crc kubenswrapper[4612]: I0227 07:49:20.997906 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 07:49:20 crc kubenswrapper[4612]: I0227 07:49:20.997966 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 07:49:20 crc kubenswrapper[4612]: I0227 07:49:20.998030 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 07:49:20 crc kubenswrapper[4612]: I0227 07:49:20.998921 4612 scope.go:117] "RemoveContainer" containerID="8cf2ba693080e1b29495be0244dbb644aad0b6d082341df3fd4ce54c741c5b23" Feb 27 07:49:20 crc kubenswrapper[4612]: E0227 07:49:20.999290 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Feb 27 07:49:21 crc kubenswrapper[4612]: I0227 07:49:21.003822 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 27 07:49:21 crc kubenswrapper[4612]: I0227 07:49:21.779979 4612 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T07:49:21Z is after 2026-02-23T05:33:13Z Feb 27 07:49:22 crc kubenswrapper[4612]: I0227 07:49:22.000349 4612 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 07:49:22 crc kubenswrapper[4612]: I0227 07:49:22.002342 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 07:49:22 crc kubenswrapper[4612]: I0227 07:49:22.002379 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 07:49:22 crc kubenswrapper[4612]: I0227 07:49:22.002391 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 07:49:22 crc kubenswrapper[4612]: I0227 07:49:22.002937 4612 scope.go:117] "RemoveContainer" containerID="8cf2ba693080e1b29495be0244dbb644aad0b6d082341df3fd4ce54c741c5b23" Feb 27 07:49:22 crc kubenswrapper[4612]: E0227 07:49:22.003155 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Feb 27 07:49:22 crc kubenswrapper[4612]: I0227 07:49:22.779075 4612 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T07:49:22Z is after 2026-02-23T05:33:13Z Feb 27 07:49:22 crc kubenswrapper[4612]: W0227 07:49:22.829863 4612 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T07:49:22Z is after 2026-02-23T05:33:13Z Feb 27 07:49:22 crc kubenswrapper[4612]: E0227 07:49:22.830292 4612 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T07:49:22Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Feb 27 07:49:22 crc kubenswrapper[4612]: E0227 07:49:22.922222 4612 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Feb 27 07:49:23 crc kubenswrapper[4612]: I0227 07:49:23.001614 4612 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 07:49:23 crc kubenswrapper[4612]: I0227 07:49:23.003324 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 07:49:23 crc kubenswrapper[4612]: I0227 07:49:23.003400 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 07:49:23 crc kubenswrapper[4612]: I0227 07:49:23.003422 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 07:49:23 crc kubenswrapper[4612]: I0227 07:49:23.004403 4612 scope.go:117] "RemoveContainer" containerID="8cf2ba693080e1b29495be0244dbb644aad0b6d082341df3fd4ce54c741c5b23" Feb 27 07:49:23 crc kubenswrapper[4612]: E0227 07:49:23.004751 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Feb 27 07:49:23 crc kubenswrapper[4612]: I0227 07:49:23.707802 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 27 07:49:23 crc kubenswrapper[4612]: I0227 07:49:23.708024 4612 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 07:49:23 crc kubenswrapper[4612]: I0227 07:49:23.709628 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 07:49:23 crc kubenswrapper[4612]: I0227 07:49:23.709724 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 07:49:23 crc kubenswrapper[4612]: I0227 07:49:23.709744 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 07:49:23 crc kubenswrapper[4612]: I0227 07:49:23.779003 4612 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T07:49:23Z is after 2026-02-23T05:33:13Z Feb 27 07:49:24 crc kubenswrapper[4612]: I0227 07:49:24.608052 4612 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 07:49:24 crc kubenswrapper[4612]: I0227 07:49:24.610014 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 07:49:24 crc kubenswrapper[4612]: I0227 07:49:24.610060 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 07:49:24 crc kubenswrapper[4612]: I0227 07:49:24.610081 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 07:49:24 crc kubenswrapper[4612]: I0227 07:49:24.610120 4612 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 27 07:49:24 crc kubenswrapper[4612]: E0227 07:49:24.615510 4612 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T07:49:24Z is after 2026-02-23T05:33:13Z" node="crc" Feb 27 07:49:24 crc kubenswrapper[4612]: E0227 07:49:24.631817 4612 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T07:49:24Z is after 2026-02-23T05:33:13Z" interval="7s" Feb 27 07:49:24 crc kubenswrapper[4612]: I0227 07:49:24.777727 4612 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T07:49:24Z is after 2026-02-23T05:33:13Z Feb 27 07:49:25 crc kubenswrapper[4612]: I0227 07:49:25.063202 4612 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Feb 27 07:49:25 crc kubenswrapper[4612]: I0227 07:49:25.063891 4612 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Feb 27 07:49:25 crc kubenswrapper[4612]: I0227 07:49:25.311585 4612 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Feb 27 07:49:25 crc kubenswrapper[4612]: I0227 07:49:25.311788 4612 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 07:49:25 crc kubenswrapper[4612]: I0227 07:49:25.313313 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 07:49:25 crc kubenswrapper[4612]: I0227 07:49:25.313387 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 07:49:25 crc kubenswrapper[4612]: I0227 07:49:25.313431 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 07:49:25 crc kubenswrapper[4612]: I0227 07:49:25.332911 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Feb 27 07:49:25 crc kubenswrapper[4612]: I0227 07:49:25.777994 4612 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T07:49:25Z is after 2026-02-23T05:33:13Z Feb 27 07:49:26 crc kubenswrapper[4612]: I0227 07:49:26.010016 4612 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 07:49:26 crc kubenswrapper[4612]: I0227 07:49:26.011645 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 07:49:26 crc kubenswrapper[4612]: I0227 07:49:26.011737 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 07:49:26 crc kubenswrapper[4612]: I0227 07:49:26.011756 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 07:49:26 crc kubenswrapper[4612]: I0227 07:49:26.491469 4612 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Feb 27 07:49:26 crc kubenswrapper[4612]: E0227 07:49:26.495328 4612 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T07:49:26Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Feb 27 07:49:26 crc kubenswrapper[4612]: W0227 07:49:26.690787 4612 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T07:49:26Z is after 2026-02-23T05:33:13Z Feb 27 07:49:26 crc kubenswrapper[4612]: E0227 07:49:26.690906 4612 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T07:49:26Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Feb 27 07:49:26 crc kubenswrapper[4612]: I0227 07:49:26.778825 4612 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T07:49:26Z is after 2026-02-23T05:33:13Z Feb 27 07:49:27 crc kubenswrapper[4612]: I0227 07:49:27.777835 4612 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T07:49:27Z is after 2026-02-23T05:33:13Z Feb 27 07:49:28 crc kubenswrapper[4612]: E0227 07:49:28.228013 4612 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T07:49:28Z is after 2026-02-23T05:33:13Z" event="&Event{ObjectMeta:{crc.18980afec884cc0b default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 07:49:02.772407307 +0000 UTC m=+0.626337345,LastTimestamp:2026-02-27 07:49:02.772407307 +0000 UTC m=+0.626337345,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 07:49:28 crc kubenswrapper[4612]: I0227 07:49:28.777680 4612 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T07:49:28Z is after 2026-02-23T05:33:13Z Feb 27 07:49:29 crc kubenswrapper[4612]: I0227 07:49:29.778456 4612 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T07:49:29Z is after 2026-02-23T05:33:13Z Feb 27 07:49:30 crc kubenswrapper[4612]: W0227 07:49:30.334174 4612 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T07:49:30Z is after 2026-02-23T05:33:13Z Feb 27 07:49:30 crc kubenswrapper[4612]: E0227 07:49:30.334263 4612 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T07:49:30Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Feb 27 07:49:30 crc kubenswrapper[4612]: I0227 07:49:30.778508 4612 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T07:49:30Z is after 2026-02-23T05:33:13Z Feb 27 07:49:31 crc kubenswrapper[4612]: W0227 07:49:31.377035 4612 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T07:49:31Z is after 2026-02-23T05:33:13Z Feb 27 07:49:31 crc kubenswrapper[4612]: E0227 07:49:31.377169 4612 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T07:49:31Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Feb 27 07:49:31 crc kubenswrapper[4612]: I0227 07:49:31.616344 4612 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 07:49:31 crc kubenswrapper[4612]: I0227 07:49:31.617993 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 07:49:31 crc kubenswrapper[4612]: I0227 07:49:31.618059 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 07:49:31 crc kubenswrapper[4612]: I0227 07:49:31.618073 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 07:49:31 crc kubenswrapper[4612]: I0227 07:49:31.618111 4612 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 27 07:49:31 crc kubenswrapper[4612]: E0227 07:49:31.623073 4612 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T07:49:31Z is after 2026-02-23T05:33:13Z" node="crc" Feb 27 07:49:31 crc kubenswrapper[4612]: E0227 07:49:31.637200 4612 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T07:49:31Z is after 2026-02-23T05:33:13Z" interval="7s" Feb 27 07:49:31 crc kubenswrapper[4612]: I0227 07:49:31.779223 4612 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T07:49:31Z is after 2026-02-23T05:33:13Z Feb 27 07:49:32 crc kubenswrapper[4612]: W0227 07:49:32.349358 4612 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T07:49:32Z is after 2026-02-23T05:33:13Z Feb 27 07:49:32 crc kubenswrapper[4612]: E0227 07:49:32.349940 4612 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T07:49:32Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Feb 27 07:49:32 crc kubenswrapper[4612]: I0227 07:49:32.778807 4612 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T07:49:32Z is after 2026-02-23T05:33:13Z Feb 27 07:49:32 crc kubenswrapper[4612]: E0227 07:49:32.922343 4612 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Feb 27 07:49:33 crc kubenswrapper[4612]: I0227 07:49:33.778517 4612 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T07:49:33Z is after 2026-02-23T05:33:13Z Feb 27 07:49:34 crc kubenswrapper[4612]: I0227 07:49:34.776806 4612 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T07:49:34Z is after 2026-02-23T05:33:13Z Feb 27 07:49:34 crc kubenswrapper[4612]: I0227 07:49:34.852674 4612 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 07:49:34 crc kubenswrapper[4612]: I0227 07:49:34.854007 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 07:49:34 crc kubenswrapper[4612]: I0227 07:49:34.854051 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 07:49:34 crc kubenswrapper[4612]: I0227 07:49:34.854065 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 07:49:34 crc kubenswrapper[4612]: I0227 07:49:34.854766 4612 scope.go:117] "RemoveContainer" containerID="8cf2ba693080e1b29495be0244dbb644aad0b6d082341df3fd4ce54c741c5b23" Feb 27 07:49:35 crc kubenswrapper[4612]: I0227 07:49:35.063465 4612 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Feb 27 07:49:35 crc kubenswrapper[4612]: I0227 07:49:35.063562 4612 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Feb 27 07:49:35 crc kubenswrapper[4612]: I0227 07:49:35.063641 4612 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 27 07:49:35 crc kubenswrapper[4612]: I0227 07:49:35.063889 4612 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 07:49:35 crc kubenswrapper[4612]: I0227 07:49:35.065540 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 07:49:35 crc kubenswrapper[4612]: I0227 07:49:35.065577 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 07:49:35 crc kubenswrapper[4612]: I0227 07:49:35.065589 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 07:49:35 crc kubenswrapper[4612]: I0227 07:49:35.066390 4612 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="cluster-policy-controller" containerStatusID={"Type":"cri-o","ID":"0a55081c1229c10dee891581d85f34e62567e1fd7f76a9c96d601a2e30bec08f"} pod="openshift-kube-controller-manager/kube-controller-manager-crc" containerMessage="Container cluster-policy-controller failed startup probe, will be restarted" Feb 27 07:49:35 crc kubenswrapper[4612]: I0227 07:49:35.066601 4612 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" containerID="cri-o://0a55081c1229c10dee891581d85f34e62567e1fd7f76a9c96d601a2e30bec08f" gracePeriod=30 Feb 27 07:49:35 crc kubenswrapper[4612]: I0227 07:49:35.776456 4612 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T07:49:35Z is after 2026-02-23T05:33:13Z Feb 27 07:49:36 crc kubenswrapper[4612]: I0227 07:49:36.042038 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/cluster-policy-controller/0.log" Feb 27 07:49:36 crc kubenswrapper[4612]: I0227 07:49:36.042453 4612 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="0a55081c1229c10dee891581d85f34e62567e1fd7f76a9c96d601a2e30bec08f" exitCode=255 Feb 27 07:49:36 crc kubenswrapper[4612]: I0227 07:49:36.042526 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"0a55081c1229c10dee891581d85f34e62567e1fd7f76a9c96d601a2e30bec08f"} Feb 27 07:49:36 crc kubenswrapper[4612]: I0227 07:49:36.042561 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"016abce8809b28fae5bb861d181cf6c53cda44abc6a3637502f791365646ba06"} Feb 27 07:49:36 crc kubenswrapper[4612]: I0227 07:49:36.042667 4612 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 07:49:36 crc kubenswrapper[4612]: I0227 07:49:36.043785 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 07:49:36 crc kubenswrapper[4612]: I0227 07:49:36.043852 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 07:49:36 crc kubenswrapper[4612]: I0227 07:49:36.043866 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 07:49:36 crc kubenswrapper[4612]: I0227 07:49:36.046224 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/2.log" Feb 27 07:49:36 crc kubenswrapper[4612]: I0227 07:49:36.046740 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Feb 27 07:49:36 crc kubenswrapper[4612]: I0227 07:49:36.048974 4612 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="a2d539ca19556575737da94d7c2b8242335c3d7df5c3508645937c43f5d56ce5" exitCode=255 Feb 27 07:49:36 crc kubenswrapper[4612]: I0227 07:49:36.049021 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"a2d539ca19556575737da94d7c2b8242335c3d7df5c3508645937c43f5d56ce5"} Feb 27 07:49:36 crc kubenswrapper[4612]: I0227 07:49:36.049065 4612 scope.go:117] "RemoveContainer" containerID="8cf2ba693080e1b29495be0244dbb644aad0b6d082341df3fd4ce54c741c5b23" Feb 27 07:49:36 crc kubenswrapper[4612]: I0227 07:49:36.049466 4612 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 07:49:36 crc kubenswrapper[4612]: I0227 07:49:36.050544 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 07:49:36 crc kubenswrapper[4612]: I0227 07:49:36.050584 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 07:49:36 crc kubenswrapper[4612]: I0227 07:49:36.050595 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 07:49:36 crc kubenswrapper[4612]: I0227 07:49:36.051228 4612 scope.go:117] "RemoveContainer" containerID="a2d539ca19556575737da94d7c2b8242335c3d7df5c3508645937c43f5d56ce5" Feb 27 07:49:36 crc kubenswrapper[4612]: E0227 07:49:36.051412 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Feb 27 07:49:36 crc kubenswrapper[4612]: I0227 07:49:36.778128 4612 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T07:49:36Z is after 2026-02-23T05:33:13Z Feb 27 07:49:37 crc kubenswrapper[4612]: I0227 07:49:37.055274 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/2.log" Feb 27 07:49:37 crc kubenswrapper[4612]: I0227 07:49:37.577373 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 27 07:49:37 crc kubenswrapper[4612]: I0227 07:49:37.578110 4612 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 07:49:37 crc kubenswrapper[4612]: I0227 07:49:37.579476 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 07:49:37 crc kubenswrapper[4612]: I0227 07:49:37.579642 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 07:49:37 crc kubenswrapper[4612]: I0227 07:49:37.579770 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 07:49:37 crc kubenswrapper[4612]: I0227 07:49:37.778596 4612 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T07:49:37Z is after 2026-02-23T05:33:13Z Feb 27 07:49:38 crc kubenswrapper[4612]: E0227 07:49:38.234195 4612 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T07:49:38Z is after 2026-02-23T05:33:13Z" event="&Event{ObjectMeta:{crc.18980afec884cc0b default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 07:49:02.772407307 +0000 UTC m=+0.626337345,LastTimestamp:2026-02-27 07:49:02.772407307 +0000 UTC m=+0.626337345,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 07:49:38 crc kubenswrapper[4612]: I0227 07:49:38.623955 4612 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 07:49:38 crc kubenswrapper[4612]: I0227 07:49:38.625943 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 07:49:38 crc kubenswrapper[4612]: I0227 07:49:38.626018 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 07:49:38 crc kubenswrapper[4612]: I0227 07:49:38.626044 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 07:49:38 crc kubenswrapper[4612]: I0227 07:49:38.626081 4612 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 27 07:49:38 crc kubenswrapper[4612]: E0227 07:49:38.630952 4612 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T07:49:38Z is after 2026-02-23T05:33:13Z" node="crc" Feb 27 07:49:38 crc kubenswrapper[4612]: E0227 07:49:38.642554 4612 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T07:49:38Z is after 2026-02-23T05:33:13Z" interval="7s" Feb 27 07:49:38 crc kubenswrapper[4612]: I0227 07:49:38.776669 4612 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T07:49:38Z is after 2026-02-23T05:33:13Z Feb 27 07:49:39 crc kubenswrapper[4612]: I0227 07:49:39.270497 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 27 07:49:39 crc kubenswrapper[4612]: I0227 07:49:39.271437 4612 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 07:49:39 crc kubenswrapper[4612]: I0227 07:49:39.272997 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 07:49:39 crc kubenswrapper[4612]: I0227 07:49:39.273070 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 07:49:39 crc kubenswrapper[4612]: I0227 07:49:39.273152 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 07:49:39 crc kubenswrapper[4612]: I0227 07:49:39.274119 4612 scope.go:117] "RemoveContainer" containerID="a2d539ca19556575737da94d7c2b8242335c3d7df5c3508645937c43f5d56ce5" Feb 27 07:49:39 crc kubenswrapper[4612]: E0227 07:49:39.274415 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Feb 27 07:49:39 crc kubenswrapper[4612]: I0227 07:49:39.778953 4612 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T07:49:39Z is after 2026-02-23T05:33:13Z Feb 27 07:49:40 crc kubenswrapper[4612]: I0227 07:49:40.030219 4612 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 27 07:49:40 crc kubenswrapper[4612]: I0227 07:49:40.067344 4612 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 07:49:40 crc kubenswrapper[4612]: I0227 07:49:40.068855 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 07:49:40 crc kubenswrapper[4612]: I0227 07:49:40.069073 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 07:49:40 crc kubenswrapper[4612]: I0227 07:49:40.069226 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 07:49:40 crc kubenswrapper[4612]: I0227 07:49:40.070203 4612 scope.go:117] "RemoveContainer" containerID="a2d539ca19556575737da94d7c2b8242335c3d7df5c3508645937c43f5d56ce5" Feb 27 07:49:40 crc kubenswrapper[4612]: E0227 07:49:40.070673 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Feb 27 07:49:40 crc kubenswrapper[4612]: I0227 07:49:40.777207 4612 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T07:49:40Z is after 2026-02-23T05:33:13Z Feb 27 07:49:41 crc kubenswrapper[4612]: I0227 07:49:41.778621 4612 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T07:49:41Z is after 2026-02-23T05:33:13Z Feb 27 07:49:42 crc kubenswrapper[4612]: I0227 07:49:42.062343 4612 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 27 07:49:42 crc kubenswrapper[4612]: I0227 07:49:42.062498 4612 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 07:49:42 crc kubenswrapper[4612]: I0227 07:49:42.063597 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 07:49:42 crc kubenswrapper[4612]: I0227 07:49:42.063633 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 07:49:42 crc kubenswrapper[4612]: I0227 07:49:42.063643 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 07:49:42 crc kubenswrapper[4612]: I0227 07:49:42.753608 4612 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Feb 27 07:49:42 crc kubenswrapper[4612]: E0227 07:49:42.757454 4612 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T07:49:42Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Feb 27 07:49:42 crc kubenswrapper[4612]: E0227 07:49:42.758648 4612 certificate_manager.go:440] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Reached backoff limit, still unable to rotate certs: timed out waiting for the condition" logger="UnhandledError" Feb 27 07:49:42 crc kubenswrapper[4612]: I0227 07:49:42.775663 4612 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T07:49:42Z is after 2026-02-23T05:33:13Z Feb 27 07:49:42 crc kubenswrapper[4612]: E0227 07:49:42.922664 4612 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Feb 27 07:49:43 crc kubenswrapper[4612]: I0227 07:49:43.778432 4612 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T07:49:43Z is after 2026-02-23T05:33:13Z Feb 27 07:49:44 crc kubenswrapper[4612]: I0227 07:49:44.779056 4612 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T07:49:44Z is after 2026-02-23T05:33:13Z Feb 27 07:49:45 crc kubenswrapper[4612]: I0227 07:49:45.063079 4612 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Feb 27 07:49:45 crc kubenswrapper[4612]: I0227 07:49:45.063640 4612 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Feb 27 07:49:45 crc kubenswrapper[4612]: I0227 07:49:45.631382 4612 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 07:49:45 crc kubenswrapper[4612]: I0227 07:49:45.633047 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 07:49:45 crc kubenswrapper[4612]: I0227 07:49:45.633219 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 07:49:45 crc kubenswrapper[4612]: I0227 07:49:45.633323 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 07:49:45 crc kubenswrapper[4612]: I0227 07:49:45.633479 4612 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 27 07:49:45 crc kubenswrapper[4612]: E0227 07:49:45.637096 4612 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T07:49:45Z is after 2026-02-23T05:33:13Z" node="crc" Feb 27 07:49:45 crc kubenswrapper[4612]: E0227 07:49:45.646931 4612 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T07:49:45Z is after 2026-02-23T05:33:13Z" interval="7s" Feb 27 07:49:45 crc kubenswrapper[4612]: I0227 07:49:45.778391 4612 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T07:49:45Z is after 2026-02-23T05:33:13Z Feb 27 07:49:46 crc kubenswrapper[4612]: I0227 07:49:46.777387 4612 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T07:49:46Z is after 2026-02-23T05:33:13Z Feb 27 07:49:47 crc kubenswrapper[4612]: I0227 07:49:47.779645 4612 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T07:49:47Z is after 2026-02-23T05:33:13Z Feb 27 07:49:48 crc kubenswrapper[4612]: E0227 07:49:48.239959 4612 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T07:49:48Z is after 2026-02-23T05:33:13Z" event="&Event{ObjectMeta:{crc.18980afec884cc0b default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 07:49:02.772407307 +0000 UTC m=+0.626337345,LastTimestamp:2026-02-27 07:49:02.772407307 +0000 UTC m=+0.626337345,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 07:49:48 crc kubenswrapper[4612]: I0227 07:49:48.777145 4612 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T07:49:48Z is after 2026-02-23T05:33:13Z Feb 27 07:49:49 crc kubenswrapper[4612]: W0227 07:49:49.575824 4612 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T07:49:49Z is after 2026-02-23T05:33:13Z Feb 27 07:49:49 crc kubenswrapper[4612]: E0227 07:49:49.575912 4612 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T07:49:49Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Feb 27 07:49:49 crc kubenswrapper[4612]: I0227 07:49:49.779039 4612 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 27 07:49:49 crc kubenswrapper[4612]: W0227 07:49:49.887008 4612 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: csidrivers.storage.k8s.io is forbidden: User "system:anonymous" cannot list resource "csidrivers" in API group "storage.k8s.io" at the cluster scope Feb 27 07:49:49 crc kubenswrapper[4612]: E0227 07:49:49.887072 4612 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: csidrivers.storage.k8s.io is forbidden: User \"system:anonymous\" cannot list resource \"csidrivers\" in API group \"storage.k8s.io\" at the cluster scope" logger="UnhandledError" Feb 27 07:49:50 crc kubenswrapper[4612]: I0227 07:49:50.781093 4612 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 27 07:49:51 crc kubenswrapper[4612]: I0227 07:49:51.779810 4612 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 27 07:49:51 crc kubenswrapper[4612]: I0227 07:49:51.852470 4612 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 07:49:51 crc kubenswrapper[4612]: I0227 07:49:51.854214 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 07:49:51 crc kubenswrapper[4612]: I0227 07:49:51.854311 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 07:49:51 crc kubenswrapper[4612]: I0227 07:49:51.854328 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 07:49:51 crc kubenswrapper[4612]: I0227 07:49:51.855337 4612 scope.go:117] "RemoveContainer" containerID="a2d539ca19556575737da94d7c2b8242335c3d7df5c3508645937c43f5d56ce5" Feb 27 07:49:51 crc kubenswrapper[4612]: E0227 07:49:51.855625 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Feb 27 07:49:52 crc kubenswrapper[4612]: I0227 07:49:52.637492 4612 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 07:49:52 crc kubenswrapper[4612]: I0227 07:49:52.638981 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 07:49:52 crc kubenswrapper[4612]: I0227 07:49:52.639136 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 07:49:52 crc kubenswrapper[4612]: I0227 07:49:52.639239 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 07:49:52 crc kubenswrapper[4612]: I0227 07:49:52.639351 4612 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 27 07:49:52 crc kubenswrapper[4612]: E0227 07:49:52.644419 4612 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes is forbidden: User \"system:anonymous\" cannot create resource \"nodes\" in API group \"\" at the cluster scope" node="crc" Feb 27 07:49:52 crc kubenswrapper[4612]: E0227 07:49:52.651216 4612 controller.go:145] "Failed to ensure lease exists, will retry" err="leases.coordination.k8s.io \"crc\" is forbidden: User \"system:anonymous\" cannot get resource \"leases\" in API group \"coordination.k8s.io\" in the namespace \"kube-node-lease\"" interval="7s" Feb 27 07:49:52 crc kubenswrapper[4612]: I0227 07:49:52.781545 4612 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 27 07:49:52 crc kubenswrapper[4612]: E0227 07:49:52.923738 4612 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Feb 27 07:49:53 crc kubenswrapper[4612]: I0227 07:49:53.777141 4612 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 27 07:49:54 crc kubenswrapper[4612]: W0227 07:49:54.430511 4612 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: nodes "crc" is forbidden: User "system:anonymous" cannot list resource "nodes" in API group "" at the cluster scope Feb 27 07:49:54 crc kubenswrapper[4612]: E0227 07:49:54.430577 4612 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: nodes \"crc\" is forbidden: User \"system:anonymous\" cannot list resource \"nodes\" in API group \"\" at the cluster scope" logger="UnhandledError" Feb 27 07:49:54 crc kubenswrapper[4612]: I0227 07:49:54.780235 4612 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 27 07:49:55 crc kubenswrapper[4612]: I0227 07:49:55.063264 4612 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Feb 27 07:49:55 crc kubenswrapper[4612]: I0227 07:49:55.063766 4612 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Feb 27 07:49:55 crc kubenswrapper[4612]: I0227 07:49:55.776623 4612 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 27 07:49:56 crc kubenswrapper[4612]: I0227 07:49:56.527197 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 27 07:49:56 crc kubenswrapper[4612]: I0227 07:49:56.527753 4612 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 07:49:56 crc kubenswrapper[4612]: I0227 07:49:56.529602 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 07:49:56 crc kubenswrapper[4612]: I0227 07:49:56.529844 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 07:49:56 crc kubenswrapper[4612]: I0227 07:49:56.530005 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 07:49:56 crc kubenswrapper[4612]: I0227 07:49:56.780797 4612 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 27 07:49:57 crc kubenswrapper[4612]: W0227 07:49:57.164667 4612 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: runtimeclasses.node.k8s.io is forbidden: User "system:anonymous" cannot list resource "runtimeclasses" in API group "node.k8s.io" at the cluster scope Feb 27 07:49:57 crc kubenswrapper[4612]: E0227 07:49:57.164749 4612 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: runtimeclasses.node.k8s.io is forbidden: User \"system:anonymous\" cannot list resource \"runtimeclasses\" in API group \"node.k8s.io\" at the cluster scope" logger="UnhandledError" Feb 27 07:49:57 crc kubenswrapper[4612]: I0227 07:49:57.777028 4612 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 27 07:49:58 crc kubenswrapper[4612]: E0227 07:49:58.249230 4612 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.18980afec884cc0b default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 07:49:02.772407307 +0000 UTC m=+0.626337345,LastTimestamp:2026-02-27 07:49:02.772407307 +0000 UTC m=+0.626337345,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 07:49:58 crc kubenswrapper[4612]: E0227 07:49:58.257248 4612 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.18980afecca496c3 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 07:49:02.841599683 +0000 UTC m=+0.695529691,LastTimestamp:2026-02-27 07:49:02.841599683 +0000 UTC m=+0.695529691,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 07:49:58 crc kubenswrapper[4612]: E0227 07:49:58.264983 4612 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.18980afecca4ee1b default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 07:49:02.841622043 +0000 UTC m=+0.695552051,LastTimestamp:2026-02-27 07:49:02.841622043 +0000 UTC m=+0.695552051,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 07:49:58 crc kubenswrapper[4612]: E0227 07:49:58.270833 4612 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.18980afecca51ba8 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 07:49:02.841633704 +0000 UTC m=+0.695563702,LastTimestamp:2026-02-27 07:49:02.841633704 +0000 UTC m=+0.695563702,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 07:49:58 crc kubenswrapper[4612]: E0227 07:49:58.277019 4612 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.18980afed189f9ff default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeAllocatableEnforced,Message:Updated Node Allocatable limit across pods,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 07:49:02.923741695 +0000 UTC m=+0.777671703,LastTimestamp:2026-02-27 07:49:02.923741695 +0000 UTC m=+0.777671703,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 07:49:58 crc kubenswrapper[4612]: E0227 07:49:58.283420 4612 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.18980afecca496c3\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.18980afecca496c3 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 07:49:02.841599683 +0000 UTC m=+0.695529691,LastTimestamp:2026-02-27 07:49:02.954133524 +0000 UTC m=+0.808063512,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 07:49:58 crc kubenswrapper[4612]: E0227 07:49:58.290789 4612 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.18980afecca4ee1b\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.18980afecca4ee1b default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 07:49:02.841622043 +0000 UTC m=+0.695552051,LastTimestamp:2026-02-27 07:49:02.954151334 +0000 UTC m=+0.808081332,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 07:49:58 crc kubenswrapper[4612]: E0227 07:49:58.297287 4612 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.18980afecca51ba8\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.18980afecca51ba8 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 07:49:02.841633704 +0000 UTC m=+0.695563702,LastTimestamp:2026-02-27 07:49:02.954162595 +0000 UTC m=+0.808092593,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 07:49:58 crc kubenswrapper[4612]: E0227 07:49:58.303157 4612 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.18980afecca496c3\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.18980afecca496c3 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 07:49:02.841599683 +0000 UTC m=+0.695529691,LastTimestamp:2026-02-27 07:49:02.955213744 +0000 UTC m=+0.809143742,Count:3,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 07:49:58 crc kubenswrapper[4612]: E0227 07:49:58.308315 4612 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.18980afecca4ee1b\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.18980afecca4ee1b default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 07:49:02.841622043 +0000 UTC m=+0.695552051,LastTimestamp:2026-02-27 07:49:02.955224725 +0000 UTC m=+0.809154723,Count:3,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 07:49:58 crc kubenswrapper[4612]: E0227 07:49:58.314496 4612 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.18980afecca51ba8\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.18980afecca51ba8 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 07:49:02.841633704 +0000 UTC m=+0.695563702,LastTimestamp:2026-02-27 07:49:02.955231865 +0000 UTC m=+0.809161863,Count:3,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 07:49:58 crc kubenswrapper[4612]: E0227 07:49:58.321028 4612 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.18980afecca496c3\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.18980afecca496c3 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 07:49:02.841599683 +0000 UTC m=+0.695529691,LastTimestamp:2026-02-27 07:49:02.957129228 +0000 UTC m=+0.811059246,Count:4,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 07:49:58 crc kubenswrapper[4612]: E0227 07:49:58.328098 4612 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.18980afecca4ee1b\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.18980afecca4ee1b default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 07:49:02.841622043 +0000 UTC m=+0.695552051,LastTimestamp:2026-02-27 07:49:02.957152489 +0000 UTC m=+0.811082487,Count:4,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 07:49:58 crc kubenswrapper[4612]: E0227 07:49:58.333049 4612 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.18980afecca51ba8\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.18980afecca51ba8 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 07:49:02.841633704 +0000 UTC m=+0.695563702,LastTimestamp:2026-02-27 07:49:02.957161789 +0000 UTC m=+0.811091787,Count:4,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 07:49:58 crc kubenswrapper[4612]: E0227 07:49:58.338312 4612 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.18980afecca496c3\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.18980afecca496c3 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 07:49:02.841599683 +0000 UTC m=+0.695529691,LastTimestamp:2026-02-27 07:49:02.95719951 +0000 UTC m=+0.811129548,Count:5,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 07:49:58 crc kubenswrapper[4612]: E0227 07:49:58.343585 4612 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.18980afecca4ee1b\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.18980afecca4ee1b default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 07:49:02.841622043 +0000 UTC m=+0.695552051,LastTimestamp:2026-02-27 07:49:02.957252832 +0000 UTC m=+0.811182870,Count:5,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 07:49:58 crc kubenswrapper[4612]: E0227 07:49:58.345057 4612 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.18980afecca51ba8\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.18980afecca51ba8 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 07:49:02.841633704 +0000 UTC m=+0.695563702,LastTimestamp:2026-02-27 07:49:02.957277903 +0000 UTC m=+0.811207931,Count:5,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 07:49:58 crc kubenswrapper[4612]: E0227 07:49:58.353453 4612 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.18980afecca496c3\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.18980afecca496c3 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 07:49:02.841599683 +0000 UTC m=+0.695529691,LastTimestamp:2026-02-27 07:49:02.958380504 +0000 UTC m=+0.812310502,Count:6,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 07:49:58 crc kubenswrapper[4612]: E0227 07:49:58.359155 4612 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.18980afecca4ee1b\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.18980afecca4ee1b default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 07:49:02.841622043 +0000 UTC m=+0.695552051,LastTimestamp:2026-02-27 07:49:02.958395734 +0000 UTC m=+0.812325732,Count:6,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 07:49:58 crc kubenswrapper[4612]: E0227 07:49:58.364393 4612 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.18980afecca51ba8\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.18980afecca51ba8 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 07:49:02.841633704 +0000 UTC m=+0.695563702,LastTimestamp:2026-02-27 07:49:02.958405344 +0000 UTC m=+0.812335342,Count:6,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 07:49:58 crc kubenswrapper[4612]: E0227 07:49:58.369648 4612 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.18980afecca496c3\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.18980afecca496c3 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 07:49:02.841599683 +0000 UTC m=+0.695529691,LastTimestamp:2026-02-27 07:49:02.959221227 +0000 UTC m=+0.813151215,Count:7,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 07:49:58 crc kubenswrapper[4612]: E0227 07:49:58.375785 4612 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.18980afecca4ee1b\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.18980afecca4ee1b default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 07:49:02.841622043 +0000 UTC m=+0.695552051,LastTimestamp:2026-02-27 07:49:02.959232768 +0000 UTC m=+0.813162756,Count:7,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 07:49:58 crc kubenswrapper[4612]: E0227 07:49:58.380303 4612 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.18980afecca51ba8\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.18980afecca51ba8 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 07:49:02.841633704 +0000 UTC m=+0.695563702,LastTimestamp:2026-02-27 07:49:02.959240848 +0000 UTC m=+0.813170836,Count:7,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 07:49:58 crc kubenswrapper[4612]: E0227 07:49:58.385662 4612 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.18980afecca496c3\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.18980afecca496c3 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 07:49:02.841599683 +0000 UTC m=+0.695529691,LastTimestamp:2026-02-27 07:49:02.95931285 +0000 UTC m=+0.813242848,Count:8,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 07:49:58 crc kubenswrapper[4612]: E0227 07:49:58.390899 4612 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.18980afecca4ee1b\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.18980afecca4ee1b default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 07:49:02.841622043 +0000 UTC m=+0.695552051,LastTimestamp:2026-02-27 07:49:02.95932716 +0000 UTC m=+0.813257158,Count:8,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 07:49:58 crc kubenswrapper[4612]: E0227 07:49:58.398161 4612 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.18980afeebb24e67 openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 07:49:03.362592359 +0000 UTC m=+1.216522377,LastTimestamp:2026-02-27 07:49:03.362592359 +0000 UTC m=+1.216522377,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 07:49:58 crc kubenswrapper[4612]: E0227 07:49:58.403585 4612 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.18980afeebbb5964 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 07:49:03.363184996 +0000 UTC m=+1.217115034,LastTimestamp:2026-02-27 07:49:03.363184996 +0000 UTC m=+1.217115034,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 07:49:58 crc kubenswrapper[4612]: E0227 07:49:58.410605 4612 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.18980afeec3936f7 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 07:49:03.371433719 +0000 UTC m=+1.225363757,LastTimestamp:2026-02-27 07:49:03.371433719 +0000 UTC m=+1.225363757,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 07:49:58 crc kubenswrapper[4612]: E0227 07:49:58.417967 4612 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.18980afeecb3b744 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 07:49:03.379461956 +0000 UTC m=+1.233391974,LastTimestamp:2026-02-27 07:49:03.379461956 +0000 UTC m=+1.233391974,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 07:49:58 crc kubenswrapper[4612]: E0227 07:49:58.425073 4612 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.18980afeed2c6792 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{wait-for-host-port},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 07:49:03.38737141 +0000 UTC m=+1.241301438,LastTimestamp:2026-02-27 07:49:03.38737141 +0000 UTC m=+1.241301438,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 07:49:58 crc kubenswrapper[4612]: E0227 07:49:58.430123 4612 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.18980aff124b63e1 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{wait-for-host-port},},Reason:Created,Message:Created container wait-for-host-port,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 07:49:04.010159073 +0000 UTC m=+1.864089081,LastTimestamp:2026-02-27 07:49:04.010159073 +0000 UTC m=+1.864089081,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 07:49:58 crc kubenswrapper[4612]: E0227 07:49:58.436114 4612 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.18980aff12ea17bf openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Created,Message:Created container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 07:49:04.020559807 +0000 UTC m=+1.874489815,LastTimestamp:2026-02-27 07:49:04.020559807 +0000 UTC m=+1.874489815,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 07:49:58 crc kubenswrapper[4612]: E0227 07:49:58.440682 4612 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.18980aff13251ecd openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{wait-for-host-port},},Reason:Started,Message:Started container wait-for-host-port,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 07:49:04.024428237 +0000 UTC m=+1.878358245,LastTimestamp:2026-02-27 07:49:04.024428237 +0000 UTC m=+1.878358245,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 07:49:58 crc kubenswrapper[4612]: E0227 07:49:58.449208 4612 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.18980aff1332d662 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager},},Reason:Created,Message:Created container kube-controller-manager,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 07:49:04.025327202 +0000 UTC m=+1.879257210,LastTimestamp:2026-02-27 07:49:04.025327202 +0000 UTC m=+1.879257210,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 07:49:58 crc kubenswrapper[4612]: E0227 07:49:58.453327 4612 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.18980aff13416ebd openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Created,Message:Created container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 07:49:04.026283709 +0000 UTC m=+1.880213717,LastTimestamp:2026-02-27 07:49:04.026283709 +0000 UTC m=+1.880213717,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 07:49:58 crc kubenswrapper[4612]: E0227 07:49:58.457799 4612 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.18980aff13ab3957 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Created,Message:Created container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 07:49:04.033216855 +0000 UTC m=+1.887146863,LastTimestamp:2026-02-27 07:49:04.033216855 +0000 UTC m=+1.887146863,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 07:49:58 crc kubenswrapper[4612]: E0227 07:49:58.461861 4612 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.18980aff13f9f6f1 openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Started,Message:Started container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 07:49:04.038377201 +0000 UTC m=+1.892307209,LastTimestamp:2026-02-27 07:49:04.038377201 +0000 UTC m=+1.892307209,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 07:49:58 crc kubenswrapper[4612]: E0227 07:49:58.466674 4612 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.18980aff1402fef6 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager},},Reason:Started,Message:Started container kube-controller-manager,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 07:49:04.038969078 +0000 UTC m=+1.892899086,LastTimestamp:2026-02-27 07:49:04.038969078 +0000 UTC m=+1.892899086,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 07:49:58 crc kubenswrapper[4612]: E0227 07:49:58.471516 4612 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.18980aff1418a31e openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 07:49:04.040387358 +0000 UTC m=+1.894317366,LastTimestamp:2026-02-27 07:49:04.040387358 +0000 UTC m=+1.894317366,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 07:49:58 crc kubenswrapper[4612]: E0227 07:49:58.476719 4612 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.18980aff144a1de1 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Started,Message:Started container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 07:49:04.043630049 +0000 UTC m=+1.897560087,LastTimestamp:2026-02-27 07:49:04.043630049 +0000 UTC m=+1.897560087,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 07:49:58 crc kubenswrapper[4612]: E0227 07:49:58.480851 4612 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.18980aff151a6231 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Started,Message:Started container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 07:49:04.057279025 +0000 UTC m=+1.911209043,LastTimestamp:2026-02-27 07:49:04.057279025 +0000 UTC m=+1.911209043,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 07:49:58 crc kubenswrapper[4612]: E0227 07:49:58.484861 4612 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.18980aff2396a829 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Created,Message:Created container cluster-policy-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 07:49:04.300304425 +0000 UTC m=+2.154234453,LastTimestamp:2026-02-27 07:49:04.300304425 +0000 UTC m=+2.154234453,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 07:49:58 crc kubenswrapper[4612]: E0227 07:49:58.488399 4612 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.18980aff246d93e1 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Started,Message:Started container cluster-policy-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 07:49:04.314389473 +0000 UTC m=+2.168319511,LastTimestamp:2026-02-27 07:49:04.314389473 +0000 UTC m=+2.168319511,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 07:49:58 crc kubenswrapper[4612]: E0227 07:49:58.492919 4612 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.18980aff248555f7 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-cert-syncer},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 07:49:04.315946487 +0000 UTC m=+2.169876515,LastTimestamp:2026-02-27 07:49:04.315946487 +0000 UTC m=+2.169876515,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 07:49:58 crc kubenswrapper[4612]: E0227 07:49:58.497780 4612 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.18980aff3380a302 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-cert-syncer},},Reason:Created,Message:Created container kube-controller-manager-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 07:49:04.56729677 +0000 UTC m=+2.421226778,LastTimestamp:2026-02-27 07:49:04.56729677 +0000 UTC m=+2.421226778,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 07:49:58 crc kubenswrapper[4612]: E0227 07:49:58.504857 4612 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.18980aff345b12f3 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-cert-syncer},},Reason:Started,Message:Started container kube-controller-manager-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 07:49:04.581612275 +0000 UTC m=+2.435542313,LastTimestamp:2026-02-27 07:49:04.581612275 +0000 UTC m=+2.435542313,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 07:49:58 crc kubenswrapper[4612]: E0227 07:49:58.510161 4612 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.18980aff3472ad77 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-recovery-controller},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 07:49:04.583159159 +0000 UTC m=+2.437089157,LastTimestamp:2026-02-27 07:49:04.583159159 +0000 UTC m=+2.437089157,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 07:49:58 crc kubenswrapper[4612]: E0227 07:49:58.516576 4612 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.18980aff415f0b81 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-recovery-controller},},Reason:Created,Message:Created container kube-controller-manager-recovery-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 07:49:04.799976321 +0000 UTC m=+2.653906319,LastTimestamp:2026-02-27 07:49:04.799976321 +0000 UTC m=+2.653906319,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 07:49:58 crc kubenswrapper[4612]: E0227 07:49:58.521000 4612 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.18980aff42130ebf openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-recovery-controller},},Reason:Started,Message:Started container kube-controller-manager-recovery-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 07:49:04.811773631 +0000 UTC m=+2.665703629,LastTimestamp:2026-02-27 07:49:04.811773631 +0000 UTC m=+2.665703629,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 07:49:58 crc kubenswrapper[4612]: E0227 07:49:58.527662 4612 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.18980aff45c9c641 openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-rbac-proxy-crio},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 07:49:04.874079809 +0000 UTC m=+2.728009807,LastTimestamp:2026-02-27 07:49:04.874079809 +0000 UTC m=+2.728009807,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 07:49:58 crc kubenswrapper[4612]: E0227 07:49:58.533754 4612 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.18980aff45ec822c openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-ensure-env-vars},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 07:49:04.87635614 +0000 UTC m=+2.730286138,LastTimestamp:2026-02-27 07:49:04.87635614 +0000 UTC m=+2.730286138,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 07:49:58 crc kubenswrapper[4612]: E0227 07:49:58.538980 4612 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.18980aff45f377ab openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 07:49:04.876812203 +0000 UTC m=+2.730742201,LastTimestamp:2026-02-27 07:49:04.876812203 +0000 UTC m=+2.730742201,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 07:49:58 crc kubenswrapper[4612]: E0227 07:49:58.545941 4612 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.18980aff46b54e93 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 07:49:04.889515667 +0000 UTC m=+2.743445675,LastTimestamp:2026-02-27 07:49:04.889515667 +0000 UTC m=+2.743445675,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 07:49:58 crc kubenswrapper[4612]: E0227 07:49:58.551350 4612 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.18980aff53bac522 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-ensure-env-vars},},Reason:Created,Message:Created container etcd-ensure-env-vars,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 07:49:05.107977506 +0000 UTC m=+2.961907504,LastTimestamp:2026-02-27 07:49:05.107977506 +0000 UTC m=+2.961907504,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 07:49:58 crc kubenswrapper[4612]: E0227 07:49:58.552684 4612 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.18980aff541e3c71 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler},},Reason:Created,Message:Created container kube-scheduler,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 07:49:05.114496113 +0000 UTC m=+2.968426111,LastTimestamp:2026-02-27 07:49:05.114496113 +0000 UTC m=+2.968426111,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 07:49:58 crc kubenswrapper[4612]: E0227 07:49:58.557381 4612 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.18980aff5434d3ff openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:Created,Message:Created container kube-apiserver,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 07:49:05.115976703 +0000 UTC m=+2.969906711,LastTimestamp:2026-02-27 07:49:05.115976703 +0000 UTC m=+2.969906711,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 07:49:58 crc kubenswrapper[4612]: E0227 07:49:58.562027 4612 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.18980aff5480545f openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-rbac-proxy-crio},},Reason:Created,Message:Created container kube-rbac-proxy-crio,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 07:49:05.120924767 +0000 UTC m=+2.974854765,LastTimestamp:2026-02-27 07:49:05.120924767 +0000 UTC m=+2.974854765,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 07:49:58 crc kubenswrapper[4612]: E0227 07:49:58.566390 4612 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.18980aff54864abd openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-ensure-env-vars},},Reason:Started,Message:Started container etcd-ensure-env-vars,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 07:49:05.121315517 +0000 UTC m=+2.975245505,LastTimestamp:2026-02-27 07:49:05.121315517 +0000 UTC m=+2.975245505,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 07:49:58 crc kubenswrapper[4612]: E0227 07:49:58.572979 4612 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.18980aff54ddc7fd openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler},},Reason:Started,Message:Started container kube-scheduler,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 07:49:05.127049213 +0000 UTC m=+2.980979211,LastTimestamp:2026-02-27 07:49:05.127049213 +0000 UTC m=+2.980979211,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 07:49:58 crc kubenswrapper[4612]: E0227 07:49:58.577472 4612 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.18980aff54eed38b openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-cert-syncer},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 07:49:05.128166283 +0000 UTC m=+2.982096281,LastTimestamp:2026-02-27 07:49:05.128166283 +0000 UTC m=+2.982096281,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 07:49:58 crc kubenswrapper[4612]: E0227 07:49:58.580899 4612 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.18980aff5540bade openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:Started,Message:Started container kube-apiserver,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 07:49:05.133533918 +0000 UTC m=+2.987463916,LastTimestamp:2026-02-27 07:49:05.133533918 +0000 UTC m=+2.987463916,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 07:49:58 crc kubenswrapper[4612]: E0227 07:49:58.583988 4612 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.18980aff55a214f3 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-syncer},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 07:49:05.139913971 +0000 UTC m=+2.993843969,LastTimestamp:2026-02-27 07:49:05.139913971 +0000 UTC m=+2.993843969,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 07:49:58 crc kubenswrapper[4612]: E0227 07:49:58.587850 4612 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.18980aff5614866c openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-rbac-proxy-crio},},Reason:Started,Message:Started container kube-rbac-proxy-crio,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 07:49:05.147414124 +0000 UTC m=+3.001344122,LastTimestamp:2026-02-27 07:49:05.147414124 +0000 UTC m=+3.001344122,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 07:49:58 crc kubenswrapper[4612]: E0227 07:49:58.592195 4612 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.18980aff60837f93 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-cert-syncer},},Reason:Created,Message:Created container kube-scheduler-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 07:49:05.322459027 +0000 UTC m=+3.176389025,LastTimestamp:2026-02-27 07:49:05.322459027 +0000 UTC m=+3.176389025,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 07:49:58 crc kubenswrapper[4612]: E0227 07:49:58.598417 4612 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.18980aff608be750 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-syncer},},Reason:Created,Message:Created container kube-apiserver-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 07:49:05.323009872 +0000 UTC m=+3.176939870,LastTimestamp:2026-02-27 07:49:05.323009872 +0000 UTC m=+3.176939870,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 07:49:58 crc kubenswrapper[4612]: E0227 07:49:58.605134 4612 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.18980aff620d9ab5 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-cert-syncer},},Reason:Started,Message:Started container kube-scheduler-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 07:49:05.348287157 +0000 UTC m=+3.202217155,LastTimestamp:2026-02-27 07:49:05.348287157 +0000 UTC m=+3.202217155,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 07:49:58 crc kubenswrapper[4612]: E0227 07:49:58.608856 4612 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.18980aff621f63cc openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-syncer},},Reason:Started,Message:Started container kube-apiserver-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 07:49:05.349452748 +0000 UTC m=+3.203382766,LastTimestamp:2026-02-27 07:49:05.349452748 +0000 UTC m=+3.203382766,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 07:49:58 crc kubenswrapper[4612]: E0227 07:49:58.614005 4612 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.18980aff621f9dc5 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-recovery-controller},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 07:49:05.349467589 +0000 UTC m=+3.203397587,LastTimestamp:2026-02-27 07:49:05.349467589 +0000 UTC m=+3.203397587,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 07:49:58 crc kubenswrapper[4612]: E0227 07:49:58.619617 4612 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.18980aff6230e7f2 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-regeneration-controller},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 07:49:05.35060069 +0000 UTC m=+3.204530718,LastTimestamp:2026-02-27 07:49:05.35060069 +0000 UTC m=+3.204530718,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 07:49:58 crc kubenswrapper[4612]: E0227 07:49:58.625928 4612 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.18980aff6d3844f9 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-recovery-controller},},Reason:Created,Message:Created container kube-scheduler-recovery-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 07:49:05.535632633 +0000 UTC m=+3.389562631,LastTimestamp:2026-02-27 07:49:05.535632633 +0000 UTC m=+3.389562631,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 07:49:58 crc kubenswrapper[4612]: E0227 07:49:58.631992 4612 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.18980aff6d472a1f openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-regeneration-controller},},Reason:Created,Message:Created container kube-apiserver-cert-regeneration-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 07:49:05.536608799 +0000 UTC m=+3.390538797,LastTimestamp:2026-02-27 07:49:05.536608799 +0000 UTC m=+3.390538797,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 07:49:58 crc kubenswrapper[4612]: E0227 07:49:58.638936 4612 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.18980aff6e31d300 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-recovery-controller},},Reason:Started,Message:Started container kube-scheduler-recovery-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 07:49:05.551987456 +0000 UTC m=+3.405917454,LastTimestamp:2026-02-27 07:49:05.551987456 +0000 UTC m=+3.405917454,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 07:49:58 crc kubenswrapper[4612]: E0227 07:49:58.645527 4612 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.18980aff6e94ae3f openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-regeneration-controller},},Reason:Started,Message:Started container kube-apiserver-cert-regeneration-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 07:49:05.558466111 +0000 UTC m=+3.412396109,LastTimestamp:2026-02-27 07:49:05.558466111 +0000 UTC m=+3.412396109,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 07:49:58 crc kubenswrapper[4612]: E0227 07:49:58.649277 4612 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.18980aff6ea6f6c4 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-insecure-readyz},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 07:49:05.559664324 +0000 UTC m=+3.413594322,LastTimestamp:2026-02-27 07:49:05.559664324 +0000 UTC m=+3.413594322,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 07:49:58 crc kubenswrapper[4612]: E0227 07:49:58.654920 4612 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.18980aff7903e74f openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-insecure-readyz},},Reason:Created,Message:Created container kube-apiserver-insecure-readyz,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 07:49:05.733527375 +0000 UTC m=+3.587457373,LastTimestamp:2026-02-27 07:49:05.733527375 +0000 UTC m=+3.587457373,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 07:49:58 crc kubenswrapper[4612]: E0227 07:49:58.658617 4612 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.18980aff7a147e54 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-insecure-readyz},},Reason:Started,Message:Started container kube-apiserver-insecure-readyz,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 07:49:05.751391828 +0000 UTC m=+3.605321827,LastTimestamp:2026-02-27 07:49:05.751391828 +0000 UTC m=+3.605321827,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 07:49:58 crc kubenswrapper[4612]: E0227 07:49:58.664684 4612 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.18980aff7a2a29dd openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 07:49:05.752811997 +0000 UTC m=+3.606741995,LastTimestamp:2026-02-27 07:49:05.752811997 +0000 UTC m=+3.606741995,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 07:49:58 crc kubenswrapper[4612]: E0227 07:49:58.669676 4612 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.18980aff82c31841 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-resources-copy},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 07:49:05.897052225 +0000 UTC m=+3.750982223,LastTimestamp:2026-02-27 07:49:05.897052225 +0000 UTC m=+3.750982223,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 07:49:58 crc kubenswrapper[4612]: E0227 07:49:58.677378 4612 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.18980aff84d99cbc openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Created,Message:Created container kube-apiserver-check-endpoints,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 07:49:05.932082364 +0000 UTC m=+3.786012382,LastTimestamp:2026-02-27 07:49:05.932082364 +0000 UTC m=+3.786012382,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 07:49:58 crc kubenswrapper[4612]: E0227 07:49:58.681886 4612 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.18980aff85a7b544 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Started,Message:Started container kube-apiserver-check-endpoints,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 07:49:05.94558906 +0000 UTC m=+3.799519068,LastTimestamp:2026-02-27 07:49:05.94558906 +0000 UTC m=+3.799519068,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 07:49:58 crc kubenswrapper[4612]: E0227 07:49:58.687505 4612 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.18980aff8dd48ce8 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-resources-copy},},Reason:Created,Message:Created container etcd-resources-copy,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 07:49:06.082745576 +0000 UTC m=+3.936675574,LastTimestamp:2026-02-27 07:49:06.082745576 +0000 UTC m=+3.936675574,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 07:49:58 crc kubenswrapper[4612]: E0227 07:49:58.694518 4612 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.18980aff8e9de50a openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-resources-copy},},Reason:Started,Message:Started container etcd-resources-copy,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 07:49:06.095940874 +0000 UTC m=+3.949870872,LastTimestamp:2026-02-27 07:49:06.095940874 +0000 UTC m=+3.949870872,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 07:49:58 crc kubenswrapper[4612]: E0227 07:49:58.701350 4612 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.18980affbf789703 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcdctl},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 07:49:06.915579651 +0000 UTC m=+4.769509659,LastTimestamp:2026-02-27 07:49:06.915579651 +0000 UTC m=+4.769509659,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 07:49:58 crc kubenswrapper[4612]: E0227 07:49:58.707670 4612 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.18980affc90e8bb2 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcdctl},},Reason:Created,Message:Created container etcdctl,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 07:49:07.076402098 +0000 UTC m=+4.930332106,LastTimestamp:2026-02-27 07:49:07.076402098 +0000 UTC m=+4.930332106,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 07:49:58 crc kubenswrapper[4612]: E0227 07:49:58.712319 4612 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.18980affca0580dd openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcdctl},},Reason:Started,Message:Started container etcdctl,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 07:49:07.092586717 +0000 UTC m=+4.946516725,LastTimestamp:2026-02-27 07:49:07.092586717 +0000 UTC m=+4.946516725,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 07:49:58 crc kubenswrapper[4612]: E0227 07:49:58.717138 4612 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.18980affca1c3f17 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 07:49:07.094077207 +0000 UTC m=+4.948007225,LastTimestamp:2026-02-27 07:49:07.094077207 +0000 UTC m=+4.948007225,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 07:49:58 crc kubenswrapper[4612]: E0227 07:49:58.721286 4612 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.18980affd684b621 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd},},Reason:Created,Message:Created container etcd,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 07:49:07.302250017 +0000 UTC m=+5.156180025,LastTimestamp:2026-02-27 07:49:07.302250017 +0000 UTC m=+5.156180025,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 07:49:58 crc kubenswrapper[4612]: E0227 07:49:58.726025 4612 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.18980affd76571b0 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd},},Reason:Started,Message:Started container etcd,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 07:49:07.316978096 +0000 UTC m=+5.170908104,LastTimestamp:2026-02-27 07:49:07.316978096 +0000 UTC m=+5.170908104,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 07:49:58 crc kubenswrapper[4612]: E0227 07:49:58.730448 4612 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.18980affd77982a0 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-metrics},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 07:49:07.318293152 +0000 UTC m=+5.172223150,LastTimestamp:2026-02-27 07:49:07.318293152 +0000 UTC m=+5.172223150,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 07:49:58 crc kubenswrapper[4612]: E0227 07:49:58.735578 4612 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.18980affe51a4025 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-metrics},},Reason:Created,Message:Created container etcd-metrics,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 07:49:07.546931237 +0000 UTC m=+5.400861235,LastTimestamp:2026-02-27 07:49:07.546931237 +0000 UTC m=+5.400861235,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 07:49:58 crc kubenswrapper[4612]: E0227 07:49:58.739041 4612 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.18980affe62feee0 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-metrics},},Reason:Started,Message:Started container etcd-metrics,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 07:49:07.56512944 +0000 UTC m=+5.419059428,LastTimestamp:2026-02-27 07:49:07.56512944 +0000 UTC m=+5.419059428,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 07:49:58 crc kubenswrapper[4612]: E0227 07:49:58.742633 4612 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.18980affe649b60a openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-readyz},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 07:49:07.566818826 +0000 UTC m=+5.420748854,LastTimestamp:2026-02-27 07:49:07.566818826 +0000 UTC m=+5.420748854,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 07:49:58 crc kubenswrapper[4612]: E0227 07:49:58.747658 4612 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.18980afff3d3830e openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-readyz},},Reason:Created,Message:Created container etcd-readyz,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 07:49:07.79395355 +0000 UTC m=+5.647883558,LastTimestamp:2026-02-27 07:49:07.79395355 +0000 UTC m=+5.647883558,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 07:49:58 crc kubenswrapper[4612]: E0227 07:49:58.752234 4612 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.18980afff4c26720 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-readyz},},Reason:Started,Message:Started container etcd-readyz,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 07:49:07.809609504 +0000 UTC m=+5.663539522,LastTimestamp:2026-02-27 07:49:07.809609504 +0000 UTC m=+5.663539522,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 07:49:58 crc kubenswrapper[4612]: E0227 07:49:58.756896 4612 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.18980afff4d3e357 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-rev},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 07:49:07.810755415 +0000 UTC m=+5.664685423,LastTimestamp:2026-02-27 07:49:07.810755415 +0000 UTC m=+5.664685423,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 07:49:58 crc kubenswrapper[4612]: E0227 07:49:58.758913 4612 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.18980b0001b3cd22 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-rev},},Reason:Created,Message:Created container etcd-rev,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 07:49:08.026756386 +0000 UTC m=+5.880686424,LastTimestamp:2026-02-27 07:49:08.026756386 +0000 UTC m=+5.880686424,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 07:49:58 crc kubenswrapper[4612]: E0227 07:49:58.762904 4612 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.18980b0002b43460 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-rev},},Reason:Started,Message:Started container etcd-rev,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 07:49:08.043560032 +0000 UTC m=+5.897490040,LastTimestamp:2026-02-27 07:49:08.043560032 +0000 UTC m=+5.897490040,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 07:49:58 crc kubenswrapper[4612]: E0227 07:49:58.769877 4612 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event=< Feb 27 07:49:58 crc kubenswrapper[4612]: &Event{ObjectMeta:{kube-controller-manager-crc.18980b01a50dbdee openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:ProbeError,Message:Startup probe error: Get "https://192.168.126.11:10357/healthz": context deadline exceeded (Client.Timeout exceeded while awaiting headers) Feb 27 07:49:58 crc kubenswrapper[4612]: body: Feb 27 07:49:58 crc kubenswrapper[4612]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 07:49:15.062304238 +0000 UTC m=+12.916234246,LastTimestamp:2026-02-27 07:49:15.062304238 +0000 UTC m=+12.916234246,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Feb 27 07:49:58 crc kubenswrapper[4612]: > Feb 27 07:49:58 crc kubenswrapper[4612]: I0227 07:49:58.775303 4612 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 27 07:49:58 crc kubenswrapper[4612]: E0227 07:49:58.775172 4612 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.18980b01a50ed2e3 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Unhealthy,Message:Startup probe failed: Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers),Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 07:49:15.062375139 +0000 UTC m=+12.916305157,LastTimestamp:2026-02-27 07:49:15.062375139 +0000 UTC m=+12.916305157,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 07:49:58 crc kubenswrapper[4612]: E0227 07:49:58.781098 4612 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-apiserver-crc.18980aff7a2a29dd\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.18980aff7a2a29dd openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 07:49:05.752811997 +0000 UTC m=+3.606741995,LastTimestamp:2026-02-27 07:49:16.962327846 +0000 UTC m=+14.816257844,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 07:49:58 crc kubenswrapper[4612]: E0227 07:49:58.785893 4612 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-apiserver-crc.18980aff84d99cbc\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.18980aff84d99cbc openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Created,Message:Created container kube-apiserver-check-endpoints,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 07:49:05.932082364 +0000 UTC m=+3.786012382,LastTimestamp:2026-02-27 07:49:17.19278045 +0000 UTC m=+15.046710458,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 07:49:58 crc kubenswrapper[4612]: E0227 07:49:58.791467 4612 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-apiserver-crc.18980aff85a7b544\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.18980aff85a7b544 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Started,Message:Started container kube-apiserver-check-endpoints,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 07:49:05.94558906 +0000 UTC m=+3.799519068,LastTimestamp:2026-02-27 07:49:17.208831225 +0000 UTC m=+15.062761243,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 07:49:58 crc kubenswrapper[4612]: E0227 07:49:58.796679 4612 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event=< Feb 27 07:49:58 crc kubenswrapper[4612]: &Event{ObjectMeta:{kube-apiserver-crc.18980b02611b292c openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:ProbeError,Message:Startup probe error: HTTP probe failed with statuscode: 403 Feb 27 07:49:58 crc kubenswrapper[4612]: body: {"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Feb 27 07:49:58 crc kubenswrapper[4612]: Feb 27 07:49:58 crc kubenswrapper[4612]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 07:49:18.217300268 +0000 UTC m=+16.071230266,LastTimestamp:2026-02-27 07:49:18.217300268 +0000 UTC m=+16.071230266,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Feb 27 07:49:58 crc kubenswrapper[4612]: > Feb 27 07:49:58 crc kubenswrapper[4612]: E0227 07:49:58.802228 4612 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.18980b02611ba335 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:Unhealthy,Message:Startup probe failed: HTTP probe failed with statuscode: 403,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 07:49:18.217331509 +0000 UTC m=+16.071261507,LastTimestamp:2026-02-27 07:49:18.217331509 +0000 UTC m=+16.071261507,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 07:49:58 crc kubenswrapper[4612]: E0227 07:49:58.807222 4612 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-apiserver-crc.18980b02611b292c\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event=< Feb 27 07:49:58 crc kubenswrapper[4612]: &Event{ObjectMeta:{kube-apiserver-crc.18980b02611b292c openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:ProbeError,Message:Startup probe error: HTTP probe failed with statuscode: 403 Feb 27 07:49:58 crc kubenswrapper[4612]: body: {"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Feb 27 07:49:58 crc kubenswrapper[4612]: Feb 27 07:49:58 crc kubenswrapper[4612]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 07:49:18.217300268 +0000 UTC m=+16.071230266,LastTimestamp:2026-02-27 07:49:18.22289228 +0000 UTC m=+16.076822298,Count:2,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Feb 27 07:49:58 crc kubenswrapper[4612]: > Feb 27 07:49:58 crc kubenswrapper[4612]: E0227 07:49:58.811629 4612 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-apiserver-crc.18980b02611ba335\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.18980b02611ba335 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:Unhealthy,Message:Startup probe failed: HTTP probe failed with statuscode: 403,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 07:49:18.217331509 +0000 UTC m=+16.071261507,LastTimestamp:2026-02-27 07:49:18.222962602 +0000 UTC m=+16.076892620,Count:2,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 07:49:58 crc kubenswrapper[4612]: E0227 07:49:58.817781 4612 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event=< Feb 27 07:49:58 crc kubenswrapper[4612]: &Event{ObjectMeta:{kube-controller-manager-crc.18980b03f93120d9 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:ProbeError,Message:Startup probe error: Get "https://192.168.126.11:10357/healthz": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers) Feb 27 07:49:58 crc kubenswrapper[4612]: body: Feb 27 07:49:58 crc kubenswrapper[4612]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 07:49:25.063844057 +0000 UTC m=+22.917774125,LastTimestamp:2026-02-27 07:49:25.063844057 +0000 UTC m=+22.917774125,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Feb 27 07:49:58 crc kubenswrapper[4612]: > Feb 27 07:49:58 crc kubenswrapper[4612]: E0227 07:49:58.822433 4612 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.18980b03f9329fb6 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Unhealthy,Message:Startup probe failed: Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers),Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 07:49:25.06394207 +0000 UTC m=+22.917872108,LastTimestamp:2026-02-27 07:49:25.06394207 +0000 UTC m=+22.917872108,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 07:49:58 crc kubenswrapper[4612]: E0227 07:49:58.826765 4612 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.18980b03f93120d9\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event=< Feb 27 07:49:58 crc kubenswrapper[4612]: &Event{ObjectMeta:{kube-controller-manager-crc.18980b03f93120d9 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:ProbeError,Message:Startup probe error: Get "https://192.168.126.11:10357/healthz": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers) Feb 27 07:49:58 crc kubenswrapper[4612]: body: Feb 27 07:49:58 crc kubenswrapper[4612]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 07:49:25.063844057 +0000 UTC m=+22.917774125,LastTimestamp:2026-02-27 07:49:35.063534895 +0000 UTC m=+32.917464923,Count:2,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Feb 27 07:49:58 crc kubenswrapper[4612]: > Feb 27 07:49:58 crc kubenswrapper[4612]: E0227 07:49:58.837317 4612 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.18980b03f9329fb6\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.18980b03f9329fb6 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Unhealthy,Message:Startup probe failed: Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers),Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 07:49:25.06394207 +0000 UTC m=+22.917872108,LastTimestamp:2026-02-27 07:49:35.063601617 +0000 UTC m=+32.917531655,Count:2,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 07:49:58 crc kubenswrapper[4612]: E0227 07:49:58.842614 4612 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.18980b064d66b8e7 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Killing,Message:Container cluster-policy-controller failed startup probe, will be restarted,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 07:49:35.066577127 +0000 UTC m=+32.920507135,LastTimestamp:2026-02-27 07:49:35.066577127 +0000 UTC m=+32.920507135,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 07:49:58 crc kubenswrapper[4612]: E0227 07:49:58.847346 4612 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.18980aff1418a31e\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.18980aff1418a31e openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 07:49:04.040387358 +0000 UTC m=+1.894317366,LastTimestamp:2026-02-27 07:49:35.192950481 +0000 UTC m=+33.046880519,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 07:49:58 crc kubenswrapper[4612]: E0227 07:49:58.852329 4612 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.18980aff2396a829\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.18980aff2396a829 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Created,Message:Created container cluster-policy-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 07:49:04.300304425 +0000 UTC m=+2.154234453,LastTimestamp:2026-02-27 07:49:35.396014103 +0000 UTC m=+33.249944111,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 07:49:58 crc kubenswrapper[4612]: E0227 07:49:58.857579 4612 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.18980aff246d93e1\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.18980aff246d93e1 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Started,Message:Started container cluster-policy-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 07:49:04.314389473 +0000 UTC m=+2.168319511,LastTimestamp:2026-02-27 07:49:35.404365879 +0000 UTC m=+33.258295887,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 07:49:58 crc kubenswrapper[4612]: E0227 07:49:58.866120 4612 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.18980b03f93120d9\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event=< Feb 27 07:49:58 crc kubenswrapper[4612]: &Event{ObjectMeta:{kube-controller-manager-crc.18980b03f93120d9 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:ProbeError,Message:Startup probe error: Get "https://192.168.126.11:10357/healthz": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers) Feb 27 07:49:58 crc kubenswrapper[4612]: body: Feb 27 07:49:58 crc kubenswrapper[4612]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 07:49:25.063844057 +0000 UTC m=+22.917774125,LastTimestamp:2026-02-27 07:49:45.063578023 +0000 UTC m=+42.917508101,Count:3,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Feb 27 07:49:58 crc kubenswrapper[4612]: > Feb 27 07:49:58 crc kubenswrapper[4612]: E0227 07:49:58.870911 4612 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.18980b03f9329fb6\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.18980b03f9329fb6 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Unhealthy,Message:Startup probe failed: Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers),Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 07:49:25.06394207 +0000 UTC m=+22.917872108,LastTimestamp:2026-02-27 07:49:45.063955003 +0000 UTC m=+42.917885041,Count:3,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 07:49:58 crc kubenswrapper[4612]: E0227 07:49:58.875757 4612 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.18980b03f93120d9\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event=< Feb 27 07:49:58 crc kubenswrapper[4612]: &Event{ObjectMeta:{kube-controller-manager-crc.18980b03f93120d9 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:ProbeError,Message:Startup probe error: Get "https://192.168.126.11:10357/healthz": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers) Feb 27 07:49:58 crc kubenswrapper[4612]: body: Feb 27 07:49:58 crc kubenswrapper[4612]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 07:49:25.063844057 +0000 UTC m=+22.917774125,LastTimestamp:2026-02-27 07:49:55.063739595 +0000 UTC m=+52.917669613,Count:4,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Feb 27 07:49:58 crc kubenswrapper[4612]: > Feb 27 07:49:59 crc kubenswrapper[4612]: I0227 07:49:59.645350 4612 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 07:49:59 crc kubenswrapper[4612]: I0227 07:49:59.647181 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 07:49:59 crc kubenswrapper[4612]: I0227 07:49:59.647235 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 07:49:59 crc kubenswrapper[4612]: I0227 07:49:59.647257 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 07:49:59 crc kubenswrapper[4612]: I0227 07:49:59.647298 4612 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 27 07:49:59 crc kubenswrapper[4612]: E0227 07:49:59.653311 4612 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes is forbidden: User \"system:anonymous\" cannot create resource \"nodes\" in API group \"\" at the cluster scope" node="crc" Feb 27 07:49:59 crc kubenswrapper[4612]: E0227 07:49:59.653389 4612 controller.go:145] "Failed to ensure lease exists, will retry" err="leases.coordination.k8s.io \"crc\" is forbidden: User \"system:anonymous\" cannot get resource \"leases\" in API group \"coordination.k8s.io\" in the namespace \"kube-node-lease\"" interval="7s" Feb 27 07:49:59 crc kubenswrapper[4612]: I0227 07:49:59.778666 4612 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 27 07:50:00 crc kubenswrapper[4612]: I0227 07:50:00.778250 4612 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 27 07:50:01 crc kubenswrapper[4612]: I0227 07:50:01.777809 4612 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 27 07:50:02 crc kubenswrapper[4612]: I0227 07:50:02.066959 4612 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 27 07:50:02 crc kubenswrapper[4612]: I0227 07:50:02.067162 4612 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 07:50:02 crc kubenswrapper[4612]: I0227 07:50:02.068600 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 07:50:02 crc kubenswrapper[4612]: I0227 07:50:02.068736 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 07:50:02 crc kubenswrapper[4612]: I0227 07:50:02.068845 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 07:50:02 crc kubenswrapper[4612]: I0227 07:50:02.075332 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 27 07:50:02 crc kubenswrapper[4612]: I0227 07:50:02.132439 4612 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 07:50:02 crc kubenswrapper[4612]: I0227 07:50:02.134097 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 07:50:02 crc kubenswrapper[4612]: I0227 07:50:02.134160 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 07:50:02 crc kubenswrapper[4612]: I0227 07:50:02.134184 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 07:50:02 crc kubenswrapper[4612]: I0227 07:50:02.779030 4612 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 27 07:50:02 crc kubenswrapper[4612]: E0227 07:50:02.924521 4612 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Feb 27 07:50:03 crc kubenswrapper[4612]: I0227 07:50:03.782118 4612 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 27 07:50:03 crc kubenswrapper[4612]: I0227 07:50:03.853037 4612 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 07:50:03 crc kubenswrapper[4612]: I0227 07:50:03.854927 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 07:50:03 crc kubenswrapper[4612]: I0227 07:50:03.854955 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 07:50:03 crc kubenswrapper[4612]: I0227 07:50:03.854965 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 07:50:03 crc kubenswrapper[4612]: I0227 07:50:03.855579 4612 scope.go:117] "RemoveContainer" containerID="a2d539ca19556575737da94d7c2b8242335c3d7df5c3508645937c43f5d56ce5" Feb 27 07:50:04 crc kubenswrapper[4612]: I0227 07:50:04.139352 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/2.log" Feb 27 07:50:04 crc kubenswrapper[4612]: I0227 07:50:04.141770 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"6fdb61a61c41ceec6aa497297d1853ae6fec429a9dc998e0f49e8f41f41b7e9c"} Feb 27 07:50:04 crc kubenswrapper[4612]: I0227 07:50:04.141932 4612 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 07:50:04 crc kubenswrapper[4612]: I0227 07:50:04.142684 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 07:50:04 crc kubenswrapper[4612]: I0227 07:50:04.142734 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 07:50:04 crc kubenswrapper[4612]: I0227 07:50:04.142746 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 07:50:04 crc kubenswrapper[4612]: I0227 07:50:04.778853 4612 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 27 07:50:05 crc kubenswrapper[4612]: I0227 07:50:05.146011 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/3.log" Feb 27 07:50:05 crc kubenswrapper[4612]: I0227 07:50:05.146481 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/2.log" Feb 27 07:50:05 crc kubenswrapper[4612]: I0227 07:50:05.148410 4612 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="6fdb61a61c41ceec6aa497297d1853ae6fec429a9dc998e0f49e8f41f41b7e9c" exitCode=255 Feb 27 07:50:05 crc kubenswrapper[4612]: I0227 07:50:05.148447 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"6fdb61a61c41ceec6aa497297d1853ae6fec429a9dc998e0f49e8f41f41b7e9c"} Feb 27 07:50:05 crc kubenswrapper[4612]: I0227 07:50:05.148478 4612 scope.go:117] "RemoveContainer" containerID="a2d539ca19556575737da94d7c2b8242335c3d7df5c3508645937c43f5d56ce5" Feb 27 07:50:05 crc kubenswrapper[4612]: I0227 07:50:05.148811 4612 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 07:50:05 crc kubenswrapper[4612]: I0227 07:50:05.149913 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 07:50:05 crc kubenswrapper[4612]: I0227 07:50:05.149942 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 07:50:05 crc kubenswrapper[4612]: I0227 07:50:05.149954 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 07:50:05 crc kubenswrapper[4612]: I0227 07:50:05.150517 4612 scope.go:117] "RemoveContainer" containerID="6fdb61a61c41ceec6aa497297d1853ae6fec429a9dc998e0f49e8f41f41b7e9c" Feb 27 07:50:05 crc kubenswrapper[4612]: E0227 07:50:05.150731 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Feb 27 07:50:05 crc kubenswrapper[4612]: I0227 07:50:05.778651 4612 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 27 07:50:06 crc kubenswrapper[4612]: I0227 07:50:06.156726 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/3.log" Feb 27 07:50:06 crc kubenswrapper[4612]: I0227 07:50:06.653396 4612 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 07:50:06 crc kubenswrapper[4612]: I0227 07:50:06.654732 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 07:50:06 crc kubenswrapper[4612]: I0227 07:50:06.654768 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 07:50:06 crc kubenswrapper[4612]: I0227 07:50:06.654778 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 07:50:06 crc kubenswrapper[4612]: I0227 07:50:06.654809 4612 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 27 07:50:06 crc kubenswrapper[4612]: E0227 07:50:06.657915 4612 controller.go:145] "Failed to ensure lease exists, will retry" err="leases.coordination.k8s.io \"crc\" is forbidden: User \"system:anonymous\" cannot get resource \"leases\" in API group \"coordination.k8s.io\" in the namespace \"kube-node-lease\"" interval="7s" Feb 27 07:50:06 crc kubenswrapper[4612]: E0227 07:50:06.657944 4612 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes is forbidden: User \"system:anonymous\" cannot create resource \"nodes\" in API group \"\" at the cluster scope" node="crc" Feb 27 07:50:06 crc kubenswrapper[4612]: I0227 07:50:06.777797 4612 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 27 07:50:07 crc kubenswrapper[4612]: I0227 07:50:07.777641 4612 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 27 07:50:08 crc kubenswrapper[4612]: I0227 07:50:08.778344 4612 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 27 07:50:09 crc kubenswrapper[4612]: I0227 07:50:09.270218 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 27 07:50:09 crc kubenswrapper[4612]: I0227 07:50:09.270465 4612 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 07:50:09 crc kubenswrapper[4612]: I0227 07:50:09.271766 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 07:50:09 crc kubenswrapper[4612]: I0227 07:50:09.271796 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 07:50:09 crc kubenswrapper[4612]: I0227 07:50:09.271808 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 07:50:09 crc kubenswrapper[4612]: I0227 07:50:09.272257 4612 scope.go:117] "RemoveContainer" containerID="6fdb61a61c41ceec6aa497297d1853ae6fec429a9dc998e0f49e8f41f41b7e9c" Feb 27 07:50:09 crc kubenswrapper[4612]: E0227 07:50:09.272434 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Feb 27 07:50:09 crc kubenswrapper[4612]: I0227 07:50:09.780330 4612 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 27 07:50:10 crc kubenswrapper[4612]: I0227 07:50:10.030196 4612 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 27 07:50:10 crc kubenswrapper[4612]: I0227 07:50:10.169019 4612 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 07:50:10 crc kubenswrapper[4612]: I0227 07:50:10.170243 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 07:50:10 crc kubenswrapper[4612]: I0227 07:50:10.170348 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 07:50:10 crc kubenswrapper[4612]: I0227 07:50:10.170373 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 07:50:10 crc kubenswrapper[4612]: I0227 07:50:10.171459 4612 scope.go:117] "RemoveContainer" containerID="6fdb61a61c41ceec6aa497297d1853ae6fec429a9dc998e0f49e8f41f41b7e9c" Feb 27 07:50:10 crc kubenswrapper[4612]: E0227 07:50:10.171975 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Feb 27 07:50:10 crc kubenswrapper[4612]: I0227 07:50:10.778791 4612 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 27 07:50:11 crc kubenswrapper[4612]: I0227 07:50:11.778941 4612 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 27 07:50:12 crc kubenswrapper[4612]: I0227 07:50:12.779527 4612 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 27 07:50:12 crc kubenswrapper[4612]: E0227 07:50:12.924851 4612 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Feb 27 07:50:13 crc kubenswrapper[4612]: I0227 07:50:13.658931 4612 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 07:50:13 crc kubenswrapper[4612]: I0227 07:50:13.660221 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 07:50:13 crc kubenswrapper[4612]: I0227 07:50:13.660254 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 07:50:13 crc kubenswrapper[4612]: I0227 07:50:13.660282 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 07:50:13 crc kubenswrapper[4612]: I0227 07:50:13.660302 4612 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 27 07:50:13 crc kubenswrapper[4612]: E0227 07:50:13.663104 4612 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes is forbidden: User \"system:anonymous\" cannot create resource \"nodes\" in API group \"\" at the cluster scope" node="crc" Feb 27 07:50:13 crc kubenswrapper[4612]: E0227 07:50:13.664399 4612 controller.go:145] "Failed to ensure lease exists, will retry" err="leases.coordination.k8s.io \"crc\" is forbidden: User \"system:anonymous\" cannot get resource \"leases\" in API group \"coordination.k8s.io\" in the namespace \"kube-node-lease\"" interval="7s" Feb 27 07:50:13 crc kubenswrapper[4612]: I0227 07:50:13.778482 4612 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 27 07:50:14 crc kubenswrapper[4612]: I0227 07:50:14.760903 4612 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Feb 27 07:50:14 crc kubenswrapper[4612]: I0227 07:50:14.776883 4612 reflector.go:368] Caches populated for *v1.CertificateSigningRequest from k8s.io/client-go/tools/watch/informerwatcher.go:146 Feb 27 07:50:14 crc kubenswrapper[4612]: I0227 07:50:14.777812 4612 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 27 07:50:15 crc kubenswrapper[4612]: I0227 07:50:15.779985 4612 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 27 07:50:16 crc kubenswrapper[4612]: I0227 07:50:16.778167 4612 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Feb 27 07:50:17 crc kubenswrapper[4612]: I0227 07:50:17.689871 4612 csr.go:261] certificate signing request csr-zgp6b is approved, waiting to be issued Feb 27 07:50:17 crc kubenswrapper[4612]: I0227 07:50:17.699570 4612 csr.go:257] certificate signing request csr-zgp6b is issued Feb 27 07:50:17 crc kubenswrapper[4612]: I0227 07:50:17.788216 4612 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Feb 27 07:50:18 crc kubenswrapper[4612]: I0227 07:50:18.640527 4612 transport.go:147] "Certificate rotation detected, shutting down client connections to start using new credentials" Feb 27 07:50:18 crc kubenswrapper[4612]: I0227 07:50:18.701747 4612 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2027-02-24 05:54:36 +0000 UTC, rotation deadline is 2026-12-31 08:22:06.299812831 +0000 UTC Feb 27 07:50:18 crc kubenswrapper[4612]: I0227 07:50:18.701805 4612 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 7368h31m47.598012203s for next certificate rotation Feb 27 07:50:20 crc kubenswrapper[4612]: I0227 07:50:20.663679 4612 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 07:50:20 crc kubenswrapper[4612]: I0227 07:50:20.665290 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 07:50:20 crc kubenswrapper[4612]: I0227 07:50:20.665366 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 07:50:20 crc kubenswrapper[4612]: I0227 07:50:20.665389 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 07:50:20 crc kubenswrapper[4612]: I0227 07:50:20.665571 4612 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 27 07:50:20 crc kubenswrapper[4612]: I0227 07:50:20.679983 4612 kubelet_node_status.go:115] "Node was previously registered" node="crc" Feb 27 07:50:20 crc kubenswrapper[4612]: I0227 07:50:20.680273 4612 kubelet_node_status.go:79] "Successfully registered node" node="crc" Feb 27 07:50:20 crc kubenswrapper[4612]: E0227 07:50:20.680308 4612 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": node \"crc\" not found" Feb 27 07:50:20 crc kubenswrapper[4612]: I0227 07:50:20.684580 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 07:50:20 crc kubenswrapper[4612]: I0227 07:50:20.684859 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 07:50:20 crc kubenswrapper[4612]: I0227 07:50:20.685036 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 07:50:20 crc kubenswrapper[4612]: I0227 07:50:20.685221 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 07:50:20 crc kubenswrapper[4612]: I0227 07:50:20.685407 4612 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T07:50:20Z","lastTransitionTime":"2026-02-27T07:50:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 07:50:20 crc kubenswrapper[4612]: E0227 07:50:20.704443 4612 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148060Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608860Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T07:50:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T07:50:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T07:50:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T07:50:20Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T07:50:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T07:50:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T07:50:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T07:50:20Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a8a53c94-4eff-42ba-8a96-6c53b380c5c4\\\",\\\"systemUUID\\\":\\\"6e98ef2c-c447-4ff8-86d3-9b32f302ee4b\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 27 07:50:20 crc kubenswrapper[4612]: I0227 07:50:20.715860 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 07:50:20 crc kubenswrapper[4612]: I0227 07:50:20.715912 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 07:50:20 crc kubenswrapper[4612]: I0227 07:50:20.715932 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 07:50:20 crc kubenswrapper[4612]: I0227 07:50:20.715955 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 07:50:20 crc kubenswrapper[4612]: I0227 07:50:20.715973 4612 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T07:50:20Z","lastTransitionTime":"2026-02-27T07:50:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 07:50:20 crc kubenswrapper[4612]: E0227 07:50:20.732484 4612 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148060Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608860Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T07:50:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T07:50:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T07:50:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T07:50:20Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T07:50:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T07:50:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T07:50:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T07:50:20Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a8a53c94-4eff-42ba-8a96-6c53b380c5c4\\\",\\\"systemUUID\\\":\\\"6e98ef2c-c447-4ff8-86d3-9b32f302ee4b\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 27 07:50:20 crc kubenswrapper[4612]: I0227 07:50:20.743493 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 07:50:20 crc kubenswrapper[4612]: I0227 07:50:20.743541 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 07:50:20 crc kubenswrapper[4612]: I0227 07:50:20.743553 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 07:50:20 crc kubenswrapper[4612]: I0227 07:50:20.743571 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 07:50:20 crc kubenswrapper[4612]: I0227 07:50:20.743586 4612 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T07:50:20Z","lastTransitionTime":"2026-02-27T07:50:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 07:50:20 crc kubenswrapper[4612]: E0227 07:50:20.757992 4612 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148060Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608860Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T07:50:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T07:50:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T07:50:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T07:50:20Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T07:50:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T07:50:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T07:50:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T07:50:20Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a8a53c94-4eff-42ba-8a96-6c53b380c5c4\\\",\\\"systemUUID\\\":\\\"6e98ef2c-c447-4ff8-86d3-9b32f302ee4b\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 27 07:50:20 crc kubenswrapper[4612]: I0227 07:50:20.768425 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 07:50:20 crc kubenswrapper[4612]: I0227 07:50:20.768790 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 07:50:20 crc kubenswrapper[4612]: I0227 07:50:20.769001 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 07:50:20 crc kubenswrapper[4612]: I0227 07:50:20.769409 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 07:50:20 crc kubenswrapper[4612]: I0227 07:50:20.769795 4612 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T07:50:20Z","lastTransitionTime":"2026-02-27T07:50:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 07:50:20 crc kubenswrapper[4612]: E0227 07:50:20.785927 4612 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148060Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608860Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T07:50:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T07:50:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T07:50:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T07:50:20Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T07:50:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T07:50:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T07:50:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T07:50:20Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a8a53c94-4eff-42ba-8a96-6c53b380c5c4\\\",\\\"systemUUID\\\":\\\"6e98ef2c-c447-4ff8-86d3-9b32f302ee4b\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 27 07:50:20 crc kubenswrapper[4612]: E0227 07:50:20.786127 4612 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Feb 27 07:50:20 crc kubenswrapper[4612]: E0227 07:50:20.786159 4612 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 07:50:20 crc kubenswrapper[4612]: E0227 07:50:20.887233 4612 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 07:50:20 crc kubenswrapper[4612]: E0227 07:50:20.987895 4612 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 07:50:21 crc kubenswrapper[4612]: E0227 07:50:21.087988 4612 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 07:50:21 crc kubenswrapper[4612]: E0227 07:50:21.189048 4612 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 07:50:21 crc kubenswrapper[4612]: E0227 07:50:21.289655 4612 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 07:50:21 crc kubenswrapper[4612]: E0227 07:50:21.390101 4612 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 07:50:21 crc kubenswrapper[4612]: E0227 07:50:21.490928 4612 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 07:50:21 crc kubenswrapper[4612]: E0227 07:50:21.592205 4612 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 07:50:21 crc kubenswrapper[4612]: E0227 07:50:21.693515 4612 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 07:50:21 crc kubenswrapper[4612]: E0227 07:50:21.794982 4612 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 07:50:21 crc kubenswrapper[4612]: E0227 07:50:21.896054 4612 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 07:50:21 crc kubenswrapper[4612]: E0227 07:50:21.996642 4612 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 07:50:22 crc kubenswrapper[4612]: E0227 07:50:22.097340 4612 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 07:50:22 crc kubenswrapper[4612]: E0227 07:50:22.198088 4612 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 07:50:22 crc kubenswrapper[4612]: E0227 07:50:22.299007 4612 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 07:50:22 crc kubenswrapper[4612]: E0227 07:50:22.399933 4612 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 07:50:22 crc kubenswrapper[4612]: E0227 07:50:22.500880 4612 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 07:50:22 crc kubenswrapper[4612]: E0227 07:50:22.601026 4612 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 07:50:22 crc kubenswrapper[4612]: E0227 07:50:22.701722 4612 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 07:50:22 crc kubenswrapper[4612]: E0227 07:50:22.802852 4612 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 07:50:22 crc kubenswrapper[4612]: E0227 07:50:22.903893 4612 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 07:50:22 crc kubenswrapper[4612]: E0227 07:50:22.925414 4612 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Feb 27 07:50:23 crc kubenswrapper[4612]: E0227 07:50:23.004463 4612 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 07:50:23 crc kubenswrapper[4612]: E0227 07:50:23.105660 4612 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 07:50:23 crc kubenswrapper[4612]: E0227 07:50:23.206769 4612 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 07:50:23 crc kubenswrapper[4612]: E0227 07:50:23.307895 4612 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 07:50:23 crc kubenswrapper[4612]: E0227 07:50:23.408859 4612 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 07:50:23 crc kubenswrapper[4612]: E0227 07:50:23.509767 4612 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 07:50:23 crc kubenswrapper[4612]: E0227 07:50:23.610657 4612 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 07:50:23 crc kubenswrapper[4612]: E0227 07:50:23.711008 4612 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 07:50:23 crc kubenswrapper[4612]: E0227 07:50:23.811912 4612 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 07:50:23 crc kubenswrapper[4612]: E0227 07:50:23.912757 4612 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 07:50:24 crc kubenswrapper[4612]: E0227 07:50:24.013453 4612 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 07:50:24 crc kubenswrapper[4612]: E0227 07:50:24.114630 4612 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 07:50:24 crc kubenswrapper[4612]: E0227 07:50:24.215429 4612 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 07:50:24 crc kubenswrapper[4612]: E0227 07:50:24.316020 4612 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 07:50:24 crc kubenswrapper[4612]: E0227 07:50:24.416957 4612 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 07:50:24 crc kubenswrapper[4612]: E0227 07:50:24.517129 4612 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 07:50:24 crc kubenswrapper[4612]: E0227 07:50:24.618076 4612 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 07:50:24 crc kubenswrapper[4612]: E0227 07:50:24.719405 4612 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 07:50:24 crc kubenswrapper[4612]: E0227 07:50:24.819771 4612 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 07:50:24 crc kubenswrapper[4612]: I0227 07:50:24.852927 4612 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 07:50:24 crc kubenswrapper[4612]: I0227 07:50:24.854398 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 07:50:24 crc kubenswrapper[4612]: I0227 07:50:24.854459 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 07:50:24 crc kubenswrapper[4612]: I0227 07:50:24.854489 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 07:50:24 crc kubenswrapper[4612]: I0227 07:50:24.855679 4612 scope.go:117] "RemoveContainer" containerID="6fdb61a61c41ceec6aa497297d1853ae6fec429a9dc998e0f49e8f41f41b7e9c" Feb 27 07:50:24 crc kubenswrapper[4612]: E0227 07:50:24.855979 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Feb 27 07:50:24 crc kubenswrapper[4612]: E0227 07:50:24.920617 4612 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 07:50:25 crc kubenswrapper[4612]: E0227 07:50:25.021020 4612 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 07:50:25 crc kubenswrapper[4612]: E0227 07:50:25.122368 4612 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 07:50:25 crc kubenswrapper[4612]: E0227 07:50:25.222759 4612 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 07:50:25 crc kubenswrapper[4612]: E0227 07:50:25.324212 4612 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 07:50:25 crc kubenswrapper[4612]: E0227 07:50:25.424412 4612 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 07:50:25 crc kubenswrapper[4612]: E0227 07:50:25.524684 4612 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 07:50:25 crc kubenswrapper[4612]: E0227 07:50:25.624944 4612 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 07:50:25 crc kubenswrapper[4612]: E0227 07:50:25.725576 4612 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 07:50:25 crc kubenswrapper[4612]: E0227 07:50:25.826541 4612 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 07:50:25 crc kubenswrapper[4612]: I0227 07:50:25.852374 4612 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 27 07:50:25 crc kubenswrapper[4612]: I0227 07:50:25.853521 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 07:50:25 crc kubenswrapper[4612]: I0227 07:50:25.853670 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 07:50:25 crc kubenswrapper[4612]: I0227 07:50:25.853850 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 07:50:25 crc kubenswrapper[4612]: E0227 07:50:25.927873 4612 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 07:50:26 crc kubenswrapper[4612]: E0227 07:50:26.028684 4612 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 07:50:26 crc kubenswrapper[4612]: E0227 07:50:26.129229 4612 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 07:50:26 crc kubenswrapper[4612]: E0227 07:50:26.229733 4612 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 07:50:26 crc kubenswrapper[4612]: E0227 07:50:26.330476 4612 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 07:50:26 crc kubenswrapper[4612]: E0227 07:50:26.431757 4612 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 07:50:26 crc kubenswrapper[4612]: E0227 07:50:26.532855 4612 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 07:50:26 crc kubenswrapper[4612]: E0227 07:50:26.633676 4612 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 07:50:26 crc kubenswrapper[4612]: E0227 07:50:26.734148 4612 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 07:50:26 crc kubenswrapper[4612]: E0227 07:50:26.835388 4612 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 07:50:26 crc kubenswrapper[4612]: E0227 07:50:26.936010 4612 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 07:50:27 crc kubenswrapper[4612]: E0227 07:50:27.036329 4612 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 07:50:27 crc kubenswrapper[4612]: E0227 07:50:27.137364 4612 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 07:50:27 crc kubenswrapper[4612]: E0227 07:50:27.237932 4612 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 07:50:27 crc kubenswrapper[4612]: E0227 07:50:27.338602 4612 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 07:50:27 crc kubenswrapper[4612]: E0227 07:50:27.439204 4612 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 07:50:27 crc kubenswrapper[4612]: E0227 07:50:27.539738 4612 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 07:50:27 crc kubenswrapper[4612]: E0227 07:50:27.639895 4612 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 07:50:27 crc kubenswrapper[4612]: E0227 07:50:27.740273 4612 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 07:50:27 crc kubenswrapper[4612]: E0227 07:50:27.840635 4612 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 07:50:27 crc kubenswrapper[4612]: E0227 07:50:27.940775 4612 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 07:50:28 crc kubenswrapper[4612]: E0227 07:50:28.041202 4612 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 07:50:28 crc kubenswrapper[4612]: E0227 07:50:28.141629 4612 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 07:50:28 crc kubenswrapper[4612]: E0227 07:50:28.242546 4612 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 07:50:28 crc kubenswrapper[4612]: E0227 07:50:28.343443 4612 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 07:50:28 crc kubenswrapper[4612]: E0227 07:50:28.444135 4612 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 07:50:28 crc kubenswrapper[4612]: E0227 07:50:28.544451 4612 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 07:50:28 crc kubenswrapper[4612]: E0227 07:50:28.644881 4612 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 07:50:28 crc kubenswrapper[4612]: I0227 07:50:28.646339 4612 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Feb 27 07:50:28 crc kubenswrapper[4612]: E0227 07:50:28.745247 4612 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 07:50:28 crc kubenswrapper[4612]: E0227 07:50:28.846064 4612 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 07:50:28 crc kubenswrapper[4612]: E0227 07:50:28.947153 4612 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 07:50:29 crc kubenswrapper[4612]: E0227 07:50:29.047814 4612 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 07:50:29 crc kubenswrapper[4612]: E0227 07:50:29.147967 4612 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 07:50:29 crc kubenswrapper[4612]: E0227 07:50:29.248866 4612 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 07:50:29 crc kubenswrapper[4612]: E0227 07:50:29.349364 4612 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 07:50:29 crc kubenswrapper[4612]: E0227 07:50:29.450219 4612 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 07:50:29 crc kubenswrapper[4612]: E0227 07:50:29.551263 4612 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 07:50:29 crc kubenswrapper[4612]: E0227 07:50:29.651490 4612 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 07:50:29 crc kubenswrapper[4612]: E0227 07:50:29.752223 4612 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 07:50:29 crc kubenswrapper[4612]: E0227 07:50:29.852928 4612 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 07:50:29 crc kubenswrapper[4612]: E0227 07:50:29.953728 4612 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 07:50:30 crc kubenswrapper[4612]: E0227 07:50:30.054535 4612 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 07:50:30 crc kubenswrapper[4612]: E0227 07:50:30.155462 4612 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 07:50:30 crc kubenswrapper[4612]: I0227 07:50:30.255140 4612 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Feb 27 07:50:30 crc kubenswrapper[4612]: E0227 07:50:30.256086 4612 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 07:50:30 crc kubenswrapper[4612]: E0227 07:50:30.356522 4612 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 07:50:30 crc kubenswrapper[4612]: E0227 07:50:30.457396 4612 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 07:50:30 crc kubenswrapper[4612]: E0227 07:50:30.558346 4612 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 07:50:30 crc kubenswrapper[4612]: E0227 07:50:30.658958 4612 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 07:50:30 crc kubenswrapper[4612]: E0227 07:50:30.759105 4612 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 07:50:30 crc kubenswrapper[4612]: E0227 07:50:30.860328 4612 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 07:50:30 crc kubenswrapper[4612]: E0227 07:50:30.961314 4612 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 07:50:31 crc kubenswrapper[4612]: E0227 07:50:31.062019 4612 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 07:50:31 crc kubenswrapper[4612]: E0227 07:50:31.141969 4612 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": node \"crc\" not found" Feb 27 07:50:31 crc kubenswrapper[4612]: I0227 07:50:31.145577 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 07:50:31 crc kubenswrapper[4612]: I0227 07:50:31.145623 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 07:50:31 crc kubenswrapper[4612]: I0227 07:50:31.145631 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 07:50:31 crc kubenswrapper[4612]: I0227 07:50:31.145644 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 07:50:31 crc kubenswrapper[4612]: I0227 07:50:31.145654 4612 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T07:50:31Z","lastTransitionTime":"2026-02-27T07:50:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 07:50:31 crc kubenswrapper[4612]: E0227 07:50:31.153916 4612 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148060Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608860Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T07:50:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T07:50:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T07:50:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T07:50:31Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T07:50:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T07:50:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T07:50:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T07:50:31Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a8a53c94-4eff-42ba-8a96-6c53b380c5c4\\\",\\\"systemUUID\\\":\\\"6e98ef2c-c447-4ff8-86d3-9b32f302ee4b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 27 07:50:31 crc kubenswrapper[4612]: I0227 07:50:31.157679 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 07:50:31 crc kubenswrapper[4612]: I0227 07:50:31.157742 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 07:50:31 crc kubenswrapper[4612]: I0227 07:50:31.157751 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 07:50:31 crc kubenswrapper[4612]: I0227 07:50:31.157765 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 07:50:31 crc kubenswrapper[4612]: I0227 07:50:31.157792 4612 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T07:50:31Z","lastTransitionTime":"2026-02-27T07:50:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 07:50:31 crc kubenswrapper[4612]: E0227 07:50:31.173156 4612 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148060Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608860Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T07:50:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T07:50:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T07:50:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T07:50:31Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T07:50:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T07:50:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T07:50:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T07:50:31Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a8a53c94-4eff-42ba-8a96-6c53b380c5c4\\\",\\\"systemUUID\\\":\\\"6e98ef2c-c447-4ff8-86d3-9b32f302ee4b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 27 07:50:31 crc kubenswrapper[4612]: I0227 07:50:31.176708 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 07:50:31 crc kubenswrapper[4612]: I0227 07:50:31.176731 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 07:50:31 crc kubenswrapper[4612]: I0227 07:50:31.176740 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 07:50:31 crc kubenswrapper[4612]: I0227 07:50:31.176753 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 07:50:31 crc kubenswrapper[4612]: I0227 07:50:31.176762 4612 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T07:50:31Z","lastTransitionTime":"2026-02-27T07:50:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 07:50:31 crc kubenswrapper[4612]: E0227 07:50:31.189530 4612 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148060Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608860Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T07:50:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T07:50:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T07:50:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T07:50:31Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T07:50:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T07:50:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T07:50:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T07:50:31Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a8a53c94-4eff-42ba-8a96-6c53b380c5c4\\\",\\\"systemUUID\\\":\\\"6e98ef2c-c447-4ff8-86d3-9b32f302ee4b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 27 07:50:31 crc kubenswrapper[4612]: I0227 07:50:31.194128 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 07:50:31 crc kubenswrapper[4612]: I0227 07:50:31.194178 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 07:50:31 crc kubenswrapper[4612]: I0227 07:50:31.194191 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 07:50:31 crc kubenswrapper[4612]: I0227 07:50:31.194207 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 07:50:31 crc kubenswrapper[4612]: I0227 07:50:31.194216 4612 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T07:50:31Z","lastTransitionTime":"2026-02-27T07:50:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 07:50:31 crc kubenswrapper[4612]: E0227 07:50:31.203758 4612 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148060Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608860Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T07:50:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T07:50:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T07:50:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T07:50:31Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T07:50:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T07:50:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-27T07:50:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-27T07:50:31Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a8a53c94-4eff-42ba-8a96-6c53b380c5c4\\\",\\\"systemUUID\\\":\\\"6e98ef2c-c447-4ff8-86d3-9b32f302ee4b\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 27 07:50:31 crc kubenswrapper[4612]: E0227 07:50:31.203870 4612 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Feb 27 07:50:31 crc kubenswrapper[4612]: E0227 07:50:31.203901 4612 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 07:50:31 crc kubenswrapper[4612]: E0227 07:50:31.304755 4612 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 07:50:31 crc kubenswrapper[4612]: E0227 07:50:31.406124 4612 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 07:50:31 crc kubenswrapper[4612]: E0227 07:50:31.507488 4612 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 07:50:31 crc kubenswrapper[4612]: E0227 07:50:31.608521 4612 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 07:50:31 crc kubenswrapper[4612]: E0227 07:50:31.709612 4612 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 07:50:31 crc kubenswrapper[4612]: E0227 07:50:31.810235 4612 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 07:50:31 crc kubenswrapper[4612]: E0227 07:50:31.910838 4612 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 07:50:32 crc kubenswrapper[4612]: E0227 07:50:32.010970 4612 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 07:50:32 crc kubenswrapper[4612]: E0227 07:50:32.112041 4612 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 07:50:32 crc kubenswrapper[4612]: E0227 07:50:32.212793 4612 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 07:50:32 crc kubenswrapper[4612]: E0227 07:50:32.313767 4612 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 07:50:32 crc kubenswrapper[4612]: E0227 07:50:32.414675 4612 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 07:50:32 crc kubenswrapper[4612]: E0227 07:50:32.515548 4612 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 07:50:32 crc kubenswrapper[4612]: E0227 07:50:32.616031 4612 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 07:50:32 crc kubenswrapper[4612]: E0227 07:50:32.716848 4612 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 07:50:32 crc kubenswrapper[4612]: E0227 07:50:32.818396 4612 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 07:50:32 crc kubenswrapper[4612]: E0227 07:50:32.919250 4612 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 07:50:32 crc kubenswrapper[4612]: E0227 07:50:32.925751 4612 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Feb 27 07:50:33 crc kubenswrapper[4612]: E0227 07:50:33.019631 4612 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 07:50:33 crc kubenswrapper[4612]: E0227 07:50:33.120383 4612 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 07:50:33 crc kubenswrapper[4612]: E0227 07:50:33.221433 4612 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 07:50:33 crc kubenswrapper[4612]: E0227 07:50:33.322840 4612 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 07:50:33 crc kubenswrapper[4612]: E0227 07:50:33.423790 4612 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 07:50:33 crc kubenswrapper[4612]: E0227 07:50:33.524748 4612 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 07:50:33 crc kubenswrapper[4612]: E0227 07:50:33.624872 4612 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 07:50:33 crc kubenswrapper[4612]: E0227 07:50:33.725336 4612 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 07:50:33 crc kubenswrapper[4612]: E0227 07:50:33.825925 4612 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 07:50:33 crc kubenswrapper[4612]: E0227 07:50:33.927048 4612 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 07:50:34 crc kubenswrapper[4612]: E0227 07:50:34.028413 4612 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 07:50:34 crc kubenswrapper[4612]: E0227 07:50:34.129260 4612 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 07:50:34 crc kubenswrapper[4612]: E0227 07:50:34.230209 4612 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 07:50:34 crc kubenswrapper[4612]: E0227 07:50:34.331131 4612 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 07:50:34 crc kubenswrapper[4612]: E0227 07:50:34.431458 4612 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 07:50:34 crc kubenswrapper[4612]: E0227 07:50:34.532442 4612 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 07:50:34 crc kubenswrapper[4612]: E0227 07:50:34.633041 4612 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 07:50:34 crc kubenswrapper[4612]: E0227 07:50:34.733158 4612 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 07:50:34 crc kubenswrapper[4612]: E0227 07:50:34.833231 4612 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 07:50:34 crc kubenswrapper[4612]: E0227 07:50:34.937586 4612 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 07:50:35 crc kubenswrapper[4612]: E0227 07:50:35.038049 4612 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.054035 4612 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.140333 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.140365 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.140375 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.140391 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.140405 4612 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T07:50:35Z","lastTransitionTime":"2026-02-27T07:50:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.241885 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.241924 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.241937 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.241953 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.241966 4612 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T07:50:35Z","lastTransitionTime":"2026-02-27T07:50:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.344147 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.344193 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.344202 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.344215 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.344227 4612 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T07:50:35Z","lastTransitionTime":"2026-02-27T07:50:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.446442 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.446493 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.446508 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.446529 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.446544 4612 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T07:50:35Z","lastTransitionTime":"2026-02-27T07:50:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.548437 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.548475 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.548486 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.548501 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.548512 4612 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T07:50:35Z","lastTransitionTime":"2026-02-27T07:50:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.650804 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.651042 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.651138 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.651217 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.651285 4612 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T07:50:35Z","lastTransitionTime":"2026-02-27T07:50:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.753576 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.753847 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.753937 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.754037 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.754145 4612 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T07:50:35Z","lastTransitionTime":"2026-02-27T07:50:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.819871 4612 apiserver.go:52] "Watching apiserver" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.824013 4612 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.824275 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-operator/network-operator-58b4c7f79c-55gtf","openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h"] Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.824630 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 07:50:35 crc kubenswrapper[4612]: E0227 07:50:35.824793 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.824845 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.824869 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.825007 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.825081 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 07:50:35 crc kubenswrapper[4612]: E0227 07:50:35.825288 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.825346 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 07:50:35 crc kubenswrapper[4612]: E0227 07:50:35.825500 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.828159 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.828545 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.828745 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.828935 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.829105 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.829300 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.829538 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.830023 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.830866 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.846760 4612 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T07:50:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T07:50:35Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.856549 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.856605 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.856620 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.856641 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.856657 4612 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T07:50:35Z","lastTransitionTime":"2026-02-27T07:50:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.867670 4612 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T07:50:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T07:50:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.877816 4612 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T07:50:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T07:50:35Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.878210 4612 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.890464 4612 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T07:50:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T07:50:35Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.899537 4612 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T07:50:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T07:50:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.900620 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.900740 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.900764 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.900778 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.900794 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.900810 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.900828 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.900842 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.900858 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.900872 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.900888 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.900906 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.900920 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.900934 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.900950 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.900964 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.900979 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.900997 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.901011 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.901025 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.901041 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.901056 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.901071 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.901086 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.901099 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.901114 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.901131 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.901146 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.901160 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.901173 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.901188 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.901202 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.901215 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.901230 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.901245 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.901258 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.901273 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.901287 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.901301 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.901317 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.901330 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.901344 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.901361 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.901374 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.901389 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.901404 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.901419 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.901433 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.901447 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.901462 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.901477 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.901493 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.901509 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.901523 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.901538 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.901552 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.901567 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.901581 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.901598 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.901611 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.901625 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.901639 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.901655 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.901669 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.901682 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.901711 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.901726 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.901741 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.901757 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.901771 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.901785 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.901800 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.901816 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.901861 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.901878 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.901893 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.901907 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.901922 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.901938 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.901954 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.902002 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.902018 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.902032 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.902047 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.902062 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.902076 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.902091 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.902106 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.902120 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.902135 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.902152 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.902167 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.902183 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.902197 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.902211 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.902226 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.902241 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.902256 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.902272 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.902288 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.902303 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.902318 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.902333 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.902349 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.902385 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.902402 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.902428 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.902446 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.902464 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.902483 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.902497 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.902512 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.902526 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.902541 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.902556 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.902572 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.902588 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.902602 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.902617 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.902633 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.902647 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.902662 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.902678 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.902709 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.902724 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.902740 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.902755 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.902770 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.902786 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.902801 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.902816 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.902832 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.902847 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.902863 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.902884 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.902900 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.902920 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.902935 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.902952 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.902967 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.902982 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.902999 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.903015 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.903030 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.903047 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.903062 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.903078 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.903093 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.903108 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.903122 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.903138 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.903154 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.903170 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.903186 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.903202 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.903217 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.903233 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.903249 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.903266 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.903282 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.903298 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.903313 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.903332 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.903350 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.903370 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.903391 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.903416 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.903434 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.903451 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.903467 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.903485 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.903502 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.903517 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.903534 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.903550 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.903566 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.903581 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.903596 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.903611 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.903627 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.903643 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.903660 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.903676 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.904060 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.904147 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.904340 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 07:50:35 crc kubenswrapper[4612]: E0227 07:50:35.904430 4612 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 07:50:36.404410955 +0000 UTC m=+94.258340973 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.904552 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.904570 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.904842 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.904854 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.904979 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.905004 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.905062 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.905085 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.905103 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.905956 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.905976 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.905998 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.906026 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.906045 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.906062 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.906079 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.906107 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.906123 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.906158 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.906342 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.906365 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.906386 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.906809 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.906839 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.906860 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.906882 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.907046 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.907158 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.907176 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.907200 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.907219 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.907236 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.907277 4612 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.907288 4612 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.907298 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.907308 4612 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.907325 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.907335 4612 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.907344 4612 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.907354 4612 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.907363 4612 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.905166 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.905220 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.905416 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.905441 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.905629 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.917843 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.905646 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.905814 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.905919 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.905925 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.906276 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.906309 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.906507 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.906584 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.906733 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.906767 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.906837 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.907008 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.907075 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 07:50:35 crc kubenswrapper[4612]: E0227 07:50:35.907409 4612 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.918102 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 07:50:35 crc kubenswrapper[4612]: E0227 07:50:35.918162 4612 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-27 07:50:36.418142213 +0000 UTC m=+94.272072211 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.918143 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.907929 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.908164 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.918227 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.908226 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.908276 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.908719 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.909485 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.910043 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.911479 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.911597 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.911612 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.911815 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.911990 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.912060 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.912257 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.912361 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.918501 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.918554 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.918778 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.919132 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.919210 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.913269 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.913285 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.913615 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.913690 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.913874 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.914138 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.914186 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.914271 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.914362 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.914970 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.916402 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.916431 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.916584 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.917113 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.917380 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.917713 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.917766 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.917792 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.907816 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.907753 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.919377 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.919757 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.919960 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.920232 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.920462 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.921630 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.922030 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.922594 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.922835 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.922893 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.922950 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.923021 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.922860 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.923231 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.923451 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.923502 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.923611 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.923066 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 07:50:35 crc kubenswrapper[4612]: E0227 07:50:35.923854 4612 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 27 07:50:35 crc kubenswrapper[4612]: E0227 07:50:35.923932 4612 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-27 07:50:36.423912622 +0000 UTC m=+94.277842710 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.924168 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.924341 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.924606 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.924682 4612 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T07:50:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T07:50:35Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.926855 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.924747 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.925069 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.925187 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.925832 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.925891 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.927648 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.929064 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.929310 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.929388 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.928499 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.925909 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.926291 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.926341 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.926497 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.926606 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.926718 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.926772 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.927232 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.927318 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.927488 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.926817 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.927797 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.927831 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.928095 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.928182 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.928777 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.913247 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.933673 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.926101 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.934463 4612 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.937342 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.949283 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.950015 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 27 07:50:35 crc kubenswrapper[4612]: E0227 07:50:35.950608 4612 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 27 07:50:35 crc kubenswrapper[4612]: E0227 07:50:35.950634 4612 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 27 07:50:35 crc kubenswrapper[4612]: E0227 07:50:35.950650 4612 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 27 07:50:35 crc kubenswrapper[4612]: E0227 07:50:35.950727 4612 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-27 07:50:36.45068733 +0000 UTC m=+94.304617348 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.951667 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 27 07:50:35 crc kubenswrapper[4612]: E0227 07:50:35.951835 4612 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.954332 4612 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T07:50:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T07:50:35Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.955476 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.952809 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.954512 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.955057 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.955109 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.955181 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.951462 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 07:50:35 crc kubenswrapper[4612]: E0227 07:50:35.956367 4612 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 27 07:50:35 crc kubenswrapper[4612]: E0227 07:50:35.956538 4612 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 27 07:50:35 crc kubenswrapper[4612]: E0227 07:50:35.957157 4612 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-27 07:50:36.456965723 +0000 UTC m=+94.310895721 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.961507 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.961857 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.962202 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.962534 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.962953 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.963195 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.963580 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.963608 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.963620 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.963637 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.963650 4612 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T07:50:35Z","lastTransitionTime":"2026-02-27T07:50:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.963979 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.964665 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.963447 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.964837 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.965385 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.965873 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.965943 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.966109 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.966559 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.966583 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.966805 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.967265 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.969476 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.969591 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.969778 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.969840 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.970051 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.970262 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.970494 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.970798 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.970859 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.971040 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.971294 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.971383 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.971554 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.971627 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.972180 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.973079 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.964546 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.973228 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.973385 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.973891 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.974383 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.974847 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.974487 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.975038 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.974440 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.974453 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.975128 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.974488 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.975248 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.975538 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.975288 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.975354 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.972941 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.975881 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.976224 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.977814 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.977944 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.978168 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.978296 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.978372 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.978624 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.979629 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.979749 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.980245 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.983912 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.984269 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.984568 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 07:50:35 crc kubenswrapper[4612]: I0227 07:50:35.999225 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.003893 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.005659 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.008099 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.008201 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.008290 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.008396 4612 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.008470 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.008543 4612 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.008620 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.008678 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.008758 4612 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.008828 4612 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.008905 4612 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.008967 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.008334 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.009029 4612 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.009054 4612 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.009063 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.009072 4612 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.009082 4612 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.009090 4612 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.009099 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.009107 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.009115 4612 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.009123 4612 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.009131 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.009139 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.009147 4612 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.009155 4612 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.009163 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.009171 4612 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.009179 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.009188 4612 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.009196 4612 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.009204 4612 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.009212 4612 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.009219 4612 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.009228 4612 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.009236 4612 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.009244 4612 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.009252 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.009260 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.009268 4612 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.009277 4612 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.009285 4612 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.009292 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.009301 4612 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.009309 4612 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.009317 4612 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.009325 4612 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.009333 4612 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.009343 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.009352 4612 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.009360 4612 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.009368 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.009377 4612 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.009385 4612 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.009393 4612 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.009401 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.009410 4612 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.009418 4612 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.009426 4612 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.009433 4612 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.009441 4612 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.009449 4612 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.009456 4612 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.009464 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.009472 4612 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.009480 4612 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.009488 4612 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.009495 4612 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.009504 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.009514 4612 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.009522 4612 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.009530 4612 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.009537 4612 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.009545 4612 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.009553 4612 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.009561 4612 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.009569 4612 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.009576 4612 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.009584 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.009592 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.009600 4612 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.009607 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.009616 4612 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.009624 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.009631 4612 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.009639 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.009647 4612 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.009655 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.009663 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.009672 4612 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.009680 4612 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.009703 4612 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.009711 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.009719 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.009727 4612 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.009734 4612 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.009742 4612 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.009750 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.009758 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.009766 4612 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.009773 4612 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.009780 4612 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.009787 4612 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.009796 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.009803 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.009811 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.009818 4612 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.009826 4612 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.009834 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.009842 4612 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.009850 4612 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.009857 4612 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.009864 4612 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.009871 4612 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.009879 4612 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.009886 4612 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.009894 4612 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.009902 4612 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.009910 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.009918 4612 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.009925 4612 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.009935 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.009943 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.009950 4612 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.009958 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.009966 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.009974 4612 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.009981 4612 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.009988 4612 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.009995 4612 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.010003 4612 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.010010 4612 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.010018 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.010025 4612 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.010034 4612 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.010041 4612 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.010048 4612 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.010055 4612 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.010065 4612 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.010072 4612 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.010080 4612 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.010087 4612 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.010094 4612 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.010102 4612 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.010110 4612 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.010118 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.010126 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.010134 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.010143 4612 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.010151 4612 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.010158 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.010166 4612 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.010174 4612 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.010181 4612 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.010188 4612 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.010196 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.010204 4612 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.010212 4612 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.010220 4612 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.010228 4612 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.010237 4612 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.010244 4612 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.010252 4612 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.010260 4612 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.010267 4612 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.010274 4612 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.010282 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.010290 4612 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.010296 4612 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.010304 4612 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.010312 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.010319 4612 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.010328 4612 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.010337 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.010346 4612 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.010353 4612 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.010362 4612 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.010369 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.010376 4612 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.010384 4612 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.010395 4612 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.010403 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.010442 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.010451 4612 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.010459 4612 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.010466 4612 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.010474 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.010754 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.066184 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.066233 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.066246 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.066265 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.066277 4612 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T07:50:36Z","lastTransitionTime":"2026-02-27T07:50:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.110947 4612 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.141964 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.150957 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.157140 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 27 07:50:36 crc kubenswrapper[4612]: W0227 07:50:36.158880 4612 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podef543e1b_8068_4ea3_b32a_61027b32e95d.slice/crio-ebd89d3ac59778b587b3b4fe06908d3ad72e2689962c9e3ec516bd37f1e98383 WatchSource:0}: Error finding container ebd89d3ac59778b587b3b4fe06908d3ad72e2689962c9e3ec516bd37f1e98383: Status 404 returned error can't find the container with id ebd89d3ac59778b587b3b4fe06908d3ad72e2689962c9e3ec516bd37f1e98383 Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.171190 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.171225 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.171237 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.171254 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.171266 4612 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T07:50:36Z","lastTransitionTime":"2026-02-27T07:50:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.240373 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"ebd89d3ac59778b587b3b4fe06908d3ad72e2689962c9e3ec516bd37f1e98383"} Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.241612 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"1f5eb7cfdc7bfb194cbb16d602c3bd2505e7ebdf2323f4779b84faa6f29e741d"} Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.242783 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"46bdd1422961eed69d16b2aad13a9cc9e309dfd838e1a63e2ecfff103b0d59bb"} Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.273205 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.273244 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.273255 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.273269 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.273281 4612 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T07:50:36Z","lastTransitionTime":"2026-02-27T07:50:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.375590 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.375617 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.375627 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.375642 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.375653 4612 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T07:50:36Z","lastTransitionTime":"2026-02-27T07:50:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.412787 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 07:50:36 crc kubenswrapper[4612]: E0227 07:50:36.412929 4612 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 07:50:37.412915218 +0000 UTC m=+95.266845216 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.478507 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.478552 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.478562 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.478582 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.478593 4612 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T07:50:36Z","lastTransitionTime":"2026-02-27T07:50:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.513440 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.513483 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.513513 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.513535 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 07:50:36 crc kubenswrapper[4612]: E0227 07:50:36.513558 4612 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 27 07:50:36 crc kubenswrapper[4612]: E0227 07:50:36.513624 4612 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-27 07:50:37.513607712 +0000 UTC m=+95.367537710 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 27 07:50:36 crc kubenswrapper[4612]: E0227 07:50:36.513625 4612 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 27 07:50:36 crc kubenswrapper[4612]: E0227 07:50:36.513639 4612 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 27 07:50:36 crc kubenswrapper[4612]: E0227 07:50:36.513650 4612 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 27 07:50:36 crc kubenswrapper[4612]: E0227 07:50:36.513655 4612 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 27 07:50:36 crc kubenswrapper[4612]: E0227 07:50:36.513687 4612 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 27 07:50:36 crc kubenswrapper[4612]: E0227 07:50:36.513748 4612 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 27 07:50:36 crc kubenswrapper[4612]: E0227 07:50:36.513762 4612 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 27 07:50:36 crc kubenswrapper[4612]: E0227 07:50:36.513674 4612 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-27 07:50:37.513666384 +0000 UTC m=+95.367596372 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 27 07:50:36 crc kubenswrapper[4612]: E0227 07:50:36.513821 4612 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-27 07:50:37.513800898 +0000 UTC m=+95.367730896 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 27 07:50:36 crc kubenswrapper[4612]: E0227 07:50:36.513842 4612 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-27 07:50:37.513828119 +0000 UTC m=+95.367758117 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.580847 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.580885 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.580896 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.580910 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.580920 4612 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T07:50:36Z","lastTransitionTime":"2026-02-27T07:50:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.682988 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.683032 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.683046 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.683066 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.683084 4612 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T07:50:36Z","lastTransitionTime":"2026-02-27T07:50:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.785549 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.785601 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.785615 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.785635 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.785643 4612 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T07:50:36Z","lastTransitionTime":"2026-02-27T07:50:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.856345 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.857231 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.858347 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.859291 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.860142 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.862166 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.863659 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.864901 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.866262 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.869060 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.870132 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.872565 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.873782 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.875874 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.877140 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.878310 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.880327 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.881173 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.883505 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.884791 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.885923 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.888290 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.888344 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.888362 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.888383 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.888400 4612 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T07:50:36Z","lastTransitionTime":"2026-02-27T07:50:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.888509 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.889649 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.892074 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.893008 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.895540 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.897096 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.898286 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.900750 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.902025 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.904087 4612 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.904355 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.907903 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.909186 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.909764 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.911979 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.913071 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.914315 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.915171 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.916612 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.917296 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.918603 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.920089 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.922682 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.924158 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.924920 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.926305 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.927313 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.927996 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.929194 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.929849 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.931087 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.931940 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.932857 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.991213 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.991252 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.991265 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.991288 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 07:50:36 crc kubenswrapper[4612]: I0227 07:50:36.991302 4612 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T07:50:36Z","lastTransitionTime":"2026-02-27T07:50:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 07:50:37 crc kubenswrapper[4612]: I0227 07:50:37.093725 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 07:50:37 crc kubenswrapper[4612]: I0227 07:50:37.093760 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 07:50:37 crc kubenswrapper[4612]: I0227 07:50:37.093769 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 07:50:37 crc kubenswrapper[4612]: I0227 07:50:37.093783 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 07:50:37 crc kubenswrapper[4612]: I0227 07:50:37.093794 4612 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T07:50:37Z","lastTransitionTime":"2026-02-27T07:50:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 07:50:37 crc kubenswrapper[4612]: I0227 07:50:37.196573 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 07:50:37 crc kubenswrapper[4612]: I0227 07:50:37.196628 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 07:50:37 crc kubenswrapper[4612]: I0227 07:50:37.196640 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 07:50:37 crc kubenswrapper[4612]: I0227 07:50:37.196658 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 07:50:37 crc kubenswrapper[4612]: I0227 07:50:37.196672 4612 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T07:50:37Z","lastTransitionTime":"2026-02-27T07:50:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 07:50:37 crc kubenswrapper[4612]: I0227 07:50:37.246719 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"73b41e5057e67f925ad6e0d7e58a0a353819df323d0c017b892fc0241645f877"} Feb 27 07:50:37 crc kubenswrapper[4612]: I0227 07:50:37.250865 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"07c7ec4abd65466b9a815cacbd2883943fea5b9d45f749e28508179791fd0731"} Feb 27 07:50:37 crc kubenswrapper[4612]: I0227 07:50:37.250915 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"c698add58f7dafd12af40c2b95c71671b493834624d5250451d4ca5689575baf"} Feb 27 07:50:37 crc kubenswrapper[4612]: I0227 07:50:37.264961 4612 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T07:50:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T07:50:35Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T07:50:37Z is after 2025-08-24T17:21:41Z" Feb 27 07:50:37 crc kubenswrapper[4612]: I0227 07:50:37.281826 4612 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T07:50:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T07:50:35Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T07:50:37Z is after 2025-08-24T17:21:41Z" Feb 27 07:50:37 crc kubenswrapper[4612]: I0227 07:50:37.296831 4612 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T07:50:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T07:50:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T07:50:37Z is after 2025-08-24T17:21:41Z" Feb 27 07:50:37 crc kubenswrapper[4612]: I0227 07:50:37.298309 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 07:50:37 crc kubenswrapper[4612]: I0227 07:50:37.298341 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 07:50:37 crc kubenswrapper[4612]: I0227 07:50:37.298348 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 07:50:37 crc kubenswrapper[4612]: I0227 07:50:37.298363 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 07:50:37 crc kubenswrapper[4612]: I0227 07:50:37.298372 4612 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T07:50:37Z","lastTransitionTime":"2026-02-27T07:50:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 07:50:37 crc kubenswrapper[4612]: I0227 07:50:37.311164 4612 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T07:50:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T07:50:35Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T07:50:37Z is after 2025-08-24T17:21:41Z" Feb 27 07:50:37 crc kubenswrapper[4612]: I0227 07:50:37.325906 4612 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T07:50:37Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T07:50:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T07:50:37Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://73b41e5057e67f925ad6e0d7e58a0a353819df323d0c017b892fc0241645f877\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T07:50:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T07:50:37Z is after 2025-08-24T17:21:41Z" Feb 27 07:50:37 crc kubenswrapper[4612]: I0227 07:50:37.336628 4612 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T07:50:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T07:50:35Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T07:50:37Z is after 2025-08-24T17:21:41Z" Feb 27 07:50:37 crc kubenswrapper[4612]: I0227 07:50:37.348079 4612 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T07:50:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T07:50:35Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T07:50:37Z is after 2025-08-24T17:21:41Z" Feb 27 07:50:37 crc kubenswrapper[4612]: I0227 07:50:37.358574 4612 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T07:50:37Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T07:50:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T07:50:37Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://73b41e5057e67f925ad6e0d7e58a0a353819df323d0c017b892fc0241645f877\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T07:50:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T07:50:37Z is after 2025-08-24T17:21:41Z" Feb 27 07:50:37 crc kubenswrapper[4612]: I0227 07:50:37.368155 4612 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T07:50:37Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T07:50:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T07:50:37Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://07c7ec4abd65466b9a815cacbd2883943fea5b9d45f749e28508179791fd0731\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T07:50:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c698add58f7dafd12af40c2b95c71671b493834624d5250451d4ca5689575baf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-27T07:50:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T07:50:37Z is after 2025-08-24T17:21:41Z" Feb 27 07:50:37 crc kubenswrapper[4612]: I0227 07:50:37.379063 4612 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T07:50:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T07:50:35Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T07:50:37Z is after 2025-08-24T17:21:41Z" Feb 27 07:50:37 crc kubenswrapper[4612]: I0227 07:50:37.389015 4612 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T07:50:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T07:50:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T07:50:37Z is after 2025-08-24T17:21:41Z" Feb 27 07:50:37 crc kubenswrapper[4612]: I0227 07:50:37.400907 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 07:50:37 crc kubenswrapper[4612]: I0227 07:50:37.400938 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 07:50:37 crc kubenswrapper[4612]: I0227 07:50:37.400949 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 07:50:37 crc kubenswrapper[4612]: I0227 07:50:37.400964 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 07:50:37 crc kubenswrapper[4612]: I0227 07:50:37.400977 4612 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T07:50:37Z","lastTransitionTime":"2026-02-27T07:50:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 07:50:37 crc kubenswrapper[4612]: I0227 07:50:37.404062 4612 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-27T07:50:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-27T07:50:35Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-27T07:50:37Z is after 2025-08-24T17:21:41Z" Feb 27 07:50:37 crc kubenswrapper[4612]: I0227 07:50:37.420546 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 07:50:37 crc kubenswrapper[4612]: E0227 07:50:37.420685 4612 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 07:50:39.420665128 +0000 UTC m=+97.274595126 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 07:50:37 crc kubenswrapper[4612]: I0227 07:50:37.481980 4612 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Feb 27 07:50:37 crc kubenswrapper[4612]: I0227 07:50:37.503440 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 07:50:37 crc kubenswrapper[4612]: I0227 07:50:37.503480 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 07:50:37 crc kubenswrapper[4612]: I0227 07:50:37.503490 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 07:50:37 crc kubenswrapper[4612]: I0227 07:50:37.503506 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 07:50:37 crc kubenswrapper[4612]: I0227 07:50:37.503516 4612 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T07:50:37Z","lastTransitionTime":"2026-02-27T07:50:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 07:50:37 crc kubenswrapper[4612]: I0227 07:50:37.521884 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 07:50:37 crc kubenswrapper[4612]: I0227 07:50:37.521923 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 07:50:37 crc kubenswrapper[4612]: I0227 07:50:37.521948 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 07:50:37 crc kubenswrapper[4612]: I0227 07:50:37.521977 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 07:50:37 crc kubenswrapper[4612]: E0227 07:50:37.522090 4612 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 27 07:50:37 crc kubenswrapper[4612]: E0227 07:50:37.522103 4612 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 27 07:50:37 crc kubenswrapper[4612]: E0227 07:50:37.522182 4612 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-27 07:50:39.522164224 +0000 UTC m=+97.376094222 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 27 07:50:37 crc kubenswrapper[4612]: E0227 07:50:37.522110 4612 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 27 07:50:37 crc kubenswrapper[4612]: E0227 07:50:37.522279 4612 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 27 07:50:37 crc kubenswrapper[4612]: E0227 07:50:37.522111 4612 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 27 07:50:37 crc kubenswrapper[4612]: E0227 07:50:37.522376 4612 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 27 07:50:37 crc kubenswrapper[4612]: E0227 07:50:37.522392 4612 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 27 07:50:37 crc kubenswrapper[4612]: E0227 07:50:37.522105 4612 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 27 07:50:37 crc kubenswrapper[4612]: E0227 07:50:37.522346 4612 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-27 07:50:39.522324289 +0000 UTC m=+97.376254317 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 27 07:50:37 crc kubenswrapper[4612]: E0227 07:50:37.522486 4612 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-27 07:50:39.522472423 +0000 UTC m=+97.376402411 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 27 07:50:37 crc kubenswrapper[4612]: E0227 07:50:37.522504 4612 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-27 07:50:39.522497873 +0000 UTC m=+97.376427871 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 27 07:50:37 crc kubenswrapper[4612]: I0227 07:50:37.606228 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 07:50:37 crc kubenswrapper[4612]: I0227 07:50:37.606268 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 07:50:37 crc kubenswrapper[4612]: I0227 07:50:37.606278 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 07:50:37 crc kubenswrapper[4612]: I0227 07:50:37.606293 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 07:50:37 crc kubenswrapper[4612]: I0227 07:50:37.606303 4612 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T07:50:37Z","lastTransitionTime":"2026-02-27T07:50:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 07:50:37 crc kubenswrapper[4612]: I0227 07:50:37.709019 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 07:50:37 crc kubenswrapper[4612]: I0227 07:50:37.709248 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 07:50:37 crc kubenswrapper[4612]: I0227 07:50:37.709315 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 07:50:37 crc kubenswrapper[4612]: I0227 07:50:37.709391 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 07:50:37 crc kubenswrapper[4612]: I0227 07:50:37.709456 4612 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T07:50:37Z","lastTransitionTime":"2026-02-27T07:50:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 07:50:37 crc kubenswrapper[4612]: I0227 07:50:37.811268 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 07:50:37 crc kubenswrapper[4612]: I0227 07:50:37.811509 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 07:50:37 crc kubenswrapper[4612]: I0227 07:50:37.811584 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 07:50:37 crc kubenswrapper[4612]: I0227 07:50:37.811652 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 07:50:37 crc kubenswrapper[4612]: I0227 07:50:37.811752 4612 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T07:50:37Z","lastTransitionTime":"2026-02-27T07:50:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 07:50:37 crc kubenswrapper[4612]: I0227 07:50:37.852645 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 07:50:37 crc kubenswrapper[4612]: E0227 07:50:37.853004 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 07:50:37 crc kubenswrapper[4612]: I0227 07:50:37.852682 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 07:50:37 crc kubenswrapper[4612]: E0227 07:50:37.853248 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 07:50:37 crc kubenswrapper[4612]: I0227 07:50:37.852660 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 07:50:37 crc kubenswrapper[4612]: E0227 07:50:37.853446 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 07:50:37 crc kubenswrapper[4612]: I0227 07:50:37.913952 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 07:50:37 crc kubenswrapper[4612]: I0227 07:50:37.913977 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 07:50:37 crc kubenswrapper[4612]: I0227 07:50:37.913985 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 07:50:37 crc kubenswrapper[4612]: I0227 07:50:37.913996 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 07:50:37 crc kubenswrapper[4612]: I0227 07:50:37.914007 4612 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T07:50:37Z","lastTransitionTime":"2026-02-27T07:50:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 07:50:37 crc kubenswrapper[4612]: I0227 07:50:37.915908 4612 scope.go:117] "RemoveContainer" containerID="6fdb61a61c41ceec6aa497297d1853ae6fec429a9dc998e0f49e8f41f41b7e9c" Feb 27 07:50:37 crc kubenswrapper[4612]: I0227 07:50:37.915947 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Feb 27 07:50:37 crc kubenswrapper[4612]: E0227 07:50:37.916262 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Feb 27 07:50:38 crc kubenswrapper[4612]: I0227 07:50:38.016484 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 07:50:38 crc kubenswrapper[4612]: I0227 07:50:38.016526 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 07:50:38 crc kubenswrapper[4612]: I0227 07:50:38.016536 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 07:50:38 crc kubenswrapper[4612]: I0227 07:50:38.016551 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 07:50:38 crc kubenswrapper[4612]: I0227 07:50:38.016562 4612 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T07:50:38Z","lastTransitionTime":"2026-02-27T07:50:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 07:50:38 crc kubenswrapper[4612]: I0227 07:50:38.121162 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 07:50:38 crc kubenswrapper[4612]: I0227 07:50:38.121199 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 07:50:38 crc kubenswrapper[4612]: I0227 07:50:38.121208 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 07:50:38 crc kubenswrapper[4612]: I0227 07:50:38.121222 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 07:50:38 crc kubenswrapper[4612]: I0227 07:50:38.121230 4612 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T07:50:38Z","lastTransitionTime":"2026-02-27T07:50:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 07:50:38 crc kubenswrapper[4612]: I0227 07:50:38.226800 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 07:50:38 crc kubenswrapper[4612]: I0227 07:50:38.227407 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 07:50:38 crc kubenswrapper[4612]: I0227 07:50:38.227512 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 07:50:38 crc kubenswrapper[4612]: I0227 07:50:38.227615 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 07:50:38 crc kubenswrapper[4612]: I0227 07:50:38.227743 4612 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T07:50:38Z","lastTransitionTime":"2026-02-27T07:50:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 07:50:38 crc kubenswrapper[4612]: I0227 07:50:38.253162 4612 scope.go:117] "RemoveContainer" containerID="6fdb61a61c41ceec6aa497297d1853ae6fec429a9dc998e0f49e8f41f41b7e9c" Feb 27 07:50:38 crc kubenswrapper[4612]: E0227 07:50:38.253529 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Feb 27 07:50:38 crc kubenswrapper[4612]: I0227 07:50:38.329520 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 07:50:38 crc kubenswrapper[4612]: I0227 07:50:38.329551 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 07:50:38 crc kubenswrapper[4612]: I0227 07:50:38.329561 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 07:50:38 crc kubenswrapper[4612]: I0227 07:50:38.329574 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 07:50:38 crc kubenswrapper[4612]: I0227 07:50:38.329585 4612 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T07:50:38Z","lastTransitionTime":"2026-02-27T07:50:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 07:50:38 crc kubenswrapper[4612]: I0227 07:50:38.431381 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 07:50:38 crc kubenswrapper[4612]: I0227 07:50:38.431414 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 07:50:38 crc kubenswrapper[4612]: I0227 07:50:38.431423 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 07:50:38 crc kubenswrapper[4612]: I0227 07:50:38.431436 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 07:50:38 crc kubenswrapper[4612]: I0227 07:50:38.431445 4612 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T07:50:38Z","lastTransitionTime":"2026-02-27T07:50:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 07:50:38 crc kubenswrapper[4612]: I0227 07:50:38.533993 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 07:50:38 crc kubenswrapper[4612]: I0227 07:50:38.534226 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 07:50:38 crc kubenswrapper[4612]: I0227 07:50:38.534321 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 07:50:38 crc kubenswrapper[4612]: I0227 07:50:38.534418 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 07:50:38 crc kubenswrapper[4612]: I0227 07:50:38.534481 4612 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T07:50:38Z","lastTransitionTime":"2026-02-27T07:50:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 07:50:38 crc kubenswrapper[4612]: I0227 07:50:38.636978 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 07:50:38 crc kubenswrapper[4612]: I0227 07:50:38.637046 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 07:50:38 crc kubenswrapper[4612]: I0227 07:50:38.637071 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 07:50:38 crc kubenswrapper[4612]: I0227 07:50:38.637097 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 07:50:38 crc kubenswrapper[4612]: I0227 07:50:38.637118 4612 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T07:50:38Z","lastTransitionTime":"2026-02-27T07:50:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 07:50:38 crc kubenswrapper[4612]: I0227 07:50:38.739336 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 07:50:38 crc kubenswrapper[4612]: I0227 07:50:38.739372 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 07:50:38 crc kubenswrapper[4612]: I0227 07:50:38.739383 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 07:50:38 crc kubenswrapper[4612]: I0227 07:50:38.739397 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 07:50:38 crc kubenswrapper[4612]: I0227 07:50:38.739406 4612 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T07:50:38Z","lastTransitionTime":"2026-02-27T07:50:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 07:50:38 crc kubenswrapper[4612]: I0227 07:50:38.841188 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 07:50:38 crc kubenswrapper[4612]: I0227 07:50:38.841416 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 07:50:38 crc kubenswrapper[4612]: I0227 07:50:38.841479 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 07:50:38 crc kubenswrapper[4612]: I0227 07:50:38.841565 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 07:50:38 crc kubenswrapper[4612]: I0227 07:50:38.841665 4612 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T07:50:38Z","lastTransitionTime":"2026-02-27T07:50:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 07:50:38 crc kubenswrapper[4612]: I0227 07:50:38.942987 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 07:50:38 crc kubenswrapper[4612]: I0227 07:50:38.943410 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 07:50:38 crc kubenswrapper[4612]: I0227 07:50:38.943518 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 07:50:38 crc kubenswrapper[4612]: I0227 07:50:38.943617 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 07:50:38 crc kubenswrapper[4612]: I0227 07:50:38.943743 4612 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T07:50:38Z","lastTransitionTime":"2026-02-27T07:50:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 07:50:39 crc kubenswrapper[4612]: I0227 07:50:39.046368 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 07:50:39 crc kubenswrapper[4612]: I0227 07:50:39.046433 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 07:50:39 crc kubenswrapper[4612]: I0227 07:50:39.046454 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 07:50:39 crc kubenswrapper[4612]: I0227 07:50:39.046486 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 07:50:39 crc kubenswrapper[4612]: I0227 07:50:39.046507 4612 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T07:50:39Z","lastTransitionTime":"2026-02-27T07:50:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 07:50:39 crc kubenswrapper[4612]: I0227 07:50:39.149308 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 07:50:39 crc kubenswrapper[4612]: I0227 07:50:39.149379 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 07:50:39 crc kubenswrapper[4612]: I0227 07:50:39.149396 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 07:50:39 crc kubenswrapper[4612]: I0227 07:50:39.149424 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 07:50:39 crc kubenswrapper[4612]: I0227 07:50:39.149447 4612 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T07:50:39Z","lastTransitionTime":"2026-02-27T07:50:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 07:50:39 crc kubenswrapper[4612]: I0227 07:50:39.253003 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 07:50:39 crc kubenswrapper[4612]: I0227 07:50:39.253320 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 07:50:39 crc kubenswrapper[4612]: I0227 07:50:39.253529 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 07:50:39 crc kubenswrapper[4612]: I0227 07:50:39.253610 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 07:50:39 crc kubenswrapper[4612]: I0227 07:50:39.253669 4612 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T07:50:39Z","lastTransitionTime":"2026-02-27T07:50:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 07:50:39 crc kubenswrapper[4612]: I0227 07:50:39.255706 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"e8d38797a62d8c9c43b7092205de9f2adb3f8126d8f0a73e0f21c7f41bae43a4"} Feb 27 07:50:39 crc kubenswrapper[4612]: I0227 07:50:39.356318 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 07:50:39 crc kubenswrapper[4612]: I0227 07:50:39.356640 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 07:50:39 crc kubenswrapper[4612]: I0227 07:50:39.356766 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 07:50:39 crc kubenswrapper[4612]: I0227 07:50:39.356851 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 07:50:39 crc kubenswrapper[4612]: I0227 07:50:39.356938 4612 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T07:50:39Z","lastTransitionTime":"2026-02-27T07:50:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 07:50:39 crc kubenswrapper[4612]: I0227 07:50:39.436811 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 07:50:39 crc kubenswrapper[4612]: E0227 07:50:39.436947 4612 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 07:50:43.436923019 +0000 UTC m=+101.290853017 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 07:50:39 crc kubenswrapper[4612]: I0227 07:50:39.458798 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 07:50:39 crc kubenswrapper[4612]: I0227 07:50:39.458826 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 07:50:39 crc kubenswrapper[4612]: I0227 07:50:39.458838 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 07:50:39 crc kubenswrapper[4612]: I0227 07:50:39.458853 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 07:50:39 crc kubenswrapper[4612]: I0227 07:50:39.458863 4612 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T07:50:39Z","lastTransitionTime":"2026-02-27T07:50:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 07:50:39 crc kubenswrapper[4612]: I0227 07:50:39.537610 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 07:50:39 crc kubenswrapper[4612]: E0227 07:50:39.537824 4612 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 27 07:50:39 crc kubenswrapper[4612]: I0227 07:50:39.537979 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 07:50:39 crc kubenswrapper[4612]: E0227 07:50:39.538016 4612 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-27 07:50:43.537991664 +0000 UTC m=+101.391921662 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 27 07:50:39 crc kubenswrapper[4612]: I0227 07:50:39.538054 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 07:50:39 crc kubenswrapper[4612]: I0227 07:50:39.538082 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 07:50:39 crc kubenswrapper[4612]: E0227 07:50:39.538179 4612 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 27 07:50:39 crc kubenswrapper[4612]: E0227 07:50:39.538180 4612 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 27 07:50:39 crc kubenswrapper[4612]: E0227 07:50:39.538196 4612 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 27 07:50:39 crc kubenswrapper[4612]: E0227 07:50:39.538201 4612 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 27 07:50:39 crc kubenswrapper[4612]: E0227 07:50:39.538208 4612 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 27 07:50:39 crc kubenswrapper[4612]: E0227 07:50:39.538211 4612 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 27 07:50:39 crc kubenswrapper[4612]: E0227 07:50:39.538275 4612 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-27 07:50:43.538247361 +0000 UTC m=+101.392177359 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 27 07:50:39 crc kubenswrapper[4612]: E0227 07:50:39.538304 4612 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-27 07:50:43.538293802 +0000 UTC m=+101.392223800 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 27 07:50:39 crc kubenswrapper[4612]: E0227 07:50:39.538543 4612 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 27 07:50:39 crc kubenswrapper[4612]: E0227 07:50:39.538672 4612 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-27 07:50:43.538659092 +0000 UTC m=+101.392589090 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 27 07:50:39 crc kubenswrapper[4612]: I0227 07:50:39.561327 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 07:50:39 crc kubenswrapper[4612]: I0227 07:50:39.561352 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 07:50:39 crc kubenswrapper[4612]: I0227 07:50:39.561360 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 07:50:39 crc kubenswrapper[4612]: I0227 07:50:39.561374 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 07:50:39 crc kubenswrapper[4612]: I0227 07:50:39.561383 4612 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T07:50:39Z","lastTransitionTime":"2026-02-27T07:50:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 07:50:39 crc kubenswrapper[4612]: I0227 07:50:39.663292 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 07:50:39 crc kubenswrapper[4612]: I0227 07:50:39.663334 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 07:50:39 crc kubenswrapper[4612]: I0227 07:50:39.663347 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 07:50:39 crc kubenswrapper[4612]: I0227 07:50:39.663364 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 07:50:39 crc kubenswrapper[4612]: I0227 07:50:39.663377 4612 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T07:50:39Z","lastTransitionTime":"2026-02-27T07:50:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 07:50:39 crc kubenswrapper[4612]: I0227 07:50:39.765748 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 07:50:39 crc kubenswrapper[4612]: I0227 07:50:39.765786 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 07:50:39 crc kubenswrapper[4612]: I0227 07:50:39.765797 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 07:50:39 crc kubenswrapper[4612]: I0227 07:50:39.765813 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 07:50:39 crc kubenswrapper[4612]: I0227 07:50:39.765824 4612 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T07:50:39Z","lastTransitionTime":"2026-02-27T07:50:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 07:50:39 crc kubenswrapper[4612]: I0227 07:50:39.852111 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 07:50:39 crc kubenswrapper[4612]: I0227 07:50:39.852141 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 07:50:39 crc kubenswrapper[4612]: E0227 07:50:39.852217 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 07:50:39 crc kubenswrapper[4612]: E0227 07:50:39.852808 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 07:50:39 crc kubenswrapper[4612]: I0227 07:50:39.853028 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 07:50:39 crc kubenswrapper[4612]: E0227 07:50:39.853323 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 07:50:39 crc kubenswrapper[4612]: I0227 07:50:39.868253 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 07:50:39 crc kubenswrapper[4612]: I0227 07:50:39.868324 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 07:50:39 crc kubenswrapper[4612]: I0227 07:50:39.868345 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 07:50:39 crc kubenswrapper[4612]: I0227 07:50:39.868372 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 07:50:39 crc kubenswrapper[4612]: I0227 07:50:39.868393 4612 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T07:50:39Z","lastTransitionTime":"2026-02-27T07:50:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 07:50:39 crc kubenswrapper[4612]: I0227 07:50:39.970533 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 07:50:39 crc kubenswrapper[4612]: I0227 07:50:39.970572 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 07:50:39 crc kubenswrapper[4612]: I0227 07:50:39.970796 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 07:50:39 crc kubenswrapper[4612]: I0227 07:50:39.970815 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 07:50:39 crc kubenswrapper[4612]: I0227 07:50:39.970825 4612 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T07:50:39Z","lastTransitionTime":"2026-02-27T07:50:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 07:50:40 crc kubenswrapper[4612]: I0227 07:50:40.073417 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 07:50:40 crc kubenswrapper[4612]: I0227 07:50:40.073619 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 07:50:40 crc kubenswrapper[4612]: I0227 07:50:40.073733 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 07:50:40 crc kubenswrapper[4612]: I0227 07:50:40.073807 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 07:50:40 crc kubenswrapper[4612]: I0227 07:50:40.073868 4612 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T07:50:40Z","lastTransitionTime":"2026-02-27T07:50:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 07:50:40 crc kubenswrapper[4612]: I0227 07:50:40.176257 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 07:50:40 crc kubenswrapper[4612]: I0227 07:50:40.176309 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 07:50:40 crc kubenswrapper[4612]: I0227 07:50:40.176321 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 07:50:40 crc kubenswrapper[4612]: I0227 07:50:40.176335 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 07:50:40 crc kubenswrapper[4612]: I0227 07:50:40.176346 4612 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T07:50:40Z","lastTransitionTime":"2026-02-27T07:50:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 07:50:40 crc kubenswrapper[4612]: I0227 07:50:40.278736 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 07:50:40 crc kubenswrapper[4612]: I0227 07:50:40.278970 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 07:50:40 crc kubenswrapper[4612]: I0227 07:50:40.279132 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 07:50:40 crc kubenswrapper[4612]: I0227 07:50:40.279278 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 07:50:40 crc kubenswrapper[4612]: I0227 07:50:40.279391 4612 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T07:50:40Z","lastTransitionTime":"2026-02-27T07:50:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 07:50:40 crc kubenswrapper[4612]: I0227 07:50:40.381662 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 07:50:40 crc kubenswrapper[4612]: I0227 07:50:40.381927 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 07:50:40 crc kubenswrapper[4612]: I0227 07:50:40.382025 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 07:50:40 crc kubenswrapper[4612]: I0227 07:50:40.382144 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 07:50:40 crc kubenswrapper[4612]: I0227 07:50:40.382221 4612 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T07:50:40Z","lastTransitionTime":"2026-02-27T07:50:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 07:50:40 crc kubenswrapper[4612]: I0227 07:50:40.484102 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 07:50:40 crc kubenswrapper[4612]: I0227 07:50:40.484373 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 07:50:40 crc kubenswrapper[4612]: I0227 07:50:40.484466 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 07:50:40 crc kubenswrapper[4612]: I0227 07:50:40.484562 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 07:50:40 crc kubenswrapper[4612]: I0227 07:50:40.484649 4612 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T07:50:40Z","lastTransitionTime":"2026-02-27T07:50:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 07:50:40 crc kubenswrapper[4612]: I0227 07:50:40.587808 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 07:50:40 crc kubenswrapper[4612]: I0227 07:50:40.587859 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 07:50:40 crc kubenswrapper[4612]: I0227 07:50:40.587871 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 07:50:40 crc kubenswrapper[4612]: I0227 07:50:40.587888 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 07:50:40 crc kubenswrapper[4612]: I0227 07:50:40.587900 4612 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T07:50:40Z","lastTransitionTime":"2026-02-27T07:50:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 07:50:40 crc kubenswrapper[4612]: I0227 07:50:40.690534 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 07:50:40 crc kubenswrapper[4612]: I0227 07:50:40.690571 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 07:50:40 crc kubenswrapper[4612]: I0227 07:50:40.690581 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 07:50:40 crc kubenswrapper[4612]: I0227 07:50:40.690598 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 07:50:40 crc kubenswrapper[4612]: I0227 07:50:40.690611 4612 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T07:50:40Z","lastTransitionTime":"2026-02-27T07:50:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 07:50:40 crc kubenswrapper[4612]: I0227 07:50:40.793034 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 07:50:40 crc kubenswrapper[4612]: I0227 07:50:40.793080 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 07:50:40 crc kubenswrapper[4612]: I0227 07:50:40.793092 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 07:50:40 crc kubenswrapper[4612]: I0227 07:50:40.793107 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 07:50:40 crc kubenswrapper[4612]: I0227 07:50:40.793118 4612 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T07:50:40Z","lastTransitionTime":"2026-02-27T07:50:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 07:50:40 crc kubenswrapper[4612]: I0227 07:50:40.900517 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 07:50:40 crc kubenswrapper[4612]: I0227 07:50:40.900567 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 07:50:40 crc kubenswrapper[4612]: I0227 07:50:40.900581 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 07:50:40 crc kubenswrapper[4612]: I0227 07:50:40.900600 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 07:50:40 crc kubenswrapper[4612]: I0227 07:50:40.900612 4612 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T07:50:40Z","lastTransitionTime":"2026-02-27T07:50:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 07:50:41 crc kubenswrapper[4612]: I0227 07:50:41.003233 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 07:50:41 crc kubenswrapper[4612]: I0227 07:50:41.003281 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 07:50:41 crc kubenswrapper[4612]: I0227 07:50:41.003297 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 07:50:41 crc kubenswrapper[4612]: I0227 07:50:41.003317 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 07:50:41 crc kubenswrapper[4612]: I0227 07:50:41.003331 4612 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T07:50:41Z","lastTransitionTime":"2026-02-27T07:50:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 07:50:41 crc kubenswrapper[4612]: I0227 07:50:41.106100 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 07:50:41 crc kubenswrapper[4612]: I0227 07:50:41.106138 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 07:50:41 crc kubenswrapper[4612]: I0227 07:50:41.106152 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 07:50:41 crc kubenswrapper[4612]: I0227 07:50:41.106169 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 07:50:41 crc kubenswrapper[4612]: I0227 07:50:41.106182 4612 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T07:50:41Z","lastTransitionTime":"2026-02-27T07:50:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 07:50:41 crc kubenswrapper[4612]: I0227 07:50:41.209153 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 07:50:41 crc kubenswrapper[4612]: I0227 07:50:41.209193 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 07:50:41 crc kubenswrapper[4612]: I0227 07:50:41.209204 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 07:50:41 crc kubenswrapper[4612]: I0227 07:50:41.209220 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 07:50:41 crc kubenswrapper[4612]: I0227 07:50:41.209231 4612 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T07:50:41Z","lastTransitionTime":"2026-02-27T07:50:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 07:50:41 crc kubenswrapper[4612]: I0227 07:50:41.312163 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 07:50:41 crc kubenswrapper[4612]: I0227 07:50:41.312440 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 07:50:41 crc kubenswrapper[4612]: I0227 07:50:41.312521 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 07:50:41 crc kubenswrapper[4612]: I0227 07:50:41.312630 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 07:50:41 crc kubenswrapper[4612]: I0227 07:50:41.312731 4612 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T07:50:41Z","lastTransitionTime":"2026-02-27T07:50:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 07:50:41 crc kubenswrapper[4612]: I0227 07:50:41.415084 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 07:50:41 crc kubenswrapper[4612]: I0227 07:50:41.415145 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 07:50:41 crc kubenswrapper[4612]: I0227 07:50:41.415162 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 07:50:41 crc kubenswrapper[4612]: I0227 07:50:41.415188 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 07:50:41 crc kubenswrapper[4612]: I0227 07:50:41.415204 4612 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T07:50:41Z","lastTransitionTime":"2026-02-27T07:50:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 07:50:41 crc kubenswrapper[4612]: I0227 07:50:41.517486 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 07:50:41 crc kubenswrapper[4612]: I0227 07:50:41.517833 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 07:50:41 crc kubenswrapper[4612]: I0227 07:50:41.517979 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 07:50:41 crc kubenswrapper[4612]: I0227 07:50:41.518124 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 07:50:41 crc kubenswrapper[4612]: I0227 07:50:41.518251 4612 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T07:50:41Z","lastTransitionTime":"2026-02-27T07:50:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 07:50:41 crc kubenswrapper[4612]: I0227 07:50:41.519753 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 27 07:50:41 crc kubenswrapper[4612]: I0227 07:50:41.519814 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 27 07:50:41 crc kubenswrapper[4612]: I0227 07:50:41.519836 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 27 07:50:41 crc kubenswrapper[4612]: I0227 07:50:41.519860 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 27 07:50:41 crc kubenswrapper[4612]: I0227 07:50:41.519877 4612 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-27T07:50:41Z","lastTransitionTime":"2026-02-27T07:50:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 27 07:50:41 crc kubenswrapper[4612]: I0227 07:50:41.837966 4612 certificate_manager.go:356] kubernetes.io/kubelet-serving: Rotating certificates Feb 27 07:50:41 crc kubenswrapper[4612]: I0227 07:50:41.845225 4612 reflector.go:368] Caches populated for *v1.CertificateSigningRequest from k8s.io/client-go/tools/watch/informerwatcher.go:146 Feb 27 07:50:41 crc kubenswrapper[4612]: I0227 07:50:41.852636 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 07:50:41 crc kubenswrapper[4612]: I0227 07:50:41.852646 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 07:50:41 crc kubenswrapper[4612]: I0227 07:50:41.852745 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 07:50:41 crc kubenswrapper[4612]: E0227 07:50:41.853050 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 07:50:41 crc kubenswrapper[4612]: E0227 07:50:41.852908 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 07:50:41 crc kubenswrapper[4612]: E0227 07:50:41.853144 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 07:50:43 crc kubenswrapper[4612]: I0227 07:50:43.469285 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 07:50:43 crc kubenswrapper[4612]: E0227 07:50:43.469444 4612 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 07:50:51.46942065 +0000 UTC m=+109.323350648 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 07:50:43 crc kubenswrapper[4612]: I0227 07:50:43.569741 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 07:50:43 crc kubenswrapper[4612]: I0227 07:50:43.569771 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 07:50:43 crc kubenswrapper[4612]: I0227 07:50:43.569793 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 07:50:43 crc kubenswrapper[4612]: I0227 07:50:43.569817 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 07:50:43 crc kubenswrapper[4612]: E0227 07:50:43.569992 4612 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 27 07:50:43 crc kubenswrapper[4612]: E0227 07:50:43.569998 4612 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 27 07:50:43 crc kubenswrapper[4612]: E0227 07:50:43.570126 4612 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-27 07:50:51.570100445 +0000 UTC m=+109.424030483 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 27 07:50:43 crc kubenswrapper[4612]: E0227 07:50:43.570005 4612 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 27 07:50:43 crc kubenswrapper[4612]: E0227 07:50:43.570205 4612 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 27 07:50:43 crc kubenswrapper[4612]: E0227 07:50:43.570233 4612 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 27 07:50:43 crc kubenswrapper[4612]: E0227 07:50:43.570028 4612 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 27 07:50:43 crc kubenswrapper[4612]: E0227 07:50:43.570307 4612 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-27 07:50:51.57028162 +0000 UTC m=+109.424211658 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 27 07:50:43 crc kubenswrapper[4612]: E0227 07:50:43.570314 4612 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 27 07:50:43 crc kubenswrapper[4612]: E0227 07:50:43.570358 4612 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-27 07:50:51.570347171 +0000 UTC m=+109.424277249 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 27 07:50:43 crc kubenswrapper[4612]: E0227 07:50:43.570591 4612 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 27 07:50:43 crc kubenswrapper[4612]: E0227 07:50:43.570629 4612 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-27 07:50:51.570618779 +0000 UTC m=+109.424548767 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 27 07:50:43 crc kubenswrapper[4612]: I0227 07:50:43.852773 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 07:50:43 crc kubenswrapper[4612]: I0227 07:50:43.852773 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 07:50:43 crc kubenswrapper[4612]: I0227 07:50:43.852791 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 07:50:43 crc kubenswrapper[4612]: E0227 07:50:43.853111 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 07:50:43 crc kubenswrapper[4612]: E0227 07:50:43.853289 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 07:50:43 crc kubenswrapper[4612]: E0227 07:50:43.853190 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.675302 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-hhrgk"] Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.675756 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-hhrgk" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.678323 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.678887 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.682398 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.703109 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-924vb"] Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.703810 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-924vb" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.706364 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.706960 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-ljx5w"] Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.707891 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-ljx5w" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.708031 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.709425 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.709471 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.712276 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.715336 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.718065 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.718129 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.718262 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.720680 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.728392 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-4hcsd"] Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.728829 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-4hcsd" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.731321 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.736004 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.741741 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-hb4dm"] Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.742596 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-hb4dm" Feb 27 07:50:45 crc kubenswrapper[4612]: W0227 07:50:45.745407 4612 reflector.go:561] object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl": failed to list *v1.Secret: secrets "ovn-kubernetes-node-dockercfg-pwtwl" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-ovn-kubernetes": no relationship found between node 'crc' and this object Feb 27 07:50:45 crc kubenswrapper[4612]: E0227 07:50:45.745446 4612 reflector.go:158] "Unhandled Error" err="object-\"openshift-ovn-kubernetes\"/\"ovn-kubernetes-node-dockercfg-pwtwl\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"ovn-kubernetes-node-dockercfg-pwtwl\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-ovn-kubernetes\": no relationship found between node 'crc' and this object" logger="UnhandledError" Feb 27 07:50:45 crc kubenswrapper[4612]: W0227 07:50:45.749939 4612 reflector.go:561] object-"openshift-ovn-kubernetes"/"kube-root-ca.crt": failed to list *v1.ConfigMap: configmaps "kube-root-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-ovn-kubernetes": no relationship found between node 'crc' and this object Feb 27 07:50:45 crc kubenswrapper[4612]: E0227 07:50:45.749990 4612 reflector.go:158] "Unhandled Error" err="object-\"openshift-ovn-kubernetes\"/\"kube-root-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"kube-root-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-ovn-kubernetes\": no relationship found between node 'crc' and this object" logger="UnhandledError" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.750055 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.750239 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.750424 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.750625 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.750810 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.788112 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/457f0d99-2681-403b-abc2-92af86fa76e5-systemd-units\") pod \"ovnkube-node-hb4dm\" (UID: \"457f0d99-2681-403b-abc2-92af86fa76e5\") " pod="openshift-ovn-kubernetes/ovnkube-node-hb4dm" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.788156 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/5ec92260-2494-43dc-8d39-0c554e8e161b-system-cni-dir\") pod \"multus-4hcsd\" (UID: \"5ec92260-2494-43dc-8d39-0c554e8e161b\") " pod="openshift-multus/multus-4hcsd" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.788190 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/457f0d99-2681-403b-abc2-92af86fa76e5-ovnkube-config\") pod \"ovnkube-node-hb4dm\" (UID: \"457f0d99-2681-403b-abc2-92af86fa76e5\") " pod="openshift-ovn-kubernetes/ovnkube-node-hb4dm" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.788216 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/5ec92260-2494-43dc-8d39-0c554e8e161b-host-var-lib-cni-multus\") pod \"multus-4hcsd\" (UID: \"5ec92260-2494-43dc-8d39-0c554e8e161b\") " pod="openshift-multus/multus-4hcsd" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.788241 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/5ec92260-2494-43dc-8d39-0c554e8e161b-etc-kubernetes\") pod \"multus-4hcsd\" (UID: \"5ec92260-2494-43dc-8d39-0c554e8e161b\") " pod="openshift-multus/multus-4hcsd" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.788261 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/24599373-7adc-4a1b-8bb4-797bf726f43d-mcd-auth-proxy-config\") pod \"machine-config-daemon-924vb\" (UID: \"24599373-7adc-4a1b-8bb4-797bf726f43d\") " pod="openshift-machine-config-operator/machine-config-daemon-924vb" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.788284 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/457f0d99-2681-403b-abc2-92af86fa76e5-ovn-node-metrics-cert\") pod \"ovnkube-node-hb4dm\" (UID: \"457f0d99-2681-403b-abc2-92af86fa76e5\") " pod="openshift-ovn-kubernetes/ovnkube-node-hb4dm" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.788305 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/5ec92260-2494-43dc-8d39-0c554e8e161b-cni-binary-copy\") pod \"multus-4hcsd\" (UID: \"5ec92260-2494-43dc-8d39-0c554e8e161b\") " pod="openshift-multus/multus-4hcsd" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.788323 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/ac122c97-bee5-4048-ac0b-2b9e56a194ff-system-cni-dir\") pod \"multus-additional-cni-plugins-ljx5w\" (UID: \"ac122c97-bee5-4048-ac0b-2b9e56a194ff\") " pod="openshift-multus/multus-additional-cni-plugins-ljx5w" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.788342 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/457f0d99-2681-403b-abc2-92af86fa76e5-host-cni-bin\") pod \"ovnkube-node-hb4dm\" (UID: \"457f0d99-2681-403b-abc2-92af86fa76e5\") " pod="openshift-ovn-kubernetes/ovnkube-node-hb4dm" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.788360 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5zllp\" (UniqueName: \"kubernetes.io/projected/457f0d99-2681-403b-abc2-92af86fa76e5-kube-api-access-5zllp\") pod \"ovnkube-node-hb4dm\" (UID: \"457f0d99-2681-403b-abc2-92af86fa76e5\") " pod="openshift-ovn-kubernetes/ovnkube-node-hb4dm" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.788379 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/5ec92260-2494-43dc-8d39-0c554e8e161b-host-var-lib-cni-bin\") pod \"multus-4hcsd\" (UID: \"5ec92260-2494-43dc-8d39-0c554e8e161b\") " pod="openshift-multus/multus-4hcsd" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.788412 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/457f0d99-2681-403b-abc2-92af86fa76e5-log-socket\") pod \"ovnkube-node-hb4dm\" (UID: \"457f0d99-2681-403b-abc2-92af86fa76e5\") " pod="openshift-ovn-kubernetes/ovnkube-node-hb4dm" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.788431 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/5ec92260-2494-43dc-8d39-0c554e8e161b-host-var-lib-kubelet\") pod \"multus-4hcsd\" (UID: \"5ec92260-2494-43dc-8d39-0c554e8e161b\") " pod="openshift-multus/multus-4hcsd" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.788448 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/5ec92260-2494-43dc-8d39-0c554e8e161b-multus-conf-dir\") pod \"multus-4hcsd\" (UID: \"5ec92260-2494-43dc-8d39-0c554e8e161b\") " pod="openshift-multus/multus-4hcsd" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.788481 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l9ccl\" (UniqueName: \"kubernetes.io/projected/ac122c97-bee5-4048-ac0b-2b9e56a194ff-kube-api-access-l9ccl\") pod \"multus-additional-cni-plugins-ljx5w\" (UID: \"ac122c97-bee5-4048-ac0b-2b9e56a194ff\") " pod="openshift-multus/multus-additional-cni-plugins-ljx5w" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.788504 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4zwfg\" (UniqueName: \"kubernetes.io/projected/24599373-7adc-4a1b-8bb4-797bf726f43d-kube-api-access-4zwfg\") pod \"machine-config-daemon-924vb\" (UID: \"24599373-7adc-4a1b-8bb4-797bf726f43d\") " pod="openshift-machine-config-operator/machine-config-daemon-924vb" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.788532 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/457f0d99-2681-403b-abc2-92af86fa76e5-run-openvswitch\") pod \"ovnkube-node-hb4dm\" (UID: \"457f0d99-2681-403b-abc2-92af86fa76e5\") " pod="openshift-ovn-kubernetes/ovnkube-node-hb4dm" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.788553 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/457f0d99-2681-403b-abc2-92af86fa76e5-env-overrides\") pod \"ovnkube-node-hb4dm\" (UID: \"457f0d99-2681-403b-abc2-92af86fa76e5\") " pod="openshift-ovn-kubernetes/ovnkube-node-hb4dm" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.788572 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/24599373-7adc-4a1b-8bb4-797bf726f43d-rootfs\") pod \"machine-config-daemon-924vb\" (UID: \"24599373-7adc-4a1b-8bb4-797bf726f43d\") " pod="openshift-machine-config-operator/machine-config-daemon-924vb" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.788593 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/457f0d99-2681-403b-abc2-92af86fa76e5-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-hb4dm\" (UID: \"457f0d99-2681-403b-abc2-92af86fa76e5\") " pod="openshift-ovn-kubernetes/ovnkube-node-hb4dm" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.788615 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/5ec92260-2494-43dc-8d39-0c554e8e161b-multus-cni-dir\") pod \"multus-4hcsd\" (UID: \"5ec92260-2494-43dc-8d39-0c554e8e161b\") " pod="openshift-multus/multus-4hcsd" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.788642 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/5ec92260-2494-43dc-8d39-0c554e8e161b-host-run-k8s-cni-cncf-io\") pod \"multus-4hcsd\" (UID: \"5ec92260-2494-43dc-8d39-0c554e8e161b\") " pod="openshift-multus/multus-4hcsd" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.788664 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/5ec92260-2494-43dc-8d39-0c554e8e161b-hostroot\") pod \"multus-4hcsd\" (UID: \"5ec92260-2494-43dc-8d39-0c554e8e161b\") " pod="openshift-multus/multus-4hcsd" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.788682 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/24599373-7adc-4a1b-8bb4-797bf726f43d-proxy-tls\") pod \"machine-config-daemon-924vb\" (UID: \"24599373-7adc-4a1b-8bb4-797bf726f43d\") " pod="openshift-machine-config-operator/machine-config-daemon-924vb" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.788755 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/ac122c97-bee5-4048-ac0b-2b9e56a194ff-cni-binary-copy\") pod \"multus-additional-cni-plugins-ljx5w\" (UID: \"ac122c97-bee5-4048-ac0b-2b9e56a194ff\") " pod="openshift-multus/multus-additional-cni-plugins-ljx5w" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.788809 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/ac122c97-bee5-4048-ac0b-2b9e56a194ff-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-ljx5w\" (UID: \"ac122c97-bee5-4048-ac0b-2b9e56a194ff\") " pod="openshift-multus/multus-additional-cni-plugins-ljx5w" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.788853 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/457f0d99-2681-403b-abc2-92af86fa76e5-run-ovn\") pod \"ovnkube-node-hb4dm\" (UID: \"457f0d99-2681-403b-abc2-92af86fa76e5\") " pod="openshift-ovn-kubernetes/ovnkube-node-hb4dm" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.788869 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/5ec92260-2494-43dc-8d39-0c554e8e161b-multus-daemon-config\") pod \"multus-4hcsd\" (UID: \"5ec92260-2494-43dc-8d39-0c554e8e161b\") " pod="openshift-multus/multus-4hcsd" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.788896 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/ac122c97-bee5-4048-ac0b-2b9e56a194ff-tuning-conf-dir\") pod \"multus-additional-cni-plugins-ljx5w\" (UID: \"ac122c97-bee5-4048-ac0b-2b9e56a194ff\") " pod="openshift-multus/multus-additional-cni-plugins-ljx5w" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.788911 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/457f0d99-2681-403b-abc2-92af86fa76e5-host-slash\") pod \"ovnkube-node-hb4dm\" (UID: \"457f0d99-2681-403b-abc2-92af86fa76e5\") " pod="openshift-ovn-kubernetes/ovnkube-node-hb4dm" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.788925 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/457f0d99-2681-403b-abc2-92af86fa76e5-run-systemd\") pod \"ovnkube-node-hb4dm\" (UID: \"457f0d99-2681-403b-abc2-92af86fa76e5\") " pod="openshift-ovn-kubernetes/ovnkube-node-hb4dm" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.788939 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/457f0d99-2681-403b-abc2-92af86fa76e5-host-run-ovn-kubernetes\") pod \"ovnkube-node-hb4dm\" (UID: \"457f0d99-2681-403b-abc2-92af86fa76e5\") " pod="openshift-ovn-kubernetes/ovnkube-node-hb4dm" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.788953 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/ac122c97-bee5-4048-ac0b-2b9e56a194ff-cnibin\") pod \"multus-additional-cni-plugins-ljx5w\" (UID: \"ac122c97-bee5-4048-ac0b-2b9e56a194ff\") " pod="openshift-multus/multus-additional-cni-plugins-ljx5w" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.788967 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/ac122c97-bee5-4048-ac0b-2b9e56a194ff-os-release\") pod \"multus-additional-cni-plugins-ljx5w\" (UID: \"ac122c97-bee5-4048-ac0b-2b9e56a194ff\") " pod="openshift-multus/multus-additional-cni-plugins-ljx5w" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.788986 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/457f0d99-2681-403b-abc2-92af86fa76e5-host-kubelet\") pod \"ovnkube-node-hb4dm\" (UID: \"457f0d99-2681-403b-abc2-92af86fa76e5\") " pod="openshift-ovn-kubernetes/ovnkube-node-hb4dm" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.789000 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/457f0d99-2681-403b-abc2-92af86fa76e5-var-lib-openvswitch\") pod \"ovnkube-node-hb4dm\" (UID: \"457f0d99-2681-403b-abc2-92af86fa76e5\") " pod="openshift-ovn-kubernetes/ovnkube-node-hb4dm" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.789015 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/457f0d99-2681-403b-abc2-92af86fa76e5-etc-openvswitch\") pod \"ovnkube-node-hb4dm\" (UID: \"457f0d99-2681-403b-abc2-92af86fa76e5\") " pod="openshift-ovn-kubernetes/ovnkube-node-hb4dm" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.789029 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/457f0d99-2681-403b-abc2-92af86fa76e5-ovnkube-script-lib\") pod \"ovnkube-node-hb4dm\" (UID: \"457f0d99-2681-403b-abc2-92af86fa76e5\") " pod="openshift-ovn-kubernetes/ovnkube-node-hb4dm" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.789043 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/457f0d99-2681-403b-abc2-92af86fa76e5-host-run-netns\") pod \"ovnkube-node-hb4dm\" (UID: \"457f0d99-2681-403b-abc2-92af86fa76e5\") " pod="openshift-ovn-kubernetes/ovnkube-node-hb4dm" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.789059 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/5ec92260-2494-43dc-8d39-0c554e8e161b-os-release\") pod \"multus-4hcsd\" (UID: \"5ec92260-2494-43dc-8d39-0c554e8e161b\") " pod="openshift-multus/multus-4hcsd" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.789075 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/5ec92260-2494-43dc-8d39-0c554e8e161b-multus-socket-dir-parent\") pod \"multus-4hcsd\" (UID: \"5ec92260-2494-43dc-8d39-0c554e8e161b\") " pod="openshift-multus/multus-4hcsd" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.789090 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4mwmm\" (UniqueName: \"kubernetes.io/projected/5ec92260-2494-43dc-8d39-0c554e8e161b-kube-api-access-4mwmm\") pod \"multus-4hcsd\" (UID: \"5ec92260-2494-43dc-8d39-0c554e8e161b\") " pod="openshift-multus/multus-4hcsd" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.789105 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dw7nn\" (UniqueName: \"kubernetes.io/projected/c426e325-d171-48b4-a550-f81750107445-kube-api-access-dw7nn\") pod \"node-resolver-hhrgk\" (UID: \"c426e325-d171-48b4-a550-f81750107445\") " pod="openshift-dns/node-resolver-hhrgk" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.789120 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/457f0d99-2681-403b-abc2-92af86fa76e5-host-cni-netd\") pod \"ovnkube-node-hb4dm\" (UID: \"457f0d99-2681-403b-abc2-92af86fa76e5\") " pod="openshift-ovn-kubernetes/ovnkube-node-hb4dm" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.789133 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/5ec92260-2494-43dc-8d39-0c554e8e161b-cnibin\") pod \"multus-4hcsd\" (UID: \"5ec92260-2494-43dc-8d39-0c554e8e161b\") " pod="openshift-multus/multus-4hcsd" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.789147 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/5ec92260-2494-43dc-8d39-0c554e8e161b-host-run-netns\") pod \"multus-4hcsd\" (UID: \"5ec92260-2494-43dc-8d39-0c554e8e161b\") " pod="openshift-multus/multus-4hcsd" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.789160 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/5ec92260-2494-43dc-8d39-0c554e8e161b-host-run-multus-certs\") pod \"multus-4hcsd\" (UID: \"5ec92260-2494-43dc-8d39-0c554e8e161b\") " pod="openshift-multus/multus-4hcsd" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.789174 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/457f0d99-2681-403b-abc2-92af86fa76e5-node-log\") pod \"ovnkube-node-hb4dm\" (UID: \"457f0d99-2681-403b-abc2-92af86fa76e5\") " pod="openshift-ovn-kubernetes/ovnkube-node-hb4dm" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.789186 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/c426e325-d171-48b4-a550-f81750107445-hosts-file\") pod \"node-resolver-hhrgk\" (UID: \"c426e325-d171-48b4-a550-f81750107445\") " pod="openshift-dns/node-resolver-hhrgk" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.852020 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.852107 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.852020 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 07:50:45 crc kubenswrapper[4612]: E0227 07:50:45.852111 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 07:50:45 crc kubenswrapper[4612]: E0227 07:50:45.852234 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 07:50:45 crc kubenswrapper[4612]: E0227 07:50:45.852306 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.854223 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-8dcjk"] Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.854538 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8dcjk" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.855841 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.856161 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.856280 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.859771 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.890083 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4zwfg\" (UniqueName: \"kubernetes.io/projected/24599373-7adc-4a1b-8bb4-797bf726f43d-kube-api-access-4zwfg\") pod \"machine-config-daemon-924vb\" (UID: \"24599373-7adc-4a1b-8bb4-797bf726f43d\") " pod="openshift-machine-config-operator/machine-config-daemon-924vb" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.890121 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/457f0d99-2681-403b-abc2-92af86fa76e5-log-socket\") pod \"ovnkube-node-hb4dm\" (UID: \"457f0d99-2681-403b-abc2-92af86fa76e5\") " pod="openshift-ovn-kubernetes/ovnkube-node-hb4dm" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.890138 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/5ec92260-2494-43dc-8d39-0c554e8e161b-host-var-lib-kubelet\") pod \"multus-4hcsd\" (UID: \"5ec92260-2494-43dc-8d39-0c554e8e161b\") " pod="openshift-multus/multus-4hcsd" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.890164 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/5ec92260-2494-43dc-8d39-0c554e8e161b-multus-conf-dir\") pod \"multus-4hcsd\" (UID: \"5ec92260-2494-43dc-8d39-0c554e8e161b\") " pod="openshift-multus/multus-4hcsd" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.890182 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l9ccl\" (UniqueName: \"kubernetes.io/projected/ac122c97-bee5-4048-ac0b-2b9e56a194ff-kube-api-access-l9ccl\") pod \"multus-additional-cni-plugins-ljx5w\" (UID: \"ac122c97-bee5-4048-ac0b-2b9e56a194ff\") " pod="openshift-multus/multus-additional-cni-plugins-ljx5w" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.890205 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/457f0d99-2681-403b-abc2-92af86fa76e5-run-openvswitch\") pod \"ovnkube-node-hb4dm\" (UID: \"457f0d99-2681-403b-abc2-92af86fa76e5\") " pod="openshift-ovn-kubernetes/ovnkube-node-hb4dm" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.890220 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/457f0d99-2681-403b-abc2-92af86fa76e5-env-overrides\") pod \"ovnkube-node-hb4dm\" (UID: \"457f0d99-2681-403b-abc2-92af86fa76e5\") " pod="openshift-ovn-kubernetes/ovnkube-node-hb4dm" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.890234 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/24599373-7adc-4a1b-8bb4-797bf726f43d-rootfs\") pod \"machine-config-daemon-924vb\" (UID: \"24599373-7adc-4a1b-8bb4-797bf726f43d\") " pod="openshift-machine-config-operator/machine-config-daemon-924vb" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.890247 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/5ec92260-2494-43dc-8d39-0c554e8e161b-hostroot\") pod \"multus-4hcsd\" (UID: \"5ec92260-2494-43dc-8d39-0c554e8e161b\") " pod="openshift-multus/multus-4hcsd" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.890249 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/457f0d99-2681-403b-abc2-92af86fa76e5-log-socket\") pod \"ovnkube-node-hb4dm\" (UID: \"457f0d99-2681-403b-abc2-92af86fa76e5\") " pod="openshift-ovn-kubernetes/ovnkube-node-hb4dm" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.890278 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/457f0d99-2681-403b-abc2-92af86fa76e5-run-openvswitch\") pod \"ovnkube-node-hb4dm\" (UID: \"457f0d99-2681-403b-abc2-92af86fa76e5\") " pod="openshift-ovn-kubernetes/ovnkube-node-hb4dm" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.890291 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/24599373-7adc-4a1b-8bb4-797bf726f43d-rootfs\") pod \"machine-config-daemon-924vb\" (UID: \"24599373-7adc-4a1b-8bb4-797bf726f43d\") " pod="openshift-machine-config-operator/machine-config-daemon-924vb" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.890249 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/5ec92260-2494-43dc-8d39-0c554e8e161b-host-var-lib-kubelet\") pod \"multus-4hcsd\" (UID: \"5ec92260-2494-43dc-8d39-0c554e8e161b\") " pod="openshift-multus/multus-4hcsd" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.890260 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/24599373-7adc-4a1b-8bb4-797bf726f43d-proxy-tls\") pod \"machine-config-daemon-924vb\" (UID: \"24599373-7adc-4a1b-8bb4-797bf726f43d\") " pod="openshift-machine-config-operator/machine-config-daemon-924vb" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.890261 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/5ec92260-2494-43dc-8d39-0c554e8e161b-multus-conf-dir\") pod \"multus-4hcsd\" (UID: \"5ec92260-2494-43dc-8d39-0c554e8e161b\") " pod="openshift-multus/multus-4hcsd" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.890335 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/5ec92260-2494-43dc-8d39-0c554e8e161b-hostroot\") pod \"multus-4hcsd\" (UID: \"5ec92260-2494-43dc-8d39-0c554e8e161b\") " pod="openshift-multus/multus-4hcsd" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.890343 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/457f0d99-2681-403b-abc2-92af86fa76e5-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-hb4dm\" (UID: \"457f0d99-2681-403b-abc2-92af86fa76e5\") " pod="openshift-ovn-kubernetes/ovnkube-node-hb4dm" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.890446 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/457f0d99-2681-403b-abc2-92af86fa76e5-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-hb4dm\" (UID: \"457f0d99-2681-403b-abc2-92af86fa76e5\") " pod="openshift-ovn-kubernetes/ovnkube-node-hb4dm" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.890469 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/5ec92260-2494-43dc-8d39-0c554e8e161b-multus-cni-dir\") pod \"multus-4hcsd\" (UID: \"5ec92260-2494-43dc-8d39-0c554e8e161b\") " pod="openshift-multus/multus-4hcsd" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.890520 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/5ec92260-2494-43dc-8d39-0c554e8e161b-multus-cni-dir\") pod \"multus-4hcsd\" (UID: \"5ec92260-2494-43dc-8d39-0c554e8e161b\") " pod="openshift-multus/multus-4hcsd" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.890540 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/5ec92260-2494-43dc-8d39-0c554e8e161b-host-run-k8s-cni-cncf-io\") pod \"multus-4hcsd\" (UID: \"5ec92260-2494-43dc-8d39-0c554e8e161b\") " pod="openshift-multus/multus-4hcsd" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.890574 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/ac122c97-bee5-4048-ac0b-2b9e56a194ff-cni-binary-copy\") pod \"multus-additional-cni-plugins-ljx5w\" (UID: \"ac122c97-bee5-4048-ac0b-2b9e56a194ff\") " pod="openshift-multus/multus-additional-cni-plugins-ljx5w" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.890597 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/ac122c97-bee5-4048-ac0b-2b9e56a194ff-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-ljx5w\" (UID: \"ac122c97-bee5-4048-ac0b-2b9e56a194ff\") " pod="openshift-multus/multus-additional-cni-plugins-ljx5w" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.890609 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/5ec92260-2494-43dc-8d39-0c554e8e161b-host-run-k8s-cni-cncf-io\") pod \"multus-4hcsd\" (UID: \"5ec92260-2494-43dc-8d39-0c554e8e161b\") " pod="openshift-multus/multus-4hcsd" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.890633 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/457f0d99-2681-403b-abc2-92af86fa76e5-run-ovn\") pod \"ovnkube-node-hb4dm\" (UID: \"457f0d99-2681-403b-abc2-92af86fa76e5\") " pod="openshift-ovn-kubernetes/ovnkube-node-hb4dm" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.890656 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/5ec92260-2494-43dc-8d39-0c554e8e161b-multus-daemon-config\") pod \"multus-4hcsd\" (UID: \"5ec92260-2494-43dc-8d39-0c554e8e161b\") " pod="openshift-multus/multus-4hcsd" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.890679 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/ac122c97-bee5-4048-ac0b-2b9e56a194ff-tuning-conf-dir\") pod \"multus-additional-cni-plugins-ljx5w\" (UID: \"ac122c97-bee5-4048-ac0b-2b9e56a194ff\") " pod="openshift-multus/multus-additional-cni-plugins-ljx5w" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.890718 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/ac122c97-bee5-4048-ac0b-2b9e56a194ff-os-release\") pod \"multus-additional-cni-plugins-ljx5w\" (UID: \"ac122c97-bee5-4048-ac0b-2b9e56a194ff\") " pod="openshift-multus/multus-additional-cni-plugins-ljx5w" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.890743 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/457f0d99-2681-403b-abc2-92af86fa76e5-host-slash\") pod \"ovnkube-node-hb4dm\" (UID: \"457f0d99-2681-403b-abc2-92af86fa76e5\") " pod="openshift-ovn-kubernetes/ovnkube-node-hb4dm" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.890763 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/457f0d99-2681-403b-abc2-92af86fa76e5-run-systemd\") pod \"ovnkube-node-hb4dm\" (UID: \"457f0d99-2681-403b-abc2-92af86fa76e5\") " pod="openshift-ovn-kubernetes/ovnkube-node-hb4dm" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.890776 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/457f0d99-2681-403b-abc2-92af86fa76e5-env-overrides\") pod \"ovnkube-node-hb4dm\" (UID: \"457f0d99-2681-403b-abc2-92af86fa76e5\") " pod="openshift-ovn-kubernetes/ovnkube-node-hb4dm" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.890781 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/457f0d99-2681-403b-abc2-92af86fa76e5-host-run-ovn-kubernetes\") pod \"ovnkube-node-hb4dm\" (UID: \"457f0d99-2681-403b-abc2-92af86fa76e5\") " pod="openshift-ovn-kubernetes/ovnkube-node-hb4dm" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.890826 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/ac122c97-bee5-4048-ac0b-2b9e56a194ff-cnibin\") pod \"multus-additional-cni-plugins-ljx5w\" (UID: \"ac122c97-bee5-4048-ac0b-2b9e56a194ff\") " pod="openshift-multus/multus-additional-cni-plugins-ljx5w" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.890843 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/457f0d99-2681-403b-abc2-92af86fa76e5-host-kubelet\") pod \"ovnkube-node-hb4dm\" (UID: \"457f0d99-2681-403b-abc2-92af86fa76e5\") " pod="openshift-ovn-kubernetes/ovnkube-node-hb4dm" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.890858 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/457f0d99-2681-403b-abc2-92af86fa76e5-var-lib-openvswitch\") pod \"ovnkube-node-hb4dm\" (UID: \"457f0d99-2681-403b-abc2-92af86fa76e5\") " pod="openshift-ovn-kubernetes/ovnkube-node-hb4dm" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.890872 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/457f0d99-2681-403b-abc2-92af86fa76e5-etc-openvswitch\") pod \"ovnkube-node-hb4dm\" (UID: \"457f0d99-2681-403b-abc2-92af86fa76e5\") " pod="openshift-ovn-kubernetes/ovnkube-node-hb4dm" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.890885 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/457f0d99-2681-403b-abc2-92af86fa76e5-ovnkube-script-lib\") pod \"ovnkube-node-hb4dm\" (UID: \"457f0d99-2681-403b-abc2-92af86fa76e5\") " pod="openshift-ovn-kubernetes/ovnkube-node-hb4dm" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.890899 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/457f0d99-2681-403b-abc2-92af86fa76e5-host-run-netns\") pod \"ovnkube-node-hb4dm\" (UID: \"457f0d99-2681-403b-abc2-92af86fa76e5\") " pod="openshift-ovn-kubernetes/ovnkube-node-hb4dm" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.890916 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/5ec92260-2494-43dc-8d39-0c554e8e161b-os-release\") pod \"multus-4hcsd\" (UID: \"5ec92260-2494-43dc-8d39-0c554e8e161b\") " pod="openshift-multus/multus-4hcsd" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.890914 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/457f0d99-2681-403b-abc2-92af86fa76e5-host-kubelet\") pod \"ovnkube-node-hb4dm\" (UID: \"457f0d99-2681-403b-abc2-92af86fa76e5\") " pod="openshift-ovn-kubernetes/ovnkube-node-hb4dm" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.890930 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/5ec92260-2494-43dc-8d39-0c554e8e161b-multus-socket-dir-parent\") pod \"multus-4hcsd\" (UID: \"5ec92260-2494-43dc-8d39-0c554e8e161b\") " pod="openshift-multus/multus-4hcsd" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.890805 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/457f0d99-2681-403b-abc2-92af86fa76e5-host-run-ovn-kubernetes\") pod \"ovnkube-node-hb4dm\" (UID: \"457f0d99-2681-403b-abc2-92af86fa76e5\") " pod="openshift-ovn-kubernetes/ovnkube-node-hb4dm" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.890944 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/457f0d99-2681-403b-abc2-92af86fa76e5-host-cni-netd\") pod \"ovnkube-node-hb4dm\" (UID: \"457f0d99-2681-403b-abc2-92af86fa76e5\") " pod="openshift-ovn-kubernetes/ovnkube-node-hb4dm" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.890959 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4mwmm\" (UniqueName: \"kubernetes.io/projected/5ec92260-2494-43dc-8d39-0c554e8e161b-kube-api-access-4mwmm\") pod \"multus-4hcsd\" (UID: \"5ec92260-2494-43dc-8d39-0c554e8e161b\") " pod="openshift-multus/multus-4hcsd" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.890965 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/ac122c97-bee5-4048-ac0b-2b9e56a194ff-cnibin\") pod \"multus-additional-cni-plugins-ljx5w\" (UID: \"ac122c97-bee5-4048-ac0b-2b9e56a194ff\") " pod="openshift-multus/multus-additional-cni-plugins-ljx5w" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.890974 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dw7nn\" (UniqueName: \"kubernetes.io/projected/c426e325-d171-48b4-a550-f81750107445-kube-api-access-dw7nn\") pod \"node-resolver-hhrgk\" (UID: \"c426e325-d171-48b4-a550-f81750107445\") " pod="openshift-dns/node-resolver-hhrgk" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.890991 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/5ec92260-2494-43dc-8d39-0c554e8e161b-cnibin\") pod \"multus-4hcsd\" (UID: \"5ec92260-2494-43dc-8d39-0c554e8e161b\") " pod="openshift-multus/multus-4hcsd" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.890999 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/457f0d99-2681-403b-abc2-92af86fa76e5-host-run-netns\") pod \"ovnkube-node-hb4dm\" (UID: \"457f0d99-2681-403b-abc2-92af86fa76e5\") " pod="openshift-ovn-kubernetes/ovnkube-node-hb4dm" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.891006 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/5ec92260-2494-43dc-8d39-0c554e8e161b-host-run-netns\") pod \"multus-4hcsd\" (UID: \"5ec92260-2494-43dc-8d39-0c554e8e161b\") " pod="openshift-multus/multus-4hcsd" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.891022 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/5ec92260-2494-43dc-8d39-0c554e8e161b-host-run-multus-certs\") pod \"multus-4hcsd\" (UID: \"5ec92260-2494-43dc-8d39-0c554e8e161b\") " pod="openshift-multus/multus-4hcsd" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.891032 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/457f0d99-2681-403b-abc2-92af86fa76e5-var-lib-openvswitch\") pod \"ovnkube-node-hb4dm\" (UID: \"457f0d99-2681-403b-abc2-92af86fa76e5\") " pod="openshift-ovn-kubernetes/ovnkube-node-hb4dm" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.891039 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/457f0d99-2681-403b-abc2-92af86fa76e5-node-log\") pod \"ovnkube-node-hb4dm\" (UID: \"457f0d99-2681-403b-abc2-92af86fa76e5\") " pod="openshift-ovn-kubernetes/ovnkube-node-hb4dm" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.891063 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/457f0d99-2681-403b-abc2-92af86fa76e5-node-log\") pod \"ovnkube-node-hb4dm\" (UID: \"457f0d99-2681-403b-abc2-92af86fa76e5\") " pod="openshift-ovn-kubernetes/ovnkube-node-hb4dm" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.891071 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/c426e325-d171-48b4-a550-f81750107445-hosts-file\") pod \"node-resolver-hhrgk\" (UID: \"c426e325-d171-48b4-a550-f81750107445\") " pod="openshift-dns/node-resolver-hhrgk" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.891094 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/457f0d99-2681-403b-abc2-92af86fa76e5-systemd-units\") pod \"ovnkube-node-hb4dm\" (UID: \"457f0d99-2681-403b-abc2-92af86fa76e5\") " pod="openshift-ovn-kubernetes/ovnkube-node-hb4dm" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.891107 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/5ec92260-2494-43dc-8d39-0c554e8e161b-os-release\") pod \"multus-4hcsd\" (UID: \"5ec92260-2494-43dc-8d39-0c554e8e161b\") " pod="openshift-multus/multus-4hcsd" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.891114 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/5ec92260-2494-43dc-8d39-0c554e8e161b-system-cni-dir\") pod \"multus-4hcsd\" (UID: \"5ec92260-2494-43dc-8d39-0c554e8e161b\") " pod="openshift-multus/multus-4hcsd" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.891139 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/5ec92260-2494-43dc-8d39-0c554e8e161b-multus-socket-dir-parent\") pod \"multus-4hcsd\" (UID: \"5ec92260-2494-43dc-8d39-0c554e8e161b\") " pod="openshift-multus/multus-4hcsd" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.891148 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/457f0d99-2681-403b-abc2-92af86fa76e5-ovnkube-config\") pod \"ovnkube-node-hb4dm\" (UID: \"457f0d99-2681-403b-abc2-92af86fa76e5\") " pod="openshift-ovn-kubernetes/ovnkube-node-hb4dm" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.891172 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/5ec92260-2494-43dc-8d39-0c554e8e161b-host-var-lib-cni-multus\") pod \"multus-4hcsd\" (UID: \"5ec92260-2494-43dc-8d39-0c554e8e161b\") " pod="openshift-multus/multus-4hcsd" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.891194 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/5ec92260-2494-43dc-8d39-0c554e8e161b-etc-kubernetes\") pod \"multus-4hcsd\" (UID: \"5ec92260-2494-43dc-8d39-0c554e8e161b\") " pod="openshift-multus/multus-4hcsd" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.891218 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/24599373-7adc-4a1b-8bb4-797bf726f43d-mcd-auth-proxy-config\") pod \"machine-config-daemon-924vb\" (UID: \"24599373-7adc-4a1b-8bb4-797bf726f43d\") " pod="openshift-machine-config-operator/machine-config-daemon-924vb" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.891249 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/457f0d99-2681-403b-abc2-92af86fa76e5-ovn-node-metrics-cert\") pod \"ovnkube-node-hb4dm\" (UID: \"457f0d99-2681-403b-abc2-92af86fa76e5\") " pod="openshift-ovn-kubernetes/ovnkube-node-hb4dm" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.891270 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/5ec92260-2494-43dc-8d39-0c554e8e161b-cni-binary-copy\") pod \"multus-4hcsd\" (UID: \"5ec92260-2494-43dc-8d39-0c554e8e161b\") " pod="openshift-multus/multus-4hcsd" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.891276 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/5ec92260-2494-43dc-8d39-0c554e8e161b-cnibin\") pod \"multus-4hcsd\" (UID: \"5ec92260-2494-43dc-8d39-0c554e8e161b\") " pod="openshift-multus/multus-4hcsd" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.891293 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/ac122c97-bee5-4048-ac0b-2b9e56a194ff-system-cni-dir\") pod \"multus-additional-cni-plugins-ljx5w\" (UID: \"ac122c97-bee5-4048-ac0b-2b9e56a194ff\") " pod="openshift-multus/multus-additional-cni-plugins-ljx5w" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.891297 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/5ec92260-2494-43dc-8d39-0c554e8e161b-multus-daemon-config\") pod \"multus-4hcsd\" (UID: \"5ec92260-2494-43dc-8d39-0c554e8e161b\") " pod="openshift-multus/multus-4hcsd" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.891314 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/457f0d99-2681-403b-abc2-92af86fa76e5-host-cni-bin\") pod \"ovnkube-node-hb4dm\" (UID: \"457f0d99-2681-403b-abc2-92af86fa76e5\") " pod="openshift-ovn-kubernetes/ovnkube-node-hb4dm" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.891345 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5zllp\" (UniqueName: \"kubernetes.io/projected/457f0d99-2681-403b-abc2-92af86fa76e5-kube-api-access-5zllp\") pod \"ovnkube-node-hb4dm\" (UID: \"457f0d99-2681-403b-abc2-92af86fa76e5\") " pod="openshift-ovn-kubernetes/ovnkube-node-hb4dm" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.891348 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/ac122c97-bee5-4048-ac0b-2b9e56a194ff-cni-binary-copy\") pod \"multus-additional-cni-plugins-ljx5w\" (UID: \"ac122c97-bee5-4048-ac0b-2b9e56a194ff\") " pod="openshift-multus/multus-additional-cni-plugins-ljx5w" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.891368 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/5ec92260-2494-43dc-8d39-0c554e8e161b-host-var-lib-cni-bin\") pod \"multus-4hcsd\" (UID: \"5ec92260-2494-43dc-8d39-0c554e8e161b\") " pod="openshift-multus/multus-4hcsd" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.891398 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/5ec92260-2494-43dc-8d39-0c554e8e161b-host-run-netns\") pod \"multus-4hcsd\" (UID: \"5ec92260-2494-43dc-8d39-0c554e8e161b\") " pod="openshift-multus/multus-4hcsd" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.891410 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/5ec92260-2494-43dc-8d39-0c554e8e161b-host-run-multus-certs\") pod \"multus-4hcsd\" (UID: \"5ec92260-2494-43dc-8d39-0c554e8e161b\") " pod="openshift-multus/multus-4hcsd" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.891174 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/457f0d99-2681-403b-abc2-92af86fa76e5-host-cni-netd\") pod \"ovnkube-node-hb4dm\" (UID: \"457f0d99-2681-403b-abc2-92af86fa76e5\") " pod="openshift-ovn-kubernetes/ovnkube-node-hb4dm" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.891446 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/5ec92260-2494-43dc-8d39-0c554e8e161b-host-var-lib-cni-multus\") pod \"multus-4hcsd\" (UID: \"5ec92260-2494-43dc-8d39-0c554e8e161b\") " pod="openshift-multus/multus-4hcsd" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.891447 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/457f0d99-2681-403b-abc2-92af86fa76e5-systemd-units\") pod \"ovnkube-node-hb4dm\" (UID: \"457f0d99-2681-403b-abc2-92af86fa76e5\") " pod="openshift-ovn-kubernetes/ovnkube-node-hb4dm" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.891505 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/ac122c97-bee5-4048-ac0b-2b9e56a194ff-system-cni-dir\") pod \"multus-additional-cni-plugins-ljx5w\" (UID: \"ac122c97-bee5-4048-ac0b-2b9e56a194ff\") " pod="openshift-multus/multus-additional-cni-plugins-ljx5w" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.891512 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/5ec92260-2494-43dc-8d39-0c554e8e161b-etc-kubernetes\") pod \"multus-4hcsd\" (UID: \"5ec92260-2494-43dc-8d39-0c554e8e161b\") " pod="openshift-multus/multus-4hcsd" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.891535 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/457f0d99-2681-403b-abc2-92af86fa76e5-host-slash\") pod \"ovnkube-node-hb4dm\" (UID: \"457f0d99-2681-403b-abc2-92af86fa76e5\") " pod="openshift-ovn-kubernetes/ovnkube-node-hb4dm" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.891554 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/457f0d99-2681-403b-abc2-92af86fa76e5-etc-openvswitch\") pod \"ovnkube-node-hb4dm\" (UID: \"457f0d99-2681-403b-abc2-92af86fa76e5\") " pod="openshift-ovn-kubernetes/ovnkube-node-hb4dm" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.891628 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/c426e325-d171-48b4-a550-f81750107445-hosts-file\") pod \"node-resolver-hhrgk\" (UID: \"c426e325-d171-48b4-a550-f81750107445\") " pod="openshift-dns/node-resolver-hhrgk" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.891734 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/ac122c97-bee5-4048-ac0b-2b9e56a194ff-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-ljx5w\" (UID: \"ac122c97-bee5-4048-ac0b-2b9e56a194ff\") " pod="openshift-multus/multus-additional-cni-plugins-ljx5w" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.891780 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/ac122c97-bee5-4048-ac0b-2b9e56a194ff-os-release\") pod \"multus-additional-cni-plugins-ljx5w\" (UID: \"ac122c97-bee5-4048-ac0b-2b9e56a194ff\") " pod="openshift-multus/multus-additional-cni-plugins-ljx5w" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.891794 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/457f0d99-2681-403b-abc2-92af86fa76e5-run-ovn\") pod \"ovnkube-node-hb4dm\" (UID: \"457f0d99-2681-403b-abc2-92af86fa76e5\") " pod="openshift-ovn-kubernetes/ovnkube-node-hb4dm" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.891945 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/457f0d99-2681-403b-abc2-92af86fa76e5-ovnkube-config\") pod \"ovnkube-node-hb4dm\" (UID: \"457f0d99-2681-403b-abc2-92af86fa76e5\") " pod="openshift-ovn-kubernetes/ovnkube-node-hb4dm" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.891982 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/5ec92260-2494-43dc-8d39-0c554e8e161b-host-var-lib-cni-bin\") pod \"multus-4hcsd\" (UID: \"5ec92260-2494-43dc-8d39-0c554e8e161b\") " pod="openshift-multus/multus-4hcsd" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.892007 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/5ec92260-2494-43dc-8d39-0c554e8e161b-system-cni-dir\") pod \"multus-4hcsd\" (UID: \"5ec92260-2494-43dc-8d39-0c554e8e161b\") " pod="openshift-multus/multus-4hcsd" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.892028 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/457f0d99-2681-403b-abc2-92af86fa76e5-host-cni-bin\") pod \"ovnkube-node-hb4dm\" (UID: \"457f0d99-2681-403b-abc2-92af86fa76e5\") " pod="openshift-ovn-kubernetes/ovnkube-node-hb4dm" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.892048 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/457f0d99-2681-403b-abc2-92af86fa76e5-run-systemd\") pod \"ovnkube-node-hb4dm\" (UID: \"457f0d99-2681-403b-abc2-92af86fa76e5\") " pod="openshift-ovn-kubernetes/ovnkube-node-hb4dm" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.892182 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/24599373-7adc-4a1b-8bb4-797bf726f43d-mcd-auth-proxy-config\") pod \"machine-config-daemon-924vb\" (UID: \"24599373-7adc-4a1b-8bb4-797bf726f43d\") " pod="openshift-machine-config-operator/machine-config-daemon-924vb" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.892225 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/457f0d99-2681-403b-abc2-92af86fa76e5-ovnkube-script-lib\") pod \"ovnkube-node-hb4dm\" (UID: \"457f0d99-2681-403b-abc2-92af86fa76e5\") " pod="openshift-ovn-kubernetes/ovnkube-node-hb4dm" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.892251 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/5ec92260-2494-43dc-8d39-0c554e8e161b-cni-binary-copy\") pod \"multus-4hcsd\" (UID: \"5ec92260-2494-43dc-8d39-0c554e8e161b\") " pod="openshift-multus/multus-4hcsd" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.892274 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/ac122c97-bee5-4048-ac0b-2b9e56a194ff-tuning-conf-dir\") pod \"multus-additional-cni-plugins-ljx5w\" (UID: \"ac122c97-bee5-4048-ac0b-2b9e56a194ff\") " pod="openshift-multus/multus-additional-cni-plugins-ljx5w" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.900254 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/24599373-7adc-4a1b-8bb4-797bf726f43d-proxy-tls\") pod \"machine-config-daemon-924vb\" (UID: \"24599373-7adc-4a1b-8bb4-797bf726f43d\") " pod="openshift-machine-config-operator/machine-config-daemon-924vb" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.906110 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/457f0d99-2681-403b-abc2-92af86fa76e5-ovn-node-metrics-cert\") pod \"ovnkube-node-hb4dm\" (UID: \"457f0d99-2681-403b-abc2-92af86fa76e5\") " pod="openshift-ovn-kubernetes/ovnkube-node-hb4dm" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.911303 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l9ccl\" (UniqueName: \"kubernetes.io/projected/ac122c97-bee5-4048-ac0b-2b9e56a194ff-kube-api-access-l9ccl\") pod \"multus-additional-cni-plugins-ljx5w\" (UID: \"ac122c97-bee5-4048-ac0b-2b9e56a194ff\") " pod="openshift-multus/multus-additional-cni-plugins-ljx5w" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.913955 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4zwfg\" (UniqueName: \"kubernetes.io/projected/24599373-7adc-4a1b-8bb4-797bf726f43d-kube-api-access-4zwfg\") pod \"machine-config-daemon-924vb\" (UID: \"24599373-7adc-4a1b-8bb4-797bf726f43d\") " pod="openshift-machine-config-operator/machine-config-daemon-924vb" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.915112 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dw7nn\" (UniqueName: \"kubernetes.io/projected/c426e325-d171-48b4-a550-f81750107445-kube-api-access-dw7nn\") pod \"node-resolver-hhrgk\" (UID: \"c426e325-d171-48b4-a550-f81750107445\") " pod="openshift-dns/node-resolver-hhrgk" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.922095 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4mwmm\" (UniqueName: \"kubernetes.io/projected/5ec92260-2494-43dc-8d39-0c554e8e161b-kube-api-access-4mwmm\") pod \"multus-4hcsd\" (UID: \"5ec92260-2494-43dc-8d39-0c554e8e161b\") " pod="openshift-multus/multus-4hcsd" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.944554 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-bjg4l"] Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.944866 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-bjg4l" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.946393 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.946566 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.946642 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.947576 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.992229 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-hhrgk" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.992424 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/855829cc-9c0c-48bb-a54c-fdd892b9c3ee-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-8dcjk\" (UID: \"855829cc-9c0c-48bb-a54c-fdd892b9c3ee\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8dcjk" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.992458 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/855829cc-9c0c-48bb-a54c-fdd892b9c3ee-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-8dcjk\" (UID: \"855829cc-9c0c-48bb-a54c-fdd892b9c3ee\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8dcjk" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.992489 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/855829cc-9c0c-48bb-a54c-fdd892b9c3ee-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-8dcjk\" (UID: \"855829cc-9c0c-48bb-a54c-fdd892b9c3ee\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8dcjk" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.992512 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/5a408b67-07f6-4a9a-83ba-1dc8f99b0433-host\") pod \"node-ca-bjg4l\" (UID: \"5a408b67-07f6-4a9a-83ba-1dc8f99b0433\") " pod="openshift-image-registry/node-ca-bjg4l" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.992535 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/5a408b67-07f6-4a9a-83ba-1dc8f99b0433-serviceca\") pod \"node-ca-bjg4l\" (UID: \"5a408b67-07f6-4a9a-83ba-1dc8f99b0433\") " pod="openshift-image-registry/node-ca-bjg4l" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.992583 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gr6sr\" (UniqueName: \"kubernetes.io/projected/5a408b67-07f6-4a9a-83ba-1dc8f99b0433-kube-api-access-gr6sr\") pod \"node-ca-bjg4l\" (UID: \"5a408b67-07f6-4a9a-83ba-1dc8f99b0433\") " pod="openshift-image-registry/node-ca-bjg4l" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.992617 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/855829cc-9c0c-48bb-a54c-fdd892b9c3ee-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-8dcjk\" (UID: \"855829cc-9c0c-48bb-a54c-fdd892b9c3ee\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8dcjk" Feb 27 07:50:45 crc kubenswrapper[4612]: I0227 07:50:45.992640 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/855829cc-9c0c-48bb-a54c-fdd892b9c3ee-service-ca\") pod \"cluster-version-operator-5c965bbfc6-8dcjk\" (UID: \"855829cc-9c0c-48bb-a54c-fdd892b9c3ee\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8dcjk" Feb 27 07:50:46 crc kubenswrapper[4612]: I0227 07:50:46.026126 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-924vb" Feb 27 07:50:46 crc kubenswrapper[4612]: I0227 07:50:46.038890 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-ljx5w" Feb 27 07:50:46 crc kubenswrapper[4612]: I0227 07:50:46.045823 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-4hcsd" Feb 27 07:50:46 crc kubenswrapper[4612]: W0227 07:50:46.061402 4612 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podac122c97_bee5_4048_ac0b_2b9e56a194ff.slice/crio-f33d974a2f7199afc3798467093aede28f16cc425988fe97bdc347523ac62afc WatchSource:0}: Error finding container f33d974a2f7199afc3798467093aede28f16cc425988fe97bdc347523ac62afc: Status 404 returned error can't find the container with id f33d974a2f7199afc3798467093aede28f16cc425988fe97bdc347523ac62afc Feb 27 07:50:46 crc kubenswrapper[4612]: I0227 07:50:46.082457 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-74rg2"] Feb 27 07:50:46 crc kubenswrapper[4612]: I0227 07:50:46.083498 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-74rg2" Feb 27 07:50:46 crc kubenswrapper[4612]: I0227 07:50:46.085655 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Feb 27 07:50:46 crc kubenswrapper[4612]: I0227 07:50:46.085721 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Feb 27 07:50:46 crc kubenswrapper[4612]: I0227 07:50:46.096046 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/855829cc-9c0c-48bb-a54c-fdd892b9c3ee-service-ca\") pod \"cluster-version-operator-5c965bbfc6-8dcjk\" (UID: \"855829cc-9c0c-48bb-a54c-fdd892b9c3ee\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8dcjk" Feb 27 07:50:46 crc kubenswrapper[4612]: I0227 07:50:46.096091 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/855829cc-9c0c-48bb-a54c-fdd892b9c3ee-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-8dcjk\" (UID: \"855829cc-9c0c-48bb-a54c-fdd892b9c3ee\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8dcjk" Feb 27 07:50:46 crc kubenswrapper[4612]: I0227 07:50:46.096127 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/855829cc-9c0c-48bb-a54c-fdd892b9c3ee-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-8dcjk\" (UID: \"855829cc-9c0c-48bb-a54c-fdd892b9c3ee\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8dcjk" Feb 27 07:50:46 crc kubenswrapper[4612]: I0227 07:50:46.096158 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/855829cc-9c0c-48bb-a54c-fdd892b9c3ee-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-8dcjk\" (UID: \"855829cc-9c0c-48bb-a54c-fdd892b9c3ee\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8dcjk" Feb 27 07:50:46 crc kubenswrapper[4612]: I0227 07:50:46.096187 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/5a408b67-07f6-4a9a-83ba-1dc8f99b0433-host\") pod \"node-ca-bjg4l\" (UID: \"5a408b67-07f6-4a9a-83ba-1dc8f99b0433\") " pod="openshift-image-registry/node-ca-bjg4l" Feb 27 07:50:46 crc kubenswrapper[4612]: I0227 07:50:46.096225 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/5a408b67-07f6-4a9a-83ba-1dc8f99b0433-serviceca\") pod \"node-ca-bjg4l\" (UID: \"5a408b67-07f6-4a9a-83ba-1dc8f99b0433\") " pod="openshift-image-registry/node-ca-bjg4l" Feb 27 07:50:46 crc kubenswrapper[4612]: I0227 07:50:46.096247 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gr6sr\" (UniqueName: \"kubernetes.io/projected/5a408b67-07f6-4a9a-83ba-1dc8f99b0433-kube-api-access-gr6sr\") pod \"node-ca-bjg4l\" (UID: \"5a408b67-07f6-4a9a-83ba-1dc8f99b0433\") " pod="openshift-image-registry/node-ca-bjg4l" Feb 27 07:50:46 crc kubenswrapper[4612]: I0227 07:50:46.096267 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/855829cc-9c0c-48bb-a54c-fdd892b9c3ee-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-8dcjk\" (UID: \"855829cc-9c0c-48bb-a54c-fdd892b9c3ee\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8dcjk" Feb 27 07:50:46 crc kubenswrapper[4612]: I0227 07:50:46.096335 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/855829cc-9c0c-48bb-a54c-fdd892b9c3ee-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-8dcjk\" (UID: \"855829cc-9c0c-48bb-a54c-fdd892b9c3ee\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8dcjk" Feb 27 07:50:46 crc kubenswrapper[4612]: I0227 07:50:46.097022 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-bqh6n"] Feb 27 07:50:46 crc kubenswrapper[4612]: I0227 07:50:46.097113 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/855829cc-9c0c-48bb-a54c-fdd892b9c3ee-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-8dcjk\" (UID: \"855829cc-9c0c-48bb-a54c-fdd892b9c3ee\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8dcjk" Feb 27 07:50:46 crc kubenswrapper[4612]: I0227 07:50:46.097196 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/855829cc-9c0c-48bb-a54c-fdd892b9c3ee-service-ca\") pod \"cluster-version-operator-5c965bbfc6-8dcjk\" (UID: \"855829cc-9c0c-48bb-a54c-fdd892b9c3ee\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8dcjk" Feb 27 07:50:46 crc kubenswrapper[4612]: I0227 07:50:46.097334 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/5a408b67-07f6-4a9a-83ba-1dc8f99b0433-host\") pod \"node-ca-bjg4l\" (UID: \"5a408b67-07f6-4a9a-83ba-1dc8f99b0433\") " pod="openshift-image-registry/node-ca-bjg4l" Feb 27 07:50:46 crc kubenswrapper[4612]: I0227 07:50:46.097916 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/5a408b67-07f6-4a9a-83ba-1dc8f99b0433-serviceca\") pod \"node-ca-bjg4l\" (UID: \"5a408b67-07f6-4a9a-83ba-1dc8f99b0433\") " pod="openshift-image-registry/node-ca-bjg4l" Feb 27 07:50:46 crc kubenswrapper[4612]: I0227 07:50:46.097982 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-bqh6n" Feb 27 07:50:46 crc kubenswrapper[4612]: E0227 07:50:46.098100 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-bqh6n" podUID="9421f529-9b1e-4bd0-8032-7baa608981b4" Feb 27 07:50:46 crc kubenswrapper[4612]: I0227 07:50:46.100974 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/855829cc-9c0c-48bb-a54c-fdd892b9c3ee-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-8dcjk\" (UID: \"855829cc-9c0c-48bb-a54c-fdd892b9c3ee\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8dcjk" Feb 27 07:50:46 crc kubenswrapper[4612]: I0227 07:50:46.114346 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gr6sr\" (UniqueName: \"kubernetes.io/projected/5a408b67-07f6-4a9a-83ba-1dc8f99b0433-kube-api-access-gr6sr\") pod \"node-ca-bjg4l\" (UID: \"5a408b67-07f6-4a9a-83ba-1dc8f99b0433\") " pod="openshift-image-registry/node-ca-bjg4l" Feb 27 07:50:46 crc kubenswrapper[4612]: I0227 07:50:46.114431 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/855829cc-9c0c-48bb-a54c-fdd892b9c3ee-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-8dcjk\" (UID: \"855829cc-9c0c-48bb-a54c-fdd892b9c3ee\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8dcjk" Feb 27 07:50:46 crc kubenswrapper[4612]: I0227 07:50:46.164849 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8dcjk" Feb 27 07:50:46 crc kubenswrapper[4612]: W0227 07:50:46.176507 4612 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod855829cc_9c0c_48bb_a54c_fdd892b9c3ee.slice/crio-8208010c3c9fa73154198a9523c52c715eb2121940165cbaec1a49a57b7d158d WatchSource:0}: Error finding container 8208010c3c9fa73154198a9523c52c715eb2121940165cbaec1a49a57b7d158d: Status 404 returned error can't find the container with id 8208010c3c9fa73154198a9523c52c715eb2121940165cbaec1a49a57b7d158d Feb 27 07:50:46 crc kubenswrapper[4612]: I0227 07:50:46.196884 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8brqx\" (UniqueName: \"kubernetes.io/projected/3b90f13e-804f-4363-a67d-be2c0ac4d533-kube-api-access-8brqx\") pod \"ovnkube-control-plane-749d76644c-74rg2\" (UID: \"3b90f13e-804f-4363-a67d-be2c0ac4d533\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-74rg2" Feb 27 07:50:46 crc kubenswrapper[4612]: I0227 07:50:46.196927 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9421f529-9b1e-4bd0-8032-7baa608981b4-metrics-certs\") pod \"network-metrics-daemon-bqh6n\" (UID: \"9421f529-9b1e-4bd0-8032-7baa608981b4\") " pod="openshift-multus/network-metrics-daemon-bqh6n" Feb 27 07:50:46 crc kubenswrapper[4612]: I0227 07:50:46.196954 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d4gxh\" (UniqueName: \"kubernetes.io/projected/9421f529-9b1e-4bd0-8032-7baa608981b4-kube-api-access-d4gxh\") pod \"network-metrics-daemon-bqh6n\" (UID: \"9421f529-9b1e-4bd0-8032-7baa608981b4\") " pod="openshift-multus/network-metrics-daemon-bqh6n" Feb 27 07:50:46 crc kubenswrapper[4612]: I0227 07:50:46.196971 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/3b90f13e-804f-4363-a67d-be2c0ac4d533-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-74rg2\" (UID: \"3b90f13e-804f-4363-a67d-be2c0ac4d533\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-74rg2" Feb 27 07:50:46 crc kubenswrapper[4612]: I0227 07:50:46.197007 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/3b90f13e-804f-4363-a67d-be2c0ac4d533-env-overrides\") pod \"ovnkube-control-plane-749d76644c-74rg2\" (UID: \"3b90f13e-804f-4363-a67d-be2c0ac4d533\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-74rg2" Feb 27 07:50:46 crc kubenswrapper[4612]: I0227 07:50:46.197022 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/3b90f13e-804f-4363-a67d-be2c0ac4d533-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-74rg2\" (UID: \"3b90f13e-804f-4363-a67d-be2c0ac4d533\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-74rg2" Feb 27 07:50:46 crc kubenswrapper[4612]: I0227 07:50:46.254212 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-bjg4l" Feb 27 07:50:46 crc kubenswrapper[4612]: W0227 07:50:46.271442 4612 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5a408b67_07f6_4a9a_83ba_1dc8f99b0433.slice/crio-dcd706fbcdcb55132a8bbd75015dad5241359608f9e24e69c646bb64278527fa WatchSource:0}: Error finding container dcd706fbcdcb55132a8bbd75015dad5241359608f9e24e69c646bb64278527fa: Status 404 returned error can't find the container with id dcd706fbcdcb55132a8bbd75015dad5241359608f9e24e69c646bb64278527fa Feb 27 07:50:46 crc kubenswrapper[4612]: I0227 07:50:46.279830 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-4hcsd" event={"ID":"5ec92260-2494-43dc-8d39-0c554e8e161b","Type":"ContainerStarted","Data":"143ad013598715c75c27c4f0f6b62bb97d5f8c8999150abd8b94039eb7296215"} Feb 27 07:50:46 crc kubenswrapper[4612]: I0227 07:50:46.279874 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-4hcsd" event={"ID":"5ec92260-2494-43dc-8d39-0c554e8e161b","Type":"ContainerStarted","Data":"f5110982d8965e384b1ea56354076e104c7168835c301103b0519c5627d9c387"} Feb 27 07:50:46 crc kubenswrapper[4612]: I0227 07:50:46.283843 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-hhrgk" event={"ID":"c426e325-d171-48b4-a550-f81750107445","Type":"ContainerStarted","Data":"126fb4fc00441a2749fd742be860bcaf13c78f4c0a4f78db20fd5c4ad283b5bc"} Feb 27 07:50:46 crc kubenswrapper[4612]: I0227 07:50:46.283876 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-hhrgk" event={"ID":"c426e325-d171-48b4-a550-f81750107445","Type":"ContainerStarted","Data":"49e17aa6af7de9eca058a9137f5e46de2061f30841d0af0a06cca3d246d52eba"} Feb 27 07:50:46 crc kubenswrapper[4612]: I0227 07:50:46.287990 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8dcjk" event={"ID":"855829cc-9c0c-48bb-a54c-fdd892b9c3ee","Type":"ContainerStarted","Data":"942733d09c9986db1eb30a25f2da1fa2d3edca0e846b5f8806961da3b2d052ab"} Feb 27 07:50:46 crc kubenswrapper[4612]: I0227 07:50:46.288118 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8dcjk" event={"ID":"855829cc-9c0c-48bb-a54c-fdd892b9c3ee","Type":"ContainerStarted","Data":"8208010c3c9fa73154198a9523c52c715eb2121940165cbaec1a49a57b7d158d"} Feb 27 07:50:46 crc kubenswrapper[4612]: I0227 07:50:46.289452 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-ljx5w" event={"ID":"ac122c97-bee5-4048-ac0b-2b9e56a194ff","Type":"ContainerStarted","Data":"c0c59363369df4bec13d2f0225f963ca7f48c122ce6775cc55bb25f8d7c01379"} Feb 27 07:50:46 crc kubenswrapper[4612]: I0227 07:50:46.289483 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-ljx5w" event={"ID":"ac122c97-bee5-4048-ac0b-2b9e56a194ff","Type":"ContainerStarted","Data":"f33d974a2f7199afc3798467093aede28f16cc425988fe97bdc347523ac62afc"} Feb 27 07:50:46 crc kubenswrapper[4612]: I0227 07:50:46.291170 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-924vb" event={"ID":"24599373-7adc-4a1b-8bb4-797bf726f43d","Type":"ContainerStarted","Data":"e08c4319a6ac14ff0311b23ec5ebec0e3d86580a75d91a68cd88b705da5b85ef"} Feb 27 07:50:46 crc kubenswrapper[4612]: I0227 07:50:46.291198 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-924vb" event={"ID":"24599373-7adc-4a1b-8bb4-797bf726f43d","Type":"ContainerStarted","Data":"02aa743ba66d91783716a3af076f70d769d77dce3c83913aa953859928a3c90f"} Feb 27 07:50:46 crc kubenswrapper[4612]: I0227 07:50:46.291211 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-924vb" event={"ID":"24599373-7adc-4a1b-8bb4-797bf726f43d","Type":"ContainerStarted","Data":"02b0c6c731fe2ae017aa2896cedb6bf4c2e3e92d77d8c81fa3efd20d58532f56"} Feb 27 07:50:46 crc kubenswrapper[4612]: I0227 07:50:46.297355 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d4gxh\" (UniqueName: \"kubernetes.io/projected/9421f529-9b1e-4bd0-8032-7baa608981b4-kube-api-access-d4gxh\") pod \"network-metrics-daemon-bqh6n\" (UID: \"9421f529-9b1e-4bd0-8032-7baa608981b4\") " pod="openshift-multus/network-metrics-daemon-bqh6n" Feb 27 07:50:46 crc kubenswrapper[4612]: I0227 07:50:46.297387 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/3b90f13e-804f-4363-a67d-be2c0ac4d533-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-74rg2\" (UID: \"3b90f13e-804f-4363-a67d-be2c0ac4d533\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-74rg2" Feb 27 07:50:46 crc kubenswrapper[4612]: I0227 07:50:46.297410 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/3b90f13e-804f-4363-a67d-be2c0ac4d533-env-overrides\") pod \"ovnkube-control-plane-749d76644c-74rg2\" (UID: \"3b90f13e-804f-4363-a67d-be2c0ac4d533\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-74rg2" Feb 27 07:50:46 crc kubenswrapper[4612]: I0227 07:50:46.297426 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/3b90f13e-804f-4363-a67d-be2c0ac4d533-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-74rg2\" (UID: \"3b90f13e-804f-4363-a67d-be2c0ac4d533\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-74rg2" Feb 27 07:50:46 crc kubenswrapper[4612]: I0227 07:50:46.297475 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8brqx\" (UniqueName: \"kubernetes.io/projected/3b90f13e-804f-4363-a67d-be2c0ac4d533-kube-api-access-8brqx\") pod \"ovnkube-control-plane-749d76644c-74rg2\" (UID: \"3b90f13e-804f-4363-a67d-be2c0ac4d533\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-74rg2" Feb 27 07:50:46 crc kubenswrapper[4612]: I0227 07:50:46.297491 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9421f529-9b1e-4bd0-8032-7baa608981b4-metrics-certs\") pod \"network-metrics-daemon-bqh6n\" (UID: \"9421f529-9b1e-4bd0-8032-7baa608981b4\") " pod="openshift-multus/network-metrics-daemon-bqh6n" Feb 27 07:50:46 crc kubenswrapper[4612]: E0227 07:50:46.297581 4612 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 27 07:50:46 crc kubenswrapper[4612]: E0227 07:50:46.297617 4612 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9421f529-9b1e-4bd0-8032-7baa608981b4-metrics-certs podName:9421f529-9b1e-4bd0-8032-7baa608981b4 nodeName:}" failed. No retries permitted until 2026-02-27 07:50:46.797604675 +0000 UTC m=+104.651534673 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/9421f529-9b1e-4bd0-8032-7baa608981b4-metrics-certs") pod "network-metrics-daemon-bqh6n" (UID: "9421f529-9b1e-4bd0-8032-7baa608981b4") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 27 07:50:46 crc kubenswrapper[4612]: I0227 07:50:46.297908 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-4hcsd" podStartSLOduration=43.297889123 podStartE2EDuration="43.297889123s" podCreationTimestamp="2026-02-27 07:50:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 07:50:46.297376099 +0000 UTC m=+104.151306097" watchObservedRunningTime="2026-02-27 07:50:46.297889123 +0000 UTC m=+104.151819121" Feb 27 07:50:46 crc kubenswrapper[4612]: I0227 07:50:46.298161 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/3b90f13e-804f-4363-a67d-be2c0ac4d533-env-overrides\") pod \"ovnkube-control-plane-749d76644c-74rg2\" (UID: \"3b90f13e-804f-4363-a67d-be2c0ac4d533\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-74rg2" Feb 27 07:50:46 crc kubenswrapper[4612]: I0227 07:50:46.298854 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/3b90f13e-804f-4363-a67d-be2c0ac4d533-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-74rg2\" (UID: \"3b90f13e-804f-4363-a67d-be2c0ac4d533\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-74rg2" Feb 27 07:50:46 crc kubenswrapper[4612]: I0227 07:50:46.300227 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/3b90f13e-804f-4363-a67d-be2c0ac4d533-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-74rg2\" (UID: \"3b90f13e-804f-4363-a67d-be2c0ac4d533\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-74rg2" Feb 27 07:50:46 crc kubenswrapper[4612]: I0227 07:50:46.314612 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d4gxh\" (UniqueName: \"kubernetes.io/projected/9421f529-9b1e-4bd0-8032-7baa608981b4-kube-api-access-d4gxh\") pod \"network-metrics-daemon-bqh6n\" (UID: \"9421f529-9b1e-4bd0-8032-7baa608981b4\") " pod="openshift-multus/network-metrics-daemon-bqh6n" Feb 27 07:50:46 crc kubenswrapper[4612]: I0227 07:50:46.333083 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podStartSLOduration=43.333062652 podStartE2EDuration="43.333062652s" podCreationTimestamp="2026-02-27 07:50:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 07:50:46.310555182 +0000 UTC m=+104.164485190" watchObservedRunningTime="2026-02-27 07:50:46.333062652 +0000 UTC m=+104.186992640" Feb 27 07:50:46 crc kubenswrapper[4612]: I0227 07:50:46.343513 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-hhrgk" podStartSLOduration=43.34350034 podStartE2EDuration="43.34350034s" podCreationTimestamp="2026-02-27 07:50:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 07:50:46.342741149 +0000 UTC m=+104.196671147" watchObservedRunningTime="2026-02-27 07:50:46.34350034 +0000 UTC m=+104.197430338" Feb 27 07:50:46 crc kubenswrapper[4612]: I0227 07:50:46.803452 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9421f529-9b1e-4bd0-8032-7baa608981b4-metrics-certs\") pod \"network-metrics-daemon-bqh6n\" (UID: \"9421f529-9b1e-4bd0-8032-7baa608981b4\") " pod="openshift-multus/network-metrics-daemon-bqh6n" Feb 27 07:50:46 crc kubenswrapper[4612]: E0227 07:50:46.803631 4612 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 27 07:50:46 crc kubenswrapper[4612]: E0227 07:50:46.803727 4612 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9421f529-9b1e-4bd0-8032-7baa608981b4-metrics-certs podName:9421f529-9b1e-4bd0-8032-7baa608981b4 nodeName:}" failed. No retries permitted until 2026-02-27 07:50:47.803684221 +0000 UTC m=+105.657614239 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/9421f529-9b1e-4bd0-8032-7baa608981b4-metrics-certs") pod "network-metrics-daemon-bqh6n" (UID: "9421f529-9b1e-4bd0-8032-7baa608981b4") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 27 07:50:46 crc kubenswrapper[4612]: E0227 07:50:46.905328 4612 projected.go:288] Couldn't get configMap openshift-ovn-kubernetes/kube-root-ca.crt: failed to sync configmap cache: timed out waiting for the condition Feb 27 07:50:46 crc kubenswrapper[4612]: E0227 07:50:46.905392 4612 projected.go:194] Error preparing data for projected volume kube-api-access-5zllp for pod openshift-ovn-kubernetes/ovnkube-node-hb4dm: failed to sync configmap cache: timed out waiting for the condition Feb 27 07:50:46 crc kubenswrapper[4612]: E0227 07:50:46.905475 4612 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/457f0d99-2681-403b-abc2-92af86fa76e5-kube-api-access-5zllp podName:457f0d99-2681-403b-abc2-92af86fa76e5 nodeName:}" failed. No retries permitted until 2026-02-27 07:50:47.405449085 +0000 UTC m=+105.259379113 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-5zllp" (UniqueName: "kubernetes.io/projected/457f0d99-2681-403b-abc2-92af86fa76e5-kube-api-access-5zllp") pod "ovnkube-node-hb4dm" (UID: "457f0d99-2681-403b-abc2-92af86fa76e5") : failed to sync configmap cache: timed out waiting for the condition Feb 27 07:50:46 crc kubenswrapper[4612]: I0227 07:50:46.999197 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Feb 27 07:50:47 crc kubenswrapper[4612]: I0227 07:50:47.037502 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Feb 27 07:50:47 crc kubenswrapper[4612]: I0227 07:50:47.045962 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8brqx\" (UniqueName: \"kubernetes.io/projected/3b90f13e-804f-4363-a67d-be2c0ac4d533-kube-api-access-8brqx\") pod \"ovnkube-control-plane-749d76644c-74rg2\" (UID: \"3b90f13e-804f-4363-a67d-be2c0ac4d533\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-74rg2" Feb 27 07:50:47 crc kubenswrapper[4612]: I0227 07:50:47.296038 4612 generic.go:334] "Generic (PLEG): container finished" podID="ac122c97-bee5-4048-ac0b-2b9e56a194ff" containerID="c0c59363369df4bec13d2f0225f963ca7f48c122ce6775cc55bb25f8d7c01379" exitCode=0 Feb 27 07:50:47 crc kubenswrapper[4612]: I0227 07:50:47.296138 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-ljx5w" event={"ID":"ac122c97-bee5-4048-ac0b-2b9e56a194ff","Type":"ContainerDied","Data":"c0c59363369df4bec13d2f0225f963ca7f48c122ce6775cc55bb25f8d7c01379"} Feb 27 07:50:47 crc kubenswrapper[4612]: I0227 07:50:47.307058 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-bjg4l" event={"ID":"5a408b67-07f6-4a9a-83ba-1dc8f99b0433","Type":"ContainerStarted","Data":"d245755355c0c4d440540cbc426e9675a5edc77261325beb402f0fed8a21b6ba"} Feb 27 07:50:47 crc kubenswrapper[4612]: I0227 07:50:47.307120 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-bjg4l" event={"ID":"5a408b67-07f6-4a9a-83ba-1dc8f99b0433","Type":"ContainerStarted","Data":"dcd706fbcdcb55132a8bbd75015dad5241359608f9e24e69c646bb64278527fa"} Feb 27 07:50:47 crc kubenswrapper[4612]: I0227 07:50:47.313668 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-74rg2" Feb 27 07:50:47 crc kubenswrapper[4612]: W0227 07:50:47.345951 4612 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3b90f13e_804f_4363_a67d_be2c0ac4d533.slice/crio-71b03216f4dabb549b38a366842c6453cd8a254852c148639d49e7792324081a WatchSource:0}: Error finding container 71b03216f4dabb549b38a366842c6453cd8a254852c148639d49e7792324081a: Status 404 returned error can't find the container with id 71b03216f4dabb549b38a366842c6453cd8a254852c148639d49e7792324081a Feb 27 07:50:47 crc kubenswrapper[4612]: I0227 07:50:47.352419 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8dcjk" podStartSLOduration=44.352399742 podStartE2EDuration="44.352399742s" podCreationTimestamp="2026-02-27 07:50:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 07:50:47.336179605 +0000 UTC m=+105.190109613" watchObservedRunningTime="2026-02-27 07:50:47.352399742 +0000 UTC m=+105.206329740" Feb 27 07:50:47 crc kubenswrapper[4612]: I0227 07:50:47.409675 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5zllp\" (UniqueName: \"kubernetes.io/projected/457f0d99-2681-403b-abc2-92af86fa76e5-kube-api-access-5zllp\") pod \"ovnkube-node-hb4dm\" (UID: \"457f0d99-2681-403b-abc2-92af86fa76e5\") " pod="openshift-ovn-kubernetes/ovnkube-node-hb4dm" Feb 27 07:50:47 crc kubenswrapper[4612]: I0227 07:50:47.420915 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5zllp\" (UniqueName: \"kubernetes.io/projected/457f0d99-2681-403b-abc2-92af86fa76e5-kube-api-access-5zllp\") pod \"ovnkube-node-hb4dm\" (UID: \"457f0d99-2681-403b-abc2-92af86fa76e5\") " pod="openshift-ovn-kubernetes/ovnkube-node-hb4dm" Feb 27 07:50:47 crc kubenswrapper[4612]: I0227 07:50:47.572658 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-hb4dm" Feb 27 07:50:47 crc kubenswrapper[4612]: W0227 07:50:47.590883 4612 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod457f0d99_2681_403b_abc2_92af86fa76e5.slice/crio-212df7bdcf3f29364e20e1f6189ba971500b49282e1edd5512d243a568a05c08 WatchSource:0}: Error finding container 212df7bdcf3f29364e20e1f6189ba971500b49282e1edd5512d243a568a05c08: Status 404 returned error can't find the container with id 212df7bdcf3f29364e20e1f6189ba971500b49282e1edd5512d243a568a05c08 Feb 27 07:50:47 crc kubenswrapper[4612]: I0227 07:50:47.815160 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9421f529-9b1e-4bd0-8032-7baa608981b4-metrics-certs\") pod \"network-metrics-daemon-bqh6n\" (UID: \"9421f529-9b1e-4bd0-8032-7baa608981b4\") " pod="openshift-multus/network-metrics-daemon-bqh6n" Feb 27 07:50:47 crc kubenswrapper[4612]: E0227 07:50:47.815309 4612 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 27 07:50:47 crc kubenswrapper[4612]: E0227 07:50:47.815524 4612 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9421f529-9b1e-4bd0-8032-7baa608981b4-metrics-certs podName:9421f529-9b1e-4bd0-8032-7baa608981b4 nodeName:}" failed. No retries permitted until 2026-02-27 07:50:49.815507494 +0000 UTC m=+107.669437492 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/9421f529-9b1e-4bd0-8032-7baa608981b4-metrics-certs") pod "network-metrics-daemon-bqh6n" (UID: "9421f529-9b1e-4bd0-8032-7baa608981b4") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 27 07:50:47 crc kubenswrapper[4612]: I0227 07:50:47.851933 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 07:50:47 crc kubenswrapper[4612]: I0227 07:50:47.851959 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 07:50:47 crc kubenswrapper[4612]: I0227 07:50:47.851964 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 07:50:47 crc kubenswrapper[4612]: E0227 07:50:47.852048 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 07:50:47 crc kubenswrapper[4612]: I0227 07:50:47.852081 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-bqh6n" Feb 27 07:50:47 crc kubenswrapper[4612]: E0227 07:50:47.852254 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 07:50:47 crc kubenswrapper[4612]: E0227 07:50:47.852353 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-bqh6n" podUID="9421f529-9b1e-4bd0-8032-7baa608981b4" Feb 27 07:50:47 crc kubenswrapper[4612]: E0227 07:50:47.852489 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 07:50:48 crc kubenswrapper[4612]: I0227 07:50:48.311503 4612 generic.go:334] "Generic (PLEG): container finished" podID="ac122c97-bee5-4048-ac0b-2b9e56a194ff" containerID="5096c77a912baae08963fa2360a7df05cccccb94c57b27dc12b3a3612fb1b839" exitCode=0 Feb 27 07:50:48 crc kubenswrapper[4612]: I0227 07:50:48.311574 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-ljx5w" event={"ID":"ac122c97-bee5-4048-ac0b-2b9e56a194ff","Type":"ContainerDied","Data":"5096c77a912baae08963fa2360a7df05cccccb94c57b27dc12b3a3612fb1b839"} Feb 27 07:50:48 crc kubenswrapper[4612]: I0227 07:50:48.314970 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-74rg2" event={"ID":"3b90f13e-804f-4363-a67d-be2c0ac4d533","Type":"ContainerStarted","Data":"2d874041abce89cee21b49061195782f92385723a139a8d030b8e9067c80cf8e"} Feb 27 07:50:48 crc kubenswrapper[4612]: I0227 07:50:48.315015 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-74rg2" event={"ID":"3b90f13e-804f-4363-a67d-be2c0ac4d533","Type":"ContainerStarted","Data":"895ef4821f341cb46cf70b6335e94f5c9a8a32494145c84e6d702a5ed0edd3d5"} Feb 27 07:50:48 crc kubenswrapper[4612]: I0227 07:50:48.315030 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-74rg2" event={"ID":"3b90f13e-804f-4363-a67d-be2c0ac4d533","Type":"ContainerStarted","Data":"71b03216f4dabb549b38a366842c6453cd8a254852c148639d49e7792324081a"} Feb 27 07:50:48 crc kubenswrapper[4612]: I0227 07:50:48.317723 4612 generic.go:334] "Generic (PLEG): container finished" podID="457f0d99-2681-403b-abc2-92af86fa76e5" containerID="3f34f9c07a0fd73cf62299b31f3851cf759b6f85e30aa577427a8cd07a956aaf" exitCode=0 Feb 27 07:50:48 crc kubenswrapper[4612]: I0227 07:50:48.317735 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hb4dm" event={"ID":"457f0d99-2681-403b-abc2-92af86fa76e5","Type":"ContainerDied","Data":"3f34f9c07a0fd73cf62299b31f3851cf759b6f85e30aa577427a8cd07a956aaf"} Feb 27 07:50:48 crc kubenswrapper[4612]: I0227 07:50:48.317797 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hb4dm" event={"ID":"457f0d99-2681-403b-abc2-92af86fa76e5","Type":"ContainerStarted","Data":"212df7bdcf3f29364e20e1f6189ba971500b49282e1edd5512d243a568a05c08"} Feb 27 07:50:48 crc kubenswrapper[4612]: I0227 07:50:48.345462 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-bjg4l" podStartSLOduration=45.345436976 podStartE2EDuration="45.345436976s" podCreationTimestamp="2026-02-27 07:50:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 07:50:47.352802993 +0000 UTC m=+105.206733011" watchObservedRunningTime="2026-02-27 07:50:48.345436976 +0000 UTC m=+106.199367004" Feb 27 07:50:48 crc kubenswrapper[4612]: I0227 07:50:48.404496 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-74rg2" podStartSLOduration=45.404478203 podStartE2EDuration="45.404478203s" podCreationTimestamp="2026-02-27 07:50:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 07:50:48.36843342 +0000 UTC m=+106.222363438" watchObservedRunningTime="2026-02-27 07:50:48.404478203 +0000 UTC m=+106.258408211" Feb 27 07:50:48 crc kubenswrapper[4612]: I0227 07:50:48.867234 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Feb 27 07:50:49 crc kubenswrapper[4612]: I0227 07:50:49.325623 4612 generic.go:334] "Generic (PLEG): container finished" podID="ac122c97-bee5-4048-ac0b-2b9e56a194ff" containerID="55dfddf927446058ecb2e3c3b15caaf8a6de4bf78a18521b8453f92978ea16cb" exitCode=0 Feb 27 07:50:49 crc kubenswrapper[4612]: I0227 07:50:49.325707 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-ljx5w" event={"ID":"ac122c97-bee5-4048-ac0b-2b9e56a194ff","Type":"ContainerDied","Data":"55dfddf927446058ecb2e3c3b15caaf8a6de4bf78a18521b8453f92978ea16cb"} Feb 27 07:50:49 crc kubenswrapper[4612]: I0227 07:50:49.330843 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hb4dm" event={"ID":"457f0d99-2681-403b-abc2-92af86fa76e5","Type":"ContainerStarted","Data":"f782a058f108da24717810364f65cf60ee65d23c78dbaba7cd8939c96eef7a6e"} Feb 27 07:50:49 crc kubenswrapper[4612]: I0227 07:50:49.330903 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hb4dm" event={"ID":"457f0d99-2681-403b-abc2-92af86fa76e5","Type":"ContainerStarted","Data":"afbd64ee95f7e2f410dc87e9f75d523cc8c37d0567edc382c2d4bd06eaba651b"} Feb 27 07:50:49 crc kubenswrapper[4612]: I0227 07:50:49.330916 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hb4dm" event={"ID":"457f0d99-2681-403b-abc2-92af86fa76e5","Type":"ContainerStarted","Data":"804ccd844cb7c9fe0448b798d55e5f40e8e4b949c76f7acaeba4be31bc805ea0"} Feb 27 07:50:49 crc kubenswrapper[4612]: I0227 07:50:49.330933 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hb4dm" event={"ID":"457f0d99-2681-403b-abc2-92af86fa76e5","Type":"ContainerStarted","Data":"d6a8a1c2c4961b0cae8d98041227bad13852d7dab2d4043eb13ebfe7f9e983b8"} Feb 27 07:50:49 crc kubenswrapper[4612]: I0227 07:50:49.330945 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hb4dm" event={"ID":"457f0d99-2681-403b-abc2-92af86fa76e5","Type":"ContainerStarted","Data":"3938aacd5fe03026f6a2c84286053cb2e237e9ebe6325bc6d4f1da844329e77e"} Feb 27 07:50:49 crc kubenswrapper[4612]: I0227 07:50:49.330957 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hb4dm" event={"ID":"457f0d99-2681-403b-abc2-92af86fa76e5","Type":"ContainerStarted","Data":"272efda744ba67b324d44c1a9a0e60d9c2eede817ee274bdfa1a2adcef018517"} Feb 27 07:50:49 crc kubenswrapper[4612]: I0227 07:50:49.836621 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9421f529-9b1e-4bd0-8032-7baa608981b4-metrics-certs\") pod \"network-metrics-daemon-bqh6n\" (UID: \"9421f529-9b1e-4bd0-8032-7baa608981b4\") " pod="openshift-multus/network-metrics-daemon-bqh6n" Feb 27 07:50:49 crc kubenswrapper[4612]: E0227 07:50:49.836811 4612 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 27 07:50:49 crc kubenswrapper[4612]: E0227 07:50:49.836983 4612 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9421f529-9b1e-4bd0-8032-7baa608981b4-metrics-certs podName:9421f529-9b1e-4bd0-8032-7baa608981b4 nodeName:}" failed. No retries permitted until 2026-02-27 07:50:53.836966678 +0000 UTC m=+111.690896686 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/9421f529-9b1e-4bd0-8032-7baa608981b4-metrics-certs") pod "network-metrics-daemon-bqh6n" (UID: "9421f529-9b1e-4bd0-8032-7baa608981b4") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 27 07:50:49 crc kubenswrapper[4612]: I0227 07:50:49.852147 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-bqh6n" Feb 27 07:50:49 crc kubenswrapper[4612]: I0227 07:50:49.852201 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 07:50:49 crc kubenswrapper[4612]: I0227 07:50:49.852172 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 07:50:49 crc kubenswrapper[4612]: I0227 07:50:49.852147 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 07:50:49 crc kubenswrapper[4612]: E0227 07:50:49.852335 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-bqh6n" podUID="9421f529-9b1e-4bd0-8032-7baa608981b4" Feb 27 07:50:49 crc kubenswrapper[4612]: E0227 07:50:49.852408 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 07:50:49 crc kubenswrapper[4612]: E0227 07:50:49.852485 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 07:50:49 crc kubenswrapper[4612]: E0227 07:50:49.852600 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 07:50:50 crc kubenswrapper[4612]: I0227 07:50:50.337946 4612 generic.go:334] "Generic (PLEG): container finished" podID="ac122c97-bee5-4048-ac0b-2b9e56a194ff" containerID="2f6766c7f0693831f459e493d00600d703fe193a46438ddeef874e4466d76791" exitCode=0 Feb 27 07:50:50 crc kubenswrapper[4612]: I0227 07:50:50.338008 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-ljx5w" event={"ID":"ac122c97-bee5-4048-ac0b-2b9e56a194ff","Type":"ContainerDied","Data":"2f6766c7f0693831f459e493d00600d703fe193a46438ddeef874e4466d76791"} Feb 27 07:50:50 crc kubenswrapper[4612]: I0227 07:50:50.381372 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=2.381343889 podStartE2EDuration="2.381343889s" podCreationTimestamp="2026-02-27 07:50:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 07:50:49.388042507 +0000 UTC m=+107.241972535" watchObservedRunningTime="2026-02-27 07:50:50.381343889 +0000 UTC m=+108.235273947" Feb 27 07:50:51 crc kubenswrapper[4612]: I0227 07:50:51.345235 4612 generic.go:334] "Generic (PLEG): container finished" podID="ac122c97-bee5-4048-ac0b-2b9e56a194ff" containerID="54f6f4a51593572b70ab5386c784c2fb5b3df723e0c447ab2810a8a7dbbc8480" exitCode=0 Feb 27 07:50:51 crc kubenswrapper[4612]: I0227 07:50:51.345386 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-ljx5w" event={"ID":"ac122c97-bee5-4048-ac0b-2b9e56a194ff","Type":"ContainerDied","Data":"54f6f4a51593572b70ab5386c784c2fb5b3df723e0c447ab2810a8a7dbbc8480"} Feb 27 07:50:51 crc kubenswrapper[4612]: I0227 07:50:51.354125 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hb4dm" event={"ID":"457f0d99-2681-403b-abc2-92af86fa76e5","Type":"ContainerStarted","Data":"53ab32307d4bba133b74665292857b1dad285f3ef223f0758c3752cf1355d4da"} Feb 27 07:50:51 crc kubenswrapper[4612]: I0227 07:50:51.558650 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 07:50:51 crc kubenswrapper[4612]: E0227 07:50:51.559016 4612 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 07:51:07.558980181 +0000 UTC m=+125.412910189 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 07:50:51 crc kubenswrapper[4612]: I0227 07:50:51.659739 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 07:50:51 crc kubenswrapper[4612]: I0227 07:50:51.659802 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 07:50:51 crc kubenswrapper[4612]: I0227 07:50:51.659848 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 07:50:51 crc kubenswrapper[4612]: I0227 07:50:51.659889 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 07:50:51 crc kubenswrapper[4612]: E0227 07:50:51.659917 4612 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 27 07:50:51 crc kubenswrapper[4612]: E0227 07:50:51.659971 4612 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 27 07:50:51 crc kubenswrapper[4612]: E0227 07:50:51.659983 4612 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-27 07:51:07.659963394 +0000 UTC m=+125.513893392 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 27 07:50:51 crc kubenswrapper[4612]: E0227 07:50:51.660024 4612 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-27 07:51:07.660006075 +0000 UTC m=+125.513936083 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 27 07:50:51 crc kubenswrapper[4612]: E0227 07:50:51.660114 4612 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 27 07:50:51 crc kubenswrapper[4612]: E0227 07:50:51.660131 4612 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 27 07:50:51 crc kubenswrapper[4612]: E0227 07:50:51.660146 4612 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 27 07:50:51 crc kubenswrapper[4612]: E0227 07:50:51.660175 4612 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-27 07:51:07.660165839 +0000 UTC m=+125.514095847 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 27 07:50:51 crc kubenswrapper[4612]: E0227 07:50:51.660227 4612 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 27 07:50:51 crc kubenswrapper[4612]: E0227 07:50:51.660238 4612 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 27 07:50:51 crc kubenswrapper[4612]: E0227 07:50:51.660247 4612 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 27 07:50:51 crc kubenswrapper[4612]: E0227 07:50:51.660275 4612 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-27 07:51:07.660266812 +0000 UTC m=+125.514196820 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 27 07:50:51 crc kubenswrapper[4612]: I0227 07:50:51.852332 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 07:50:51 crc kubenswrapper[4612]: E0227 07:50:51.852472 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 07:50:51 crc kubenswrapper[4612]: I0227 07:50:51.852861 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-bqh6n" Feb 27 07:50:51 crc kubenswrapper[4612]: I0227 07:50:51.853018 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 07:50:51 crc kubenswrapper[4612]: E0227 07:50:51.853140 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-bqh6n" podUID="9421f529-9b1e-4bd0-8032-7baa608981b4" Feb 27 07:50:51 crc kubenswrapper[4612]: I0227 07:50:51.853243 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 07:50:51 crc kubenswrapper[4612]: E0227 07:50:51.853444 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 07:50:51 crc kubenswrapper[4612]: I0227 07:50:51.853627 4612 scope.go:117] "RemoveContainer" containerID="6fdb61a61c41ceec6aa497297d1853ae6fec429a9dc998e0f49e8f41f41b7e9c" Feb 27 07:50:51 crc kubenswrapper[4612]: E0227 07:50:51.853976 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 07:50:52 crc kubenswrapper[4612]: I0227 07:50:52.359893 4612 generic.go:334] "Generic (PLEG): container finished" podID="ac122c97-bee5-4048-ac0b-2b9e56a194ff" containerID="3638d68d0cd981b2c1347c0f562892e065df9a12dcf6ca1894381d9df5684960" exitCode=0 Feb 27 07:50:52 crc kubenswrapper[4612]: I0227 07:50:52.360262 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-ljx5w" event={"ID":"ac122c97-bee5-4048-ac0b-2b9e56a194ff","Type":"ContainerDied","Data":"3638d68d0cd981b2c1347c0f562892e065df9a12dcf6ca1894381d9df5684960"} Feb 27 07:50:52 crc kubenswrapper[4612]: I0227 07:50:52.373124 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/3.log" Feb 27 07:50:52 crc kubenswrapper[4612]: I0227 07:50:52.375144 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"4cad3a2314cde773106537076b97d74afb9103af33e0450a5d804a1026fe06b7"} Feb 27 07:50:52 crc kubenswrapper[4612]: I0227 07:50:52.375344 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 27 07:50:53 crc kubenswrapper[4612]: I0227 07:50:53.380858 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-ljx5w" event={"ID":"ac122c97-bee5-4048-ac0b-2b9e56a194ff","Type":"ContainerStarted","Data":"33ebb0bf692cdf27ffaece7b2f4b31e9d7d50b4ca7d0645b3c72470fd4966274"} Feb 27 07:50:53 crc kubenswrapper[4612]: I0227 07:50:53.392016 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hb4dm" event={"ID":"457f0d99-2681-403b-abc2-92af86fa76e5","Type":"ContainerStarted","Data":"a920896c7db750f9e9ad9cd609c66d12c71145a4b4cc475ae07fbdc1d2694a7b"} Feb 27 07:50:53 crc kubenswrapper[4612]: I0227 07:50:53.392115 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-hb4dm" Feb 27 07:50:53 crc kubenswrapper[4612]: I0227 07:50:53.392150 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-hb4dm" Feb 27 07:50:53 crc kubenswrapper[4612]: I0227 07:50:53.406104 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=16.40608916 podStartE2EDuration="16.40608916s" podCreationTimestamp="2026-02-27 07:50:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 07:50:52.43682189 +0000 UTC m=+110.290751888" watchObservedRunningTime="2026-02-27 07:50:53.40608916 +0000 UTC m=+111.260019158" Feb 27 07:50:53 crc kubenswrapper[4612]: I0227 07:50:53.406246 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-ljx5w" podStartSLOduration=50.406241315 podStartE2EDuration="50.406241315s" podCreationTimestamp="2026-02-27 07:50:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 07:50:53.405952447 +0000 UTC m=+111.259882445" watchObservedRunningTime="2026-02-27 07:50:53.406241315 +0000 UTC m=+111.260171313" Feb 27 07:50:53 crc kubenswrapper[4612]: I0227 07:50:53.424025 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-hb4dm" Feb 27 07:50:53 crc kubenswrapper[4612]: I0227 07:50:53.475151 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-hb4dm" podStartSLOduration=50.475133193 podStartE2EDuration="50.475133193s" podCreationTimestamp="2026-02-27 07:50:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 07:50:53.439797549 +0000 UTC m=+111.293727557" watchObservedRunningTime="2026-02-27 07:50:53.475133193 +0000 UTC m=+111.329063191" Feb 27 07:50:53 crc kubenswrapper[4612]: I0227 07:50:53.852711 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-bqh6n" Feb 27 07:50:53 crc kubenswrapper[4612]: I0227 07:50:53.852771 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 07:50:53 crc kubenswrapper[4612]: I0227 07:50:53.852775 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 07:50:53 crc kubenswrapper[4612]: E0227 07:50:53.852865 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-bqh6n" podUID="9421f529-9b1e-4bd0-8032-7baa608981b4" Feb 27 07:50:53 crc kubenswrapper[4612]: I0227 07:50:53.852915 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 07:50:53 crc kubenswrapper[4612]: E0227 07:50:53.852982 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 07:50:53 crc kubenswrapper[4612]: E0227 07:50:53.853056 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 07:50:53 crc kubenswrapper[4612]: E0227 07:50:53.853122 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 07:50:53 crc kubenswrapper[4612]: I0227 07:50:53.882563 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9421f529-9b1e-4bd0-8032-7baa608981b4-metrics-certs\") pod \"network-metrics-daemon-bqh6n\" (UID: \"9421f529-9b1e-4bd0-8032-7baa608981b4\") " pod="openshift-multus/network-metrics-daemon-bqh6n" Feb 27 07:50:53 crc kubenswrapper[4612]: E0227 07:50:53.882685 4612 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 27 07:50:53 crc kubenswrapper[4612]: E0227 07:50:53.882764 4612 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9421f529-9b1e-4bd0-8032-7baa608981b4-metrics-certs podName:9421f529-9b1e-4bd0-8032-7baa608981b4 nodeName:}" failed. No retries permitted until 2026-02-27 07:51:01.882747546 +0000 UTC m=+119.736677544 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/9421f529-9b1e-4bd0-8032-7baa608981b4-metrics-certs") pod "network-metrics-daemon-bqh6n" (UID: "9421f529-9b1e-4bd0-8032-7baa608981b4") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 27 07:50:54 crc kubenswrapper[4612]: I0227 07:50:54.390375 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-hb4dm" Feb 27 07:50:54 crc kubenswrapper[4612]: I0227 07:50:54.424723 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-hb4dm" Feb 27 07:50:55 crc kubenswrapper[4612]: I0227 07:50:55.723238 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-bqh6n"] Feb 27 07:50:55 crc kubenswrapper[4612]: I0227 07:50:55.723371 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-bqh6n" Feb 27 07:50:55 crc kubenswrapper[4612]: E0227 07:50:55.723515 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-bqh6n" podUID="9421f529-9b1e-4bd0-8032-7baa608981b4" Feb 27 07:50:55 crc kubenswrapper[4612]: I0227 07:50:55.852632 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 07:50:55 crc kubenswrapper[4612]: I0227 07:50:55.852683 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 07:50:55 crc kubenswrapper[4612]: E0227 07:50:55.852766 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 07:50:55 crc kubenswrapper[4612]: I0227 07:50:55.852721 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 07:50:55 crc kubenswrapper[4612]: E0227 07:50:55.852873 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 07:50:55 crc kubenswrapper[4612]: E0227 07:50:55.852928 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 07:50:57 crc kubenswrapper[4612]: I0227 07:50:57.852467 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 07:50:57 crc kubenswrapper[4612]: E0227 07:50:57.852919 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 27 07:50:57 crc kubenswrapper[4612]: I0227 07:50:57.852517 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-bqh6n" Feb 27 07:50:57 crc kubenswrapper[4612]: E0227 07:50:57.853044 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-bqh6n" podUID="9421f529-9b1e-4bd0-8032-7baa608981b4" Feb 27 07:50:57 crc kubenswrapper[4612]: I0227 07:50:57.852594 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 07:50:57 crc kubenswrapper[4612]: E0227 07:50:57.853132 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 27 07:50:57 crc kubenswrapper[4612]: I0227 07:50:57.852468 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 07:50:57 crc kubenswrapper[4612]: E0227 07:50:57.853211 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.390639 4612 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.391003 4612 kubelet_node_status.go:538] "Fast updating node status as it just became ready" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.445553 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-k6s2g"] Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.445985 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-k6s2g" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.448283 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-q6h7h"] Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.448633 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-96vl6"] Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.448894 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-q6h7h" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.449179 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-96vl6" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.449805 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-m5b65"] Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.450040 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-m5b65" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.450458 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-jdng9"] Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.450659 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-jdng9" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.450998 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-zbccb"] Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.451251 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-zbccb" Feb 27 07:50:58 crc kubenswrapper[4612]: W0227 07:50:58.455361 4612 reflector.go:561] object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4": failed to list *v1.Secret: secrets "machine-approver-sa-dockercfg-nl2j4" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-cluster-machine-approver": no relationship found between node 'crc' and this object Feb 27 07:50:58 crc kubenswrapper[4612]: E0227 07:50:58.455400 4612 reflector.go:158] "Unhandled Error" err="object-\"openshift-cluster-machine-approver\"/\"machine-approver-sa-dockercfg-nl2j4\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"machine-approver-sa-dockercfg-nl2j4\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-cluster-machine-approver\": no relationship found between node 'crc' and this object" logger="UnhandledError" Feb 27 07:50:58 crc kubenswrapper[4612]: W0227 07:50:58.455445 4612 reflector.go:561] object-"openshift-cluster-machine-approver"/"kube-rbac-proxy": failed to list *v1.ConfigMap: configmaps "kube-rbac-proxy" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-cluster-machine-approver": no relationship found between node 'crc' and this object Feb 27 07:50:58 crc kubenswrapper[4612]: E0227 07:50:58.455456 4612 reflector.go:158] "Unhandled Error" err="object-\"openshift-cluster-machine-approver\"/\"kube-rbac-proxy\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"kube-rbac-proxy\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-cluster-machine-approver\": no relationship found between node 'crc' and this object" logger="UnhandledError" Feb 27 07:50:58 crc kubenswrapper[4612]: W0227 07:50:58.455490 4612 reflector.go:561] object-"openshift-cluster-machine-approver"/"machine-approver-config": failed to list *v1.ConfigMap: configmaps "machine-approver-config" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-cluster-machine-approver": no relationship found between node 'crc' and this object Feb 27 07:50:58 crc kubenswrapper[4612]: E0227 07:50:58.455501 4612 reflector.go:158] "Unhandled Error" err="object-\"openshift-cluster-machine-approver\"/\"machine-approver-config\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"machine-approver-config\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-cluster-machine-approver\": no relationship found between node 'crc' and this object" logger="UnhandledError" Feb 27 07:50:58 crc kubenswrapper[4612]: W0227 07:50:58.455533 4612 reflector.go:561] object-"openshift-cluster-machine-approver"/"kube-root-ca.crt": failed to list *v1.ConfigMap: configmaps "kube-root-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-cluster-machine-approver": no relationship found between node 'crc' and this object Feb 27 07:50:58 crc kubenswrapper[4612]: E0227 07:50:58.455542 4612 reflector.go:158] "Unhandled Error" err="object-\"openshift-cluster-machine-approver\"/\"kube-root-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"kube-root-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-cluster-machine-approver\": no relationship found between node 'crc' and this object" logger="UnhandledError" Feb 27 07:50:58 crc kubenswrapper[4612]: W0227 07:50:58.460899 4612 reflector.go:561] object-"openshift-cluster-machine-approver"/"machine-approver-tls": failed to list *v1.Secret: secrets "machine-approver-tls" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-cluster-machine-approver": no relationship found between node 'crc' and this object Feb 27 07:50:58 crc kubenswrapper[4612]: E0227 07:50:58.460945 4612 reflector.go:158] "Unhandled Error" err="object-\"openshift-cluster-machine-approver\"/\"machine-approver-tls\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"machine-approver-tls\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-cluster-machine-approver\": no relationship found between node 'crc' and this object" logger="UnhandledError" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.465344 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-2rrfz"] Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.465820 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-8wk54"] Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.466133 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-8wk54" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.466494 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2rrfz" Feb 27 07:50:58 crc kubenswrapper[4612]: W0227 07:50:58.466546 4612 reflector.go:561] object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt": failed to list *v1.ConfigMap: configmaps "openshift-service-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-cluster-machine-approver": no relationship found between node 'crc' and this object Feb 27 07:50:58 crc kubenswrapper[4612]: E0227 07:50:58.466582 4612 reflector.go:158] "Unhandled Error" err="object-\"openshift-cluster-machine-approver\"/\"openshift-service-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"openshift-service-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-cluster-machine-approver\": no relationship found between node 'crc' and this object" logger="UnhandledError" Feb 27 07:50:58 crc kubenswrapper[4612]: W0227 07:50:58.467651 4612 reflector.go:561] object-"openshift-machine-api"/"openshift-service-ca.crt": failed to list *v1.ConfigMap: configmaps "openshift-service-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-machine-api": no relationship found between node 'crc' and this object Feb 27 07:50:58 crc kubenswrapper[4612]: E0227 07:50:58.467683 4612 reflector.go:158] "Unhandled Error" err="object-\"openshift-machine-api\"/\"openshift-service-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"openshift-service-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-machine-api\": no relationship found between node 'crc' and this object" logger="UnhandledError" Feb 27 07:50:58 crc kubenswrapper[4612]: W0227 07:50:58.467899 4612 reflector.go:561] object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7": failed to list *v1.Secret: secrets "machine-api-operator-dockercfg-mfbb7" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-machine-api": no relationship found between node 'crc' and this object Feb 27 07:50:58 crc kubenswrapper[4612]: E0227 07:50:58.467923 4612 reflector.go:158] "Unhandled Error" err="object-\"openshift-machine-api\"/\"machine-api-operator-dockercfg-mfbb7\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"machine-api-operator-dockercfg-mfbb7\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-machine-api\": no relationship found between node 'crc' and this object" logger="UnhandledError" Feb 27 07:50:58 crc kubenswrapper[4612]: W0227 07:50:58.468285 4612 reflector.go:561] object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt": failed to list *v1.ConfigMap: configmaps "openshift-service-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-cluster-samples-operator": no relationship found between node 'crc' and this object Feb 27 07:50:58 crc kubenswrapper[4612]: E0227 07:50:58.468311 4612 reflector.go:158] "Unhandled Error" err="object-\"openshift-cluster-samples-operator\"/\"openshift-service-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"openshift-service-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-cluster-samples-operator\": no relationship found between node 'crc' and this object" logger="UnhandledError" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.482777 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.486348 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.486672 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.486778 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.486350 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.486781 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.490134 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.490525 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.490798 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.491338 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.491537 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.491548 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.491334 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.491827 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.491848 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.492005 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.492010 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.492076 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.492130 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.492216 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.492328 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.492349 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.492440 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.492120 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.492782 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.495582 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.502157 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.503279 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-nkbg2"] Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.503766 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-nkbg2" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.506703 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.506901 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.506963 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.507117 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.507185 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.507187 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.507133 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.518365 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.518427 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.526287 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.526506 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.526650 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.526822 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.526986 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.527346 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-z9zch"] Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.527826 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-z9zch" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.527351 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.527714 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.527801 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.528959 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.529073 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-spnfb"] Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.529202 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.529464 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-spnfb" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.530956 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.531322 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eb060975-a3f2-43f4-b4eb-9456a81c2f1f-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-m5b65\" (UID: \"eb060975-a3f2-43f4-b4eb-9456a81c2f1f\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-m5b65" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.531417 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ca4c97ff-9032-48de-b286-89da26f5a822-service-ca-bundle\") pod \"authentication-operator-69f744f599-jdng9\" (UID: \"ca4c97ff-9032-48de-b286-89da26f5a822\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-jdng9" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.531485 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v7d4b\" (UniqueName: \"kubernetes.io/projected/eb060975-a3f2-43f4-b4eb-9456a81c2f1f-kube-api-access-v7d4b\") pod \"openshift-controller-manager-operator-756b6f6bc6-m5b65\" (UID: \"eb060975-a3f2-43f4-b4eb-9456a81c2f1f\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-m5b65" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.531574 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4l9jx\" (UniqueName: \"kubernetes.io/projected/094b001c-e80c-4d07-a161-e00514667462-kube-api-access-4l9jx\") pod \"machine-api-operator-5694c8668f-q6h7h\" (UID: \"094b001c-e80c-4d07-a161-e00514667462\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-q6h7h" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.531650 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/eb060975-a3f2-43f4-b4eb-9456a81c2f1f-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-m5b65\" (UID: \"eb060975-a3f2-43f4-b4eb-9456a81c2f1f\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-m5b65" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.531770 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/29ab8b79-a05e-4b94-a2bb-2c6b412fd297-machine-approver-tls\") pod \"machine-approver-56656f9798-k6s2g\" (UID: \"29ab8b79-a05e-4b94-a2bb-2c6b412fd297\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-k6s2g" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.531835 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/094b001c-e80c-4d07-a161-e00514667462-images\") pod \"machine-api-operator-5694c8668f-q6h7h\" (UID: \"094b001c-e80c-4d07-a161-e00514667462\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-q6h7h" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.531578 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.539527 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k7t8s\" (UniqueName: \"kubernetes.io/projected/620a7a2f-92fc-4a76-900b-021d0dd78d79-kube-api-access-k7t8s\") pod \"cluster-samples-operator-665b6dd947-96vl6\" (UID: \"620a7a2f-92fc-4a76-900b-021d0dd78d79\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-96vl6" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.557137 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/29ab8b79-a05e-4b94-a2bb-2c6b412fd297-config\") pod \"machine-approver-56656f9798-k6s2g\" (UID: \"29ab8b79-a05e-4b94-a2bb-2c6b412fd297\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-k6s2g" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.559155 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/29ab8b79-a05e-4b94-a2bb-2c6b412fd297-auth-proxy-config\") pod \"machine-approver-56656f9798-k6s2g\" (UID: \"29ab8b79-a05e-4b94-a2bb-2c6b412fd297\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-k6s2g" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.559193 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c8t7j\" (UniqueName: \"kubernetes.io/projected/a5bbba10-16da-4fce-ae5f-311fc5d0d6c1-kube-api-access-c8t7j\") pod \"downloads-7954f5f757-zbccb\" (UID: \"a5bbba10-16da-4fce-ae5f-311fc5d0d6c1\") " pod="openshift-console/downloads-7954f5f757-zbccb" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.559211 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/620a7a2f-92fc-4a76-900b-021d0dd78d79-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-96vl6\" (UID: \"620a7a2f-92fc-4a76-900b-021d0dd78d79\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-96vl6" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.559231 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ca4c97ff-9032-48de-b286-89da26f5a822-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-jdng9\" (UID: \"ca4c97ff-9032-48de-b286-89da26f5a822\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-jdng9" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.559249 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ca4c97ff-9032-48de-b286-89da26f5a822-serving-cert\") pod \"authentication-operator-69f744f599-jdng9\" (UID: \"ca4c97ff-9032-48de-b286-89da26f5a822\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-jdng9" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.559264 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dpltn\" (UniqueName: \"kubernetes.io/projected/ca4c97ff-9032-48de-b286-89da26f5a822-kube-api-access-dpltn\") pod \"authentication-operator-69f744f599-jdng9\" (UID: \"ca4c97ff-9032-48de-b286-89da26f5a822\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-jdng9" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.559292 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/094b001c-e80c-4d07-a161-e00514667462-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-q6h7h\" (UID: \"094b001c-e80c-4d07-a161-e00514667462\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-q6h7h" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.560761 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jplnj\" (UniqueName: \"kubernetes.io/projected/29ab8b79-a05e-4b94-a2bb-2c6b412fd297-kube-api-access-jplnj\") pod \"machine-approver-56656f9798-k6s2g\" (UID: \"29ab8b79-a05e-4b94-a2bb-2c6b412fd297\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-k6s2g" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.560864 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ca4c97ff-9032-48de-b286-89da26f5a822-config\") pod \"authentication-operator-69f744f599-jdng9\" (UID: \"ca4c97ff-9032-48de-b286-89da26f5a822\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-jdng9" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.560948 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/094b001c-e80c-4d07-a161-e00514667462-config\") pod \"machine-api-operator-5694c8668f-q6h7h\" (UID: \"094b001c-e80c-4d07-a161-e00514667462\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-q6h7h" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.561021 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-cw57r"] Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.561512 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2k66r"] Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.561733 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-sxt8c"] Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.561988 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-sxt8c" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.562293 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-cw57r" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.562530 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.562721 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-qvwzc"] Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.562541 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2k66r" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.563481 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-qvwzc" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.566575 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-pdjlb"] Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.567222 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-pdjlb" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.576395 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-w7pv5"] Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.577327 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bk49v"] Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.577648 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-hp72h"] Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.579166 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-g9hq6"] Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.579670 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-kbtxf"] Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.580144 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-kbtxf" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.580210 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-2d7kh"] Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.580535 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-w7pv5" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.580659 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-2d7kh" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.580818 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bk49v" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.581137 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-hp72h" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.581571 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-g9hq6" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.583779 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.589225 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.590111 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.590367 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.590620 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.593222 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.596756 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-jzd5b"] Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.601495 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.601521 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.602461 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.602609 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.602765 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.602933 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.602943 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.603054 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.603116 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.603209 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.604115 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-jzd5b" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.605768 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-c6sdx"] Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.609678 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.610369 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-ghtdk"] Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.610917 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-f8qzz"] Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.611104 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-c6sdx" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.611126 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-ghtdk" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.611391 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-f8qzz" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.649737 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.650294 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.657332 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-5wvhc"] Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.658073 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-8l852"] Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.658787 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-5wvhc" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.662810 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29536305-vmtnx"] Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.663392 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29536305-vmtnx" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.663658 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-8l852" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.666890 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-4lpn2"] Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.667422 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-42tqr"] Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.675268 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-gcfmk"] Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.675742 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-gcfmk" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.676031 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-4lpn2" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.673178 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.676340 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-42tqr" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.673230 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.673269 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.673296 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.673299 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.673322 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.677323 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.677446 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.677556 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/ff5fcfe1-eb01-4f01-857d-a67712db3b87-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-5wvhc\" (UID: \"ff5fcfe1-eb01-4f01-857d-a67712db3b87\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-5wvhc" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.677595 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v7d4b\" (UniqueName: \"kubernetes.io/projected/eb060975-a3f2-43f4-b4eb-9456a81c2f1f-kube-api-access-v7d4b\") pod \"openshift-controller-manager-operator-756b6f6bc6-m5b65\" (UID: \"eb060975-a3f2-43f4-b4eb-9456a81c2f1f\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-m5b65" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.677616 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6ece12fd-d439-4104-a7a9-a9d174e29b5a-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-nkbg2\" (UID: \"6ece12fd-d439-4104-a7a9-a9d174e29b5a\") " pod="openshift-authentication/oauth-openshift-558db77b4-nkbg2" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.677636 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/19b65635-c86d-4439-8ece-4d8c3dbe986b-trusted-ca\") pod \"console-operator-58897d9998-qvwzc\" (UID: \"19b65635-c86d-4439-8ece-4d8c3dbe986b\") " pod="openshift-console-operator/console-operator-58897d9998-qvwzc" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.677660 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/6ece12fd-d439-4104-a7a9-a9d174e29b5a-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-nkbg2\" (UID: \"6ece12fd-d439-4104-a7a9-a9d174e29b5a\") " pod="openshift-authentication/oauth-openshift-558db77b4-nkbg2" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.677678 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4l9jx\" (UniqueName: \"kubernetes.io/projected/094b001c-e80c-4d07-a161-e00514667462-kube-api-access-4l9jx\") pod \"machine-api-operator-5694c8668f-q6h7h\" (UID: \"094b001c-e80c-4d07-a161-e00514667462\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-q6h7h" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.677710 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/da284312-b98a-48d6-8cff-f8071c334e0b-console-config\") pod \"console-f9d7485db-8wk54\" (UID: \"da284312-b98a-48d6-8cff-f8071c334e0b\") " pod="openshift-console/console-f9d7485db-8wk54" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.677727 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/6ece12fd-d439-4104-a7a9-a9d174e29b5a-audit-dir\") pod \"oauth-openshift-558db77b4-nkbg2\" (UID: \"6ece12fd-d439-4104-a7a9-a9d174e29b5a\") " pod="openshift-authentication/oauth-openshift-558db77b4-nkbg2" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.677742 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/8cc01026-2c76-4446-b78f-7d8fb5d9a887-proxy-tls\") pod \"machine-config-operator-74547568cd-w7pv5\" (UID: \"8cc01026-2c76-4446-b78f-7d8fb5d9a887\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-w7pv5" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.677757 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w8p5f\" (UniqueName: \"kubernetes.io/projected/19b65635-c86d-4439-8ece-4d8c3dbe986b-kube-api-access-w8p5f\") pod \"console-operator-58897d9998-qvwzc\" (UID: \"19b65635-c86d-4439-8ece-4d8c3dbe986b\") " pod="openshift-console-operator/console-operator-58897d9998-qvwzc" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.677771 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-69bw7\" (UniqueName: \"kubernetes.io/projected/cc3f4dca-b740-4294-93e5-73906f8dc82c-kube-api-access-69bw7\") pod \"migrator-59844c95c7-g9hq6\" (UID: \"cc3f4dca-b740-4294-93e5-73906f8dc82c\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-g9hq6" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.677787 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/eb060975-a3f2-43f4-b4eb-9456a81c2f1f-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-m5b65\" (UID: \"eb060975-a3f2-43f4-b4eb-9456a81c2f1f\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-m5b65" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.677801 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/29ab8b79-a05e-4b94-a2bb-2c6b412fd297-machine-approver-tls\") pod \"machine-approver-56656f9798-k6s2g\" (UID: \"29ab8b79-a05e-4b94-a2bb-2c6b412fd297\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-k6s2g" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.677827 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/094b001c-e80c-4d07-a161-e00514667462-images\") pod \"machine-api-operator-5694c8668f-q6h7h\" (UID: \"094b001c-e80c-4d07-a161-e00514667462\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-q6h7h" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.677843 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/6ece12fd-d439-4104-a7a9-a9d174e29b5a-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-nkbg2\" (UID: \"6ece12fd-d439-4104-a7a9-a9d174e29b5a\") " pod="openshift-authentication/oauth-openshift-558db77b4-nkbg2" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.677858 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/6ece12fd-d439-4104-a7a9-a9d174e29b5a-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-nkbg2\" (UID: \"6ece12fd-d439-4104-a7a9-a9d174e29b5a\") " pod="openshift-authentication/oauth-openshift-558db77b4-nkbg2" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.677871 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/da284312-b98a-48d6-8cff-f8071c334e0b-service-ca\") pod \"console-f9d7485db-8wk54\" (UID: \"da284312-b98a-48d6-8cff-f8071c334e0b\") " pod="openshift-console/console-f9d7485db-8wk54" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.677886 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/da284312-b98a-48d6-8cff-f8071c334e0b-trusted-ca-bundle\") pod \"console-f9d7485db-8wk54\" (UID: \"da284312-b98a-48d6-8cff-f8071c334e0b\") " pod="openshift-console/console-f9d7485db-8wk54" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.677900 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/6d061f42-06de-43e5-b4d0-dbca09e48630-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-hp72h\" (UID: \"6d061f42-06de-43e5-b4d0-dbca09e48630\") " pod="openshift-marketplace/marketplace-operator-79b997595-hp72h" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.677916 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/78554bd5-6fc8-4ce8-b1d2-b183dab17495-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-2k66r\" (UID: \"78554bd5-6fc8-4ce8-b1d2-b183dab17495\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2k66r" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.677933 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/6ece12fd-d439-4104-a7a9-a9d174e29b5a-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-nkbg2\" (UID: \"6ece12fd-d439-4104-a7a9-a9d174e29b5a\") " pod="openshift-authentication/oauth-openshift-558db77b4-nkbg2" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.677949 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/9bb5eece-3596-485e-a859-b61d201b05f5-available-featuregates\") pod \"openshift-config-operator-7777fb866f-z9zch\" (UID: \"9bb5eece-3596-485e-a859-b61d201b05f5\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-z9zch" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.677965 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/19b65635-c86d-4439-8ece-4d8c3dbe986b-serving-cert\") pod \"console-operator-58897d9998-qvwzc\" (UID: \"19b65635-c86d-4439-8ece-4d8c3dbe986b\") " pod="openshift-console-operator/console-operator-58897d9998-qvwzc" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.677980 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/3b5b5090-45a4-4acc-aa94-84c892bcf306-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-f8qzz\" (UID: \"3b5b5090-45a4-4acc-aa94-84c892bcf306\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-f8qzz" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.677999 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k7t8s\" (UniqueName: \"kubernetes.io/projected/620a7a2f-92fc-4a76-900b-021d0dd78d79-kube-api-access-k7t8s\") pod \"cluster-samples-operator-665b6dd947-96vl6\" (UID: \"620a7a2f-92fc-4a76-900b-021d0dd78d79\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-96vl6" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.678014 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/29ab8b79-a05e-4b94-a2bb-2c6b412fd297-config\") pod \"machine-approver-56656f9798-k6s2g\" (UID: \"29ab8b79-a05e-4b94-a2bb-2c6b412fd297\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-k6s2g" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.678030 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/d9aa8f66-da03-49e5-b7b4-d17135ccaa03-etcd-client\") pod \"apiserver-7bbb656c7d-2rrfz\" (UID: \"d9aa8f66-da03-49e5-b7b4-d17135ccaa03\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2rrfz" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.678044 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/d9aa8f66-da03-49e5-b7b4-d17135ccaa03-audit-dir\") pod \"apiserver-7bbb656c7d-2rrfz\" (UID: \"d9aa8f66-da03-49e5-b7b4-d17135ccaa03\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2rrfz" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.678059 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/42b49d65-d823-45af-b274-380dfbddc9b7-client-ca\") pod \"route-controller-manager-6576b87f9c-spnfb\" (UID: \"42b49d65-d823-45af-b274-380dfbddc9b7\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-spnfb" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.678075 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/3b5b5090-45a4-4acc-aa94-84c892bcf306-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-f8qzz\" (UID: \"3b5b5090-45a4-4acc-aa94-84c892bcf306\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-f8qzz" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.678091 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/29ab8b79-a05e-4b94-a2bb-2c6b412fd297-auth-proxy-config\") pod \"machine-approver-56656f9798-k6s2g\" (UID: \"29ab8b79-a05e-4b94-a2bb-2c6b412fd297\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-k6s2g" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.678106 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/78554bd5-6fc8-4ce8-b1d2-b183dab17495-config\") pod \"openshift-apiserver-operator-796bbdcf4f-2k66r\" (UID: \"78554bd5-6fc8-4ce8-b1d2-b183dab17495\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2k66r" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.678123 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lp2b5\" (UniqueName: \"kubernetes.io/projected/8cc01026-2c76-4446-b78f-7d8fb5d9a887-kube-api-access-lp2b5\") pod \"machine-config-operator-74547568cd-w7pv5\" (UID: \"8cc01026-2c76-4446-b78f-7d8fb5d9a887\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-w7pv5" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.678141 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/faae1773-bf96-4127-8060-a40da07930ce-stats-auth\") pod \"router-default-5444994796-2d7kh\" (UID: \"faae1773-bf96-4127-8060-a40da07930ce\") " pod="openshift-ingress/router-default-5444994796-2d7kh" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.678155 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/d9aa8f66-da03-49e5-b7b4-d17135ccaa03-encryption-config\") pod \"apiserver-7bbb656c7d-2rrfz\" (UID: \"d9aa8f66-da03-49e5-b7b4-d17135ccaa03\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2rrfz" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.678170 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/6ece12fd-d439-4104-a7a9-a9d174e29b5a-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-nkbg2\" (UID: \"6ece12fd-d439-4104-a7a9-a9d174e29b5a\") " pod="openshift-authentication/oauth-openshift-558db77b4-nkbg2" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.678186 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/faae1773-bf96-4127-8060-a40da07930ce-metrics-certs\") pod \"router-default-5444994796-2d7kh\" (UID: \"faae1773-bf96-4127-8060-a40da07930ce\") " pod="openshift-ingress/router-default-5444994796-2d7kh" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.678206 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/da284312-b98a-48d6-8cff-f8071c334e0b-console-serving-cert\") pod \"console-f9d7485db-8wk54\" (UID: \"da284312-b98a-48d6-8cff-f8071c334e0b\") " pod="openshift-console/console-f9d7485db-8wk54" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.678220 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cxrc9\" (UniqueName: \"kubernetes.io/projected/d9aa8f66-da03-49e5-b7b4-d17135ccaa03-kube-api-access-cxrc9\") pod \"apiserver-7bbb656c7d-2rrfz\" (UID: \"d9aa8f66-da03-49e5-b7b4-d17135ccaa03\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2rrfz" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.678233 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/d9aa8f66-da03-49e5-b7b4-d17135ccaa03-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-2rrfz\" (UID: \"d9aa8f66-da03-49e5-b7b4-d17135ccaa03\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2rrfz" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.678248 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/faae1773-bf96-4127-8060-a40da07930ce-service-ca-bundle\") pod \"router-default-5444994796-2d7kh\" (UID: \"faae1773-bf96-4127-8060-a40da07930ce\") " pod="openshift-ingress/router-default-5444994796-2d7kh" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.678262 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/6ece12fd-d439-4104-a7a9-a9d174e29b5a-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-nkbg2\" (UID: \"6ece12fd-d439-4104-a7a9-a9d174e29b5a\") " pod="openshift-authentication/oauth-openshift-558db77b4-nkbg2" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.678275 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/8cc01026-2c76-4446-b78f-7d8fb5d9a887-images\") pod \"machine-config-operator-74547568cd-w7pv5\" (UID: \"8cc01026-2c76-4446-b78f-7d8fb5d9a887\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-w7pv5" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.678290 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ctqts\" (UniqueName: \"kubernetes.io/projected/6d061f42-06de-43e5-b4d0-dbca09e48630-kube-api-access-ctqts\") pod \"marketplace-operator-79b997595-hp72h\" (UID: \"6d061f42-06de-43e5-b4d0-dbca09e48630\") " pod="openshift-marketplace/marketplace-operator-79b997595-hp72h" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.678314 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c8t7j\" (UniqueName: \"kubernetes.io/projected/a5bbba10-16da-4fce-ae5f-311fc5d0d6c1-kube-api-access-c8t7j\") pod \"downloads-7954f5f757-zbccb\" (UID: \"a5bbba10-16da-4fce-ae5f-311fc5d0d6c1\") " pod="openshift-console/downloads-7954f5f757-zbccb" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.678330 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c876r\" (UniqueName: \"kubernetes.io/projected/da284312-b98a-48d6-8cff-f8071c334e0b-kube-api-access-c876r\") pod \"console-f9d7485db-8wk54\" (UID: \"da284312-b98a-48d6-8cff-f8071c334e0b\") " pod="openshift-console/console-f9d7485db-8wk54" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.678346 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5wddd\" (UniqueName: \"kubernetes.io/projected/20e2f07c-3a73-4d78-88e9-c12b093743bb-kube-api-access-5wddd\") pod \"kube-storage-version-migrator-operator-b67b599dd-jzd5b\" (UID: \"20e2f07c-3a73-4d78-88e9-c12b093743bb\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-jzd5b" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.678363 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l9hk2\" (UniqueName: \"kubernetes.io/projected/3b5b5090-45a4-4acc-aa94-84c892bcf306-kube-api-access-l9hk2\") pod \"cluster-image-registry-operator-dc59b4c8b-f8qzz\" (UID: \"3b5b5090-45a4-4acc-aa94-84c892bcf306\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-f8qzz" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.678378 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/faae1773-bf96-4127-8060-a40da07930ce-default-certificate\") pod \"router-default-5444994796-2d7kh\" (UID: \"faae1773-bf96-4127-8060-a40da07930ce\") " pod="openshift-ingress/router-default-5444994796-2d7kh" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.678396 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/620a7a2f-92fc-4a76-900b-021d0dd78d79-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-96vl6\" (UID: \"620a7a2f-92fc-4a76-900b-021d0dd78d79\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-96vl6" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.678411 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/d9aa8f66-da03-49e5-b7b4-d17135ccaa03-audit-policies\") pod \"apiserver-7bbb656c7d-2rrfz\" (UID: \"d9aa8f66-da03-49e5-b7b4-d17135ccaa03\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2rrfz" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.678427 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/6ece12fd-d439-4104-a7a9-a9d174e29b5a-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-nkbg2\" (UID: \"6ece12fd-d439-4104-a7a9-a9d174e29b5a\") " pod="openshift-authentication/oauth-openshift-558db77b4-nkbg2" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.678443 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ca4c97ff-9032-48de-b286-89da26f5a822-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-jdng9\" (UID: \"ca4c97ff-9032-48de-b286-89da26f5a822\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-jdng9" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.678457 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d9aa8f66-da03-49e5-b7b4-d17135ccaa03-serving-cert\") pod \"apiserver-7bbb656c7d-2rrfz\" (UID: \"d9aa8f66-da03-49e5-b7b4-d17135ccaa03\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2rrfz" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.678472 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cpp6b\" (UniqueName: \"kubernetes.io/projected/ff5fcfe1-eb01-4f01-857d-a67712db3b87-kube-api-access-cpp6b\") pod \"multus-admission-controller-857f4d67dd-5wvhc\" (UID: \"ff5fcfe1-eb01-4f01-857d-a67712db3b87\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-5wvhc" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.678488 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9bb5eece-3596-485e-a859-b61d201b05f5-serving-cert\") pod \"openshift-config-operator-7777fb866f-z9zch\" (UID: \"9bb5eece-3596-485e-a859-b61d201b05f5\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-z9zch" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.678505 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/6ece12fd-d439-4104-a7a9-a9d174e29b5a-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-nkbg2\" (UID: \"6ece12fd-d439-4104-a7a9-a9d174e29b5a\") " pod="openshift-authentication/oauth-openshift-558db77b4-nkbg2" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.678520 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2l85j\" (UniqueName: \"kubernetes.io/projected/9bb5eece-3596-485e-a859-b61d201b05f5-kube-api-access-2l85j\") pod \"openshift-config-operator-7777fb866f-z9zch\" (UID: \"9bb5eece-3596-485e-a859-b61d201b05f5\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-z9zch" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.678534 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/42b49d65-d823-45af-b274-380dfbddc9b7-serving-cert\") pod \"route-controller-manager-6576b87f9c-spnfb\" (UID: \"42b49d65-d823-45af-b274-380dfbddc9b7\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-spnfb" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.678547 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/6d061f42-06de-43e5-b4d0-dbca09e48630-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-hp72h\" (UID: \"6d061f42-06de-43e5-b4d0-dbca09e48630\") " pod="openshift-marketplace/marketplace-operator-79b997595-hp72h" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.678562 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/20e2f07c-3a73-4d78-88e9-c12b093743bb-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-jzd5b\" (UID: \"20e2f07c-3a73-4d78-88e9-c12b093743bb\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-jzd5b" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.678577 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ca4c97ff-9032-48de-b286-89da26f5a822-serving-cert\") pod \"authentication-operator-69f744f599-jdng9\" (UID: \"ca4c97ff-9032-48de-b286-89da26f5a822\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-jdng9" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.678592 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dpltn\" (UniqueName: \"kubernetes.io/projected/ca4c97ff-9032-48de-b286-89da26f5a822-kube-api-access-dpltn\") pod \"authentication-operator-69f744f599-jdng9\" (UID: \"ca4c97ff-9032-48de-b286-89da26f5a822\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-jdng9" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.678607 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d9aa8f66-da03-49e5-b7b4-d17135ccaa03-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-2rrfz\" (UID: \"d9aa8f66-da03-49e5-b7b4-d17135ccaa03\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2rrfz" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.678621 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/20e2f07c-3a73-4d78-88e9-c12b093743bb-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-jzd5b\" (UID: \"20e2f07c-3a73-4d78-88e9-c12b093743bb\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-jzd5b" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.678639 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/094b001c-e80c-4d07-a161-e00514667462-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-q6h7h\" (UID: \"094b001c-e80c-4d07-a161-e00514667462\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-q6h7h" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.678655 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/19b65635-c86d-4439-8ece-4d8c3dbe986b-config\") pod \"console-operator-58897d9998-qvwzc\" (UID: \"19b65635-c86d-4439-8ece-4d8c3dbe986b\") " pod="openshift-console-operator/console-operator-58897d9998-qvwzc" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.678670 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zx4pb\" (UniqueName: \"kubernetes.io/projected/78554bd5-6fc8-4ce8-b1d2-b183dab17495-kube-api-access-zx4pb\") pod \"openshift-apiserver-operator-796bbdcf4f-2k66r\" (UID: \"78554bd5-6fc8-4ce8-b1d2-b183dab17495\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2k66r" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.678714 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ca4c97ff-9032-48de-b286-89da26f5a822-config\") pod \"authentication-operator-69f744f599-jdng9\" (UID: \"ca4c97ff-9032-48de-b286-89da26f5a822\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-jdng9" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.678729 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jplnj\" (UniqueName: \"kubernetes.io/projected/29ab8b79-a05e-4b94-a2bb-2c6b412fd297-kube-api-access-jplnj\") pod \"machine-approver-56656f9798-k6s2g\" (UID: \"29ab8b79-a05e-4b94-a2bb-2c6b412fd297\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-k6s2g" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.678747 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/6ece12fd-d439-4104-a7a9-a9d174e29b5a-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-nkbg2\" (UID: \"6ece12fd-d439-4104-a7a9-a9d174e29b5a\") " pod="openshift-authentication/oauth-openshift-558db77b4-nkbg2" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.678764 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/da284312-b98a-48d6-8cff-f8071c334e0b-console-oauth-config\") pod \"console-f9d7485db-8wk54\" (UID: \"da284312-b98a-48d6-8cff-f8071c334e0b\") " pod="openshift-console/console-f9d7485db-8wk54" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.678780 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tcxcq\" (UniqueName: \"kubernetes.io/projected/faae1773-bf96-4127-8060-a40da07930ce-kube-api-access-tcxcq\") pod \"router-default-5444994796-2d7kh\" (UID: \"faae1773-bf96-4127-8060-a40da07930ce\") " pod="openshift-ingress/router-default-5444994796-2d7kh" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.678796 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/8cc01026-2c76-4446-b78f-7d8fb5d9a887-auth-proxy-config\") pod \"machine-config-operator-74547568cd-w7pv5\" (UID: \"8cc01026-2c76-4446-b78f-7d8fb5d9a887\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-w7pv5" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.678811 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/3b5b5090-45a4-4acc-aa94-84c892bcf306-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-f8qzz\" (UID: \"3b5b5090-45a4-4acc-aa94-84c892bcf306\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-f8qzz" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.678827 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/6ece12fd-d439-4104-a7a9-a9d174e29b5a-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-nkbg2\" (UID: \"6ece12fd-d439-4104-a7a9-a9d174e29b5a\") " pod="openshift-authentication/oauth-openshift-558db77b4-nkbg2" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.678843 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/42b49d65-d823-45af-b274-380dfbddc9b7-config\") pod \"route-controller-manager-6576b87f9c-spnfb\" (UID: \"42b49d65-d823-45af-b274-380dfbddc9b7\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-spnfb" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.678866 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/094b001c-e80c-4d07-a161-e00514667462-config\") pod \"machine-api-operator-5694c8668f-q6h7h\" (UID: \"094b001c-e80c-4d07-a161-e00514667462\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-q6h7h" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.678882 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eb060975-a3f2-43f4-b4eb-9456a81c2f1f-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-m5b65\" (UID: \"eb060975-a3f2-43f4-b4eb-9456a81c2f1f\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-m5b65" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.678899 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gf2ls\" (UniqueName: \"kubernetes.io/projected/6ece12fd-d439-4104-a7a9-a9d174e29b5a-kube-api-access-gf2ls\") pod \"oauth-openshift-558db77b4-nkbg2\" (UID: \"6ece12fd-d439-4104-a7a9-a9d174e29b5a\") " pod="openshift-authentication/oauth-openshift-558db77b4-nkbg2" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.678921 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/da284312-b98a-48d6-8cff-f8071c334e0b-oauth-serving-cert\") pod \"console-f9d7485db-8wk54\" (UID: \"da284312-b98a-48d6-8cff-f8071c334e0b\") " pod="openshift-console/console-f9d7485db-8wk54" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.678938 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f566h\" (UniqueName: \"kubernetes.io/projected/42b49d65-d823-45af-b274-380dfbddc9b7-kube-api-access-f566h\") pod \"route-controller-manager-6576b87f9c-spnfb\" (UID: \"42b49d65-d823-45af-b274-380dfbddc9b7\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-spnfb" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.678963 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ca4c97ff-9032-48de-b286-89da26f5a822-service-ca-bundle\") pod \"authentication-operator-69f744f599-jdng9\" (UID: \"ca4c97ff-9032-48de-b286-89da26f5a822\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-jdng9" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.678978 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/6ece12fd-d439-4104-a7a9-a9d174e29b5a-audit-policies\") pod \"oauth-openshift-558db77b4-nkbg2\" (UID: \"6ece12fd-d439-4104-a7a9-a9d174e29b5a\") " pod="openshift-authentication/oauth-openshift-558db77b4-nkbg2" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.681768 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.682003 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.682139 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.682307 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.682506 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.682642 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.682528 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.673379 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.673399 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.673351 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.673463 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.673438 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.673536 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.673563 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.674491 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.673489 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.684153 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/094b001c-e80c-4d07-a161-e00514667462-images\") pod \"machine-api-operator-5694c8668f-q6h7h\" (UID: \"094b001c-e80c-4d07-a161-e00514667462\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-q6h7h" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.684288 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.684581 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6cz6v"] Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.685324 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6cz6v" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.686288 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.686701 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-kgkd4"] Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.687009 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-6q4q2"] Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.687335 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-6q4q2" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.687507 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-kgkd4" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.688569 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ca4c97ff-9032-48de-b286-89da26f5a822-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-jdng9\" (UID: \"ca4c97ff-9032-48de-b286-89da26f5a822\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-jdng9" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.689527 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.690921 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ca4c97ff-9032-48de-b286-89da26f5a822-config\") pod \"authentication-operator-69f744f599-jdng9\" (UID: \"ca4c97ff-9032-48de-b286-89da26f5a822\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-jdng9" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.691596 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eb060975-a3f2-43f4-b4eb-9456a81c2f1f-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-m5b65\" (UID: \"eb060975-a3f2-43f4-b4eb-9456a81c2f1f\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-m5b65" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.692238 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/094b001c-e80c-4d07-a161-e00514667462-config\") pod \"machine-api-operator-5694c8668f-q6h7h\" (UID: \"094b001c-e80c-4d07-a161-e00514667462\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-q6h7h" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.692368 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.695511 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.696182 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ca4c97ff-9032-48de-b286-89da26f5a822-serving-cert\") pod \"authentication-operator-69f744f599-jdng9\" (UID: \"ca4c97ff-9032-48de-b286-89da26f5a822\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-jdng9" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.696316 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/620a7a2f-92fc-4a76-900b-021d0dd78d79-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-96vl6\" (UID: \"620a7a2f-92fc-4a76-900b-021d0dd78d79\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-96vl6" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.696500 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.696548 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.697212 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/eb060975-a3f2-43f4-b4eb-9456a81c2f1f-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-m5b65\" (UID: \"eb060975-a3f2-43f4-b4eb-9456a81c2f1f\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-m5b65" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.710318 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.711480 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/094b001c-e80c-4d07-a161-e00514667462-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-q6h7h\" (UID: \"094b001c-e80c-4d07-a161-e00514667462\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-q6h7h" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.712460 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ca4c97ff-9032-48de-b286-89da26f5a822-service-ca-bundle\") pod \"authentication-operator-69f744f599-jdng9\" (UID: \"ca4c97ff-9032-48de-b286-89da26f5a822\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-jdng9" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.712521 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-ks964"] Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.713083 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-ks964" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.714020 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-gvv4x"] Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.714766 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-gvv4x" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.717031 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.736442 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.748939 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zpshm"] Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.752256 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-q6h7h"] Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.752760 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-zbccb"] Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.752841 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-96vl6"] Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.752771 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zpshm" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.755463 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-jdng9"] Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.757261 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.767008 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-m5b65"] Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.767993 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.769642 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-2rrfz"] Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.777445 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-nkbg2"] Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.779486 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/8cc01026-2c76-4446-b78f-7d8fb5d9a887-auth-proxy-config\") pod \"machine-config-operator-74547568cd-w7pv5\" (UID: \"8cc01026-2c76-4446-b78f-7d8fb5d9a887\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-w7pv5" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.779523 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/3b5b5090-45a4-4acc-aa94-84c892bcf306-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-f8qzz\" (UID: \"3b5b5090-45a4-4acc-aa94-84c892bcf306\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-f8qzz" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.779550 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/6ece12fd-d439-4104-a7a9-a9d174e29b5a-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-nkbg2\" (UID: \"6ece12fd-d439-4104-a7a9-a9d174e29b5a\") " pod="openshift-authentication/oauth-openshift-558db77b4-nkbg2" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.779571 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/42b49d65-d823-45af-b274-380dfbddc9b7-config\") pod \"route-controller-manager-6576b87f9c-spnfb\" (UID: \"42b49d65-d823-45af-b274-380dfbddc9b7\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-spnfb" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.779612 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gf2ls\" (UniqueName: \"kubernetes.io/projected/6ece12fd-d439-4104-a7a9-a9d174e29b5a-kube-api-access-gf2ls\") pod \"oauth-openshift-558db77b4-nkbg2\" (UID: \"6ece12fd-d439-4104-a7a9-a9d174e29b5a\") " pod="openshift-authentication/oauth-openshift-558db77b4-nkbg2" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.779636 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/da284312-b98a-48d6-8cff-f8071c334e0b-oauth-serving-cert\") pod \"console-f9d7485db-8wk54\" (UID: \"da284312-b98a-48d6-8cff-f8071c334e0b\") " pod="openshift-console/console-f9d7485db-8wk54" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.779659 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f566h\" (UniqueName: \"kubernetes.io/projected/42b49d65-d823-45af-b274-380dfbddc9b7-kube-api-access-f566h\") pod \"route-controller-manager-6576b87f9c-spnfb\" (UID: \"42b49d65-d823-45af-b274-380dfbddc9b7\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-spnfb" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.779683 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/6ece12fd-d439-4104-a7a9-a9d174e29b5a-audit-policies\") pod \"oauth-openshift-558db77b4-nkbg2\" (UID: \"6ece12fd-d439-4104-a7a9-a9d174e29b5a\") " pod="openshift-authentication/oauth-openshift-558db77b4-nkbg2" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.779755 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6ece12fd-d439-4104-a7a9-a9d174e29b5a-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-nkbg2\" (UID: \"6ece12fd-d439-4104-a7a9-a9d174e29b5a\") " pod="openshift-authentication/oauth-openshift-558db77b4-nkbg2" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.779790 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/ff5fcfe1-eb01-4f01-857d-a67712db3b87-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-5wvhc\" (UID: \"ff5fcfe1-eb01-4f01-857d-a67712db3b87\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-5wvhc" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.779820 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/19b65635-c86d-4439-8ece-4d8c3dbe986b-trusted-ca\") pod \"console-operator-58897d9998-qvwzc\" (UID: \"19b65635-c86d-4439-8ece-4d8c3dbe986b\") " pod="openshift-console-operator/console-operator-58897d9998-qvwzc" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.779858 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/6ece12fd-d439-4104-a7a9-a9d174e29b5a-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-nkbg2\" (UID: \"6ece12fd-d439-4104-a7a9-a9d174e29b5a\") " pod="openshift-authentication/oauth-openshift-558db77b4-nkbg2" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.779874 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/da284312-b98a-48d6-8cff-f8071c334e0b-console-config\") pod \"console-f9d7485db-8wk54\" (UID: \"da284312-b98a-48d6-8cff-f8071c334e0b\") " pod="openshift-console/console-f9d7485db-8wk54" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.779894 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/6ece12fd-d439-4104-a7a9-a9d174e29b5a-audit-dir\") pod \"oauth-openshift-558db77b4-nkbg2\" (UID: \"6ece12fd-d439-4104-a7a9-a9d174e29b5a\") " pod="openshift-authentication/oauth-openshift-558db77b4-nkbg2" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.779918 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/8cc01026-2c76-4446-b78f-7d8fb5d9a887-proxy-tls\") pod \"machine-config-operator-74547568cd-w7pv5\" (UID: \"8cc01026-2c76-4446-b78f-7d8fb5d9a887\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-w7pv5" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.779936 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w8p5f\" (UniqueName: \"kubernetes.io/projected/19b65635-c86d-4439-8ece-4d8c3dbe986b-kube-api-access-w8p5f\") pod \"console-operator-58897d9998-qvwzc\" (UID: \"19b65635-c86d-4439-8ece-4d8c3dbe986b\") " pod="openshift-console-operator/console-operator-58897d9998-qvwzc" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.779954 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-69bw7\" (UniqueName: \"kubernetes.io/projected/cc3f4dca-b740-4294-93e5-73906f8dc82c-kube-api-access-69bw7\") pod \"migrator-59844c95c7-g9hq6\" (UID: \"cc3f4dca-b740-4294-93e5-73906f8dc82c\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-g9hq6" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.779983 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/6ece12fd-d439-4104-a7a9-a9d174e29b5a-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-nkbg2\" (UID: \"6ece12fd-d439-4104-a7a9-a9d174e29b5a\") " pod="openshift-authentication/oauth-openshift-558db77b4-nkbg2" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.780001 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/da284312-b98a-48d6-8cff-f8071c334e0b-service-ca\") pod \"console-f9d7485db-8wk54\" (UID: \"da284312-b98a-48d6-8cff-f8071c334e0b\") " pod="openshift-console/console-f9d7485db-8wk54" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.780023 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/6ece12fd-d439-4104-a7a9-a9d174e29b5a-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-nkbg2\" (UID: \"6ece12fd-d439-4104-a7a9-a9d174e29b5a\") " pod="openshift-authentication/oauth-openshift-558db77b4-nkbg2" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.780054 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/6ece12fd-d439-4104-a7a9-a9d174e29b5a-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-nkbg2\" (UID: \"6ece12fd-d439-4104-a7a9-a9d174e29b5a\") " pod="openshift-authentication/oauth-openshift-558db77b4-nkbg2" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.780094 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/da284312-b98a-48d6-8cff-f8071c334e0b-trusted-ca-bundle\") pod \"console-f9d7485db-8wk54\" (UID: \"da284312-b98a-48d6-8cff-f8071c334e0b\") " pod="openshift-console/console-f9d7485db-8wk54" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.780122 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/6d061f42-06de-43e5-b4d0-dbca09e48630-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-hp72h\" (UID: \"6d061f42-06de-43e5-b4d0-dbca09e48630\") " pod="openshift-marketplace/marketplace-operator-79b997595-hp72h" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.780139 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/78554bd5-6fc8-4ce8-b1d2-b183dab17495-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-2k66r\" (UID: \"78554bd5-6fc8-4ce8-b1d2-b183dab17495\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2k66r" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.780160 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/9bb5eece-3596-485e-a859-b61d201b05f5-available-featuregates\") pod \"openshift-config-operator-7777fb866f-z9zch\" (UID: \"9bb5eece-3596-485e-a859-b61d201b05f5\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-z9zch" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.780190 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/19b65635-c86d-4439-8ece-4d8c3dbe986b-serving-cert\") pod \"console-operator-58897d9998-qvwzc\" (UID: \"19b65635-c86d-4439-8ece-4d8c3dbe986b\") " pod="openshift-console-operator/console-operator-58897d9998-qvwzc" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.780235 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/3b5b5090-45a4-4acc-aa94-84c892bcf306-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-f8qzz\" (UID: \"3b5b5090-45a4-4acc-aa94-84c892bcf306\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-f8qzz" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.780282 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/d9aa8f66-da03-49e5-b7b4-d17135ccaa03-etcd-client\") pod \"apiserver-7bbb656c7d-2rrfz\" (UID: \"d9aa8f66-da03-49e5-b7b4-d17135ccaa03\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2rrfz" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.780300 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/d9aa8f66-da03-49e5-b7b4-d17135ccaa03-audit-dir\") pod \"apiserver-7bbb656c7d-2rrfz\" (UID: \"d9aa8f66-da03-49e5-b7b4-d17135ccaa03\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2rrfz" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.780317 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/42b49d65-d823-45af-b274-380dfbddc9b7-client-ca\") pod \"route-controller-manager-6576b87f9c-spnfb\" (UID: \"42b49d65-d823-45af-b274-380dfbddc9b7\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-spnfb" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.780332 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/3b5b5090-45a4-4acc-aa94-84c892bcf306-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-f8qzz\" (UID: \"3b5b5090-45a4-4acc-aa94-84c892bcf306\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-f8qzz" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.780358 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/78554bd5-6fc8-4ce8-b1d2-b183dab17495-config\") pod \"openshift-apiserver-operator-796bbdcf4f-2k66r\" (UID: \"78554bd5-6fc8-4ce8-b1d2-b183dab17495\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2k66r" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.780384 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lp2b5\" (UniqueName: \"kubernetes.io/projected/8cc01026-2c76-4446-b78f-7d8fb5d9a887-kube-api-access-lp2b5\") pod \"machine-config-operator-74547568cd-w7pv5\" (UID: \"8cc01026-2c76-4446-b78f-7d8fb5d9a887\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-w7pv5" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.780402 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/faae1773-bf96-4127-8060-a40da07930ce-stats-auth\") pod \"router-default-5444994796-2d7kh\" (UID: \"faae1773-bf96-4127-8060-a40da07930ce\") " pod="openshift-ingress/router-default-5444994796-2d7kh" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.780417 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/6ece12fd-d439-4104-a7a9-a9d174e29b5a-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-nkbg2\" (UID: \"6ece12fd-d439-4104-a7a9-a9d174e29b5a\") " pod="openshift-authentication/oauth-openshift-558db77b4-nkbg2" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.780501 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/d9aa8f66-da03-49e5-b7b4-d17135ccaa03-encryption-config\") pod \"apiserver-7bbb656c7d-2rrfz\" (UID: \"d9aa8f66-da03-49e5-b7b4-d17135ccaa03\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2rrfz" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.780521 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/faae1773-bf96-4127-8060-a40da07930ce-metrics-certs\") pod \"router-default-5444994796-2d7kh\" (UID: \"faae1773-bf96-4127-8060-a40da07930ce\") " pod="openshift-ingress/router-default-5444994796-2d7kh" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.780539 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/da284312-b98a-48d6-8cff-f8071c334e0b-console-serving-cert\") pod \"console-f9d7485db-8wk54\" (UID: \"da284312-b98a-48d6-8cff-f8071c334e0b\") " pod="openshift-console/console-f9d7485db-8wk54" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.780574 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/d9aa8f66-da03-49e5-b7b4-d17135ccaa03-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-2rrfz\" (UID: \"d9aa8f66-da03-49e5-b7b4-d17135ccaa03\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2rrfz" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.780599 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cxrc9\" (UniqueName: \"kubernetes.io/projected/d9aa8f66-da03-49e5-b7b4-d17135ccaa03-kube-api-access-cxrc9\") pod \"apiserver-7bbb656c7d-2rrfz\" (UID: \"d9aa8f66-da03-49e5-b7b4-d17135ccaa03\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2rrfz" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.780637 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/8cc01026-2c76-4446-b78f-7d8fb5d9a887-images\") pod \"machine-config-operator-74547568cd-w7pv5\" (UID: \"8cc01026-2c76-4446-b78f-7d8fb5d9a887\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-w7pv5" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.780655 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ctqts\" (UniqueName: \"kubernetes.io/projected/6d061f42-06de-43e5-b4d0-dbca09e48630-kube-api-access-ctqts\") pod \"marketplace-operator-79b997595-hp72h\" (UID: \"6d061f42-06de-43e5-b4d0-dbca09e48630\") " pod="openshift-marketplace/marketplace-operator-79b997595-hp72h" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.780738 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/faae1773-bf96-4127-8060-a40da07930ce-service-ca-bundle\") pod \"router-default-5444994796-2d7kh\" (UID: \"faae1773-bf96-4127-8060-a40da07930ce\") " pod="openshift-ingress/router-default-5444994796-2d7kh" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.780756 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/6ece12fd-d439-4104-a7a9-a9d174e29b5a-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-nkbg2\" (UID: \"6ece12fd-d439-4104-a7a9-a9d174e29b5a\") " pod="openshift-authentication/oauth-openshift-558db77b4-nkbg2" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.780792 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c876r\" (UniqueName: \"kubernetes.io/projected/da284312-b98a-48d6-8cff-f8071c334e0b-kube-api-access-c876r\") pod \"console-f9d7485db-8wk54\" (UID: \"da284312-b98a-48d6-8cff-f8071c334e0b\") " pod="openshift-console/console-f9d7485db-8wk54" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.780809 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5wddd\" (UniqueName: \"kubernetes.io/projected/20e2f07c-3a73-4d78-88e9-c12b093743bb-kube-api-access-5wddd\") pod \"kube-storage-version-migrator-operator-b67b599dd-jzd5b\" (UID: \"20e2f07c-3a73-4d78-88e9-c12b093743bb\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-jzd5b" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.780846 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/d9aa8f66-da03-49e5-b7b4-d17135ccaa03-audit-policies\") pod \"apiserver-7bbb656c7d-2rrfz\" (UID: \"d9aa8f66-da03-49e5-b7b4-d17135ccaa03\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2rrfz" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.780864 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l9hk2\" (UniqueName: \"kubernetes.io/projected/3b5b5090-45a4-4acc-aa94-84c892bcf306-kube-api-access-l9hk2\") pod \"cluster-image-registry-operator-dc59b4c8b-f8qzz\" (UID: \"3b5b5090-45a4-4acc-aa94-84c892bcf306\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-f8qzz" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.780882 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/faae1773-bf96-4127-8060-a40da07930ce-default-certificate\") pod \"router-default-5444994796-2d7kh\" (UID: \"faae1773-bf96-4127-8060-a40da07930ce\") " pod="openshift-ingress/router-default-5444994796-2d7kh" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.780903 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/6ece12fd-d439-4104-a7a9-a9d174e29b5a-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-nkbg2\" (UID: \"6ece12fd-d439-4104-a7a9-a9d174e29b5a\") " pod="openshift-authentication/oauth-openshift-558db77b4-nkbg2" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.780944 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d9aa8f66-da03-49e5-b7b4-d17135ccaa03-serving-cert\") pod \"apiserver-7bbb656c7d-2rrfz\" (UID: \"d9aa8f66-da03-49e5-b7b4-d17135ccaa03\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2rrfz" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.781047 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/42b49d65-d823-45af-b274-380dfbddc9b7-config\") pod \"route-controller-manager-6576b87f9c-spnfb\" (UID: \"42b49d65-d823-45af-b274-380dfbddc9b7\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-spnfb" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.781083 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cpp6b\" (UniqueName: \"kubernetes.io/projected/ff5fcfe1-eb01-4f01-857d-a67712db3b87-kube-api-access-cpp6b\") pod \"multus-admission-controller-857f4d67dd-5wvhc\" (UID: \"ff5fcfe1-eb01-4f01-857d-a67712db3b87\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-5wvhc" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.781116 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9bb5eece-3596-485e-a859-b61d201b05f5-serving-cert\") pod \"openshift-config-operator-7777fb866f-z9zch\" (UID: \"9bb5eece-3596-485e-a859-b61d201b05f5\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-z9zch" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.781146 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/6ece12fd-d439-4104-a7a9-a9d174e29b5a-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-nkbg2\" (UID: \"6ece12fd-d439-4104-a7a9-a9d174e29b5a\") " pod="openshift-authentication/oauth-openshift-558db77b4-nkbg2" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.781175 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d9aa8f66-da03-49e5-b7b4-d17135ccaa03-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-2rrfz\" (UID: \"d9aa8f66-da03-49e5-b7b4-d17135ccaa03\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2rrfz" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.781202 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2l85j\" (UniqueName: \"kubernetes.io/projected/9bb5eece-3596-485e-a859-b61d201b05f5-kube-api-access-2l85j\") pod \"openshift-config-operator-7777fb866f-z9zch\" (UID: \"9bb5eece-3596-485e-a859-b61d201b05f5\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-z9zch" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.781219 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/42b49d65-d823-45af-b274-380dfbddc9b7-serving-cert\") pod \"route-controller-manager-6576b87f9c-spnfb\" (UID: \"42b49d65-d823-45af-b274-380dfbddc9b7\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-spnfb" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.781238 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/6d061f42-06de-43e5-b4d0-dbca09e48630-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-hp72h\" (UID: \"6d061f42-06de-43e5-b4d0-dbca09e48630\") " pod="openshift-marketplace/marketplace-operator-79b997595-hp72h" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.781268 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/20e2f07c-3a73-4d78-88e9-c12b093743bb-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-jzd5b\" (UID: \"20e2f07c-3a73-4d78-88e9-c12b093743bb\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-jzd5b" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.782184 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/8cc01026-2c76-4446-b78f-7d8fb5d9a887-auth-proxy-config\") pod \"machine-config-operator-74547568cd-w7pv5\" (UID: \"8cc01026-2c76-4446-b78f-7d8fb5d9a887\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-w7pv5" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.782600 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/3b5b5090-45a4-4acc-aa94-84c892bcf306-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-f8qzz\" (UID: \"3b5b5090-45a4-4acc-aa94-84c892bcf306\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-f8qzz" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.782830 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/78554bd5-6fc8-4ce8-b1d2-b183dab17495-config\") pod \"openshift-apiserver-operator-796bbdcf4f-2k66r\" (UID: \"78554bd5-6fc8-4ce8-b1d2-b183dab17495\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2k66r" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.781286 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/20e2f07c-3a73-4d78-88e9-c12b093743bb-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-jzd5b\" (UID: \"20e2f07c-3a73-4d78-88e9-c12b093743bb\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-jzd5b" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.782908 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/19b65635-c86d-4439-8ece-4d8c3dbe986b-config\") pod \"console-operator-58897d9998-qvwzc\" (UID: \"19b65635-c86d-4439-8ece-4d8c3dbe986b\") " pod="openshift-console-operator/console-operator-58897d9998-qvwzc" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.782930 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zx4pb\" (UniqueName: \"kubernetes.io/projected/78554bd5-6fc8-4ce8-b1d2-b183dab17495-kube-api-access-zx4pb\") pod \"openshift-apiserver-operator-796bbdcf4f-2k66r\" (UID: \"78554bd5-6fc8-4ce8-b1d2-b183dab17495\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2k66r" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.782972 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/6ece12fd-d439-4104-a7a9-a9d174e29b5a-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-nkbg2\" (UID: \"6ece12fd-d439-4104-a7a9-a9d174e29b5a\") " pod="openshift-authentication/oauth-openshift-558db77b4-nkbg2" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.783004 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/da284312-b98a-48d6-8cff-f8071c334e0b-console-oauth-config\") pod \"console-f9d7485db-8wk54\" (UID: \"da284312-b98a-48d6-8cff-f8071c334e0b\") " pod="openshift-console/console-f9d7485db-8wk54" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.783024 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tcxcq\" (UniqueName: \"kubernetes.io/projected/faae1773-bf96-4127-8060-a40da07930ce-kube-api-access-tcxcq\") pod \"router-default-5444994796-2d7kh\" (UID: \"faae1773-bf96-4127-8060-a40da07930ce\") " pod="openshift-ingress/router-default-5444994796-2d7kh" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.787898 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/6ece12fd-d439-4104-a7a9-a9d174e29b5a-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-nkbg2\" (UID: \"6ece12fd-d439-4104-a7a9-a9d174e29b5a\") " pod="openshift-authentication/oauth-openshift-558db77b4-nkbg2" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.788044 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/19b65635-c86d-4439-8ece-4d8c3dbe986b-config\") pod \"console-operator-58897d9998-qvwzc\" (UID: \"19b65635-c86d-4439-8ece-4d8c3dbe986b\") " pod="openshift-console-operator/console-operator-58897d9998-qvwzc" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.788397 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/d9aa8f66-da03-49e5-b7b4-d17135ccaa03-audit-policies\") pod \"apiserver-7bbb656c7d-2rrfz\" (UID: \"d9aa8f66-da03-49e5-b7b4-d17135ccaa03\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2rrfz" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.788967 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/6ece12fd-d439-4104-a7a9-a9d174e29b5a-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-nkbg2\" (UID: \"6ece12fd-d439-4104-a7a9-a9d174e29b5a\") " pod="openshift-authentication/oauth-openshift-558db77b4-nkbg2" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.789056 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.789390 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/da284312-b98a-48d6-8cff-f8071c334e0b-oauth-serving-cert\") pod \"console-f9d7485db-8wk54\" (UID: \"da284312-b98a-48d6-8cff-f8071c334e0b\") " pod="openshift-console/console-f9d7485db-8wk54" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.789803 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/19b65635-c86d-4439-8ece-4d8c3dbe986b-serving-cert\") pod \"console-operator-58897d9998-qvwzc\" (UID: \"19b65635-c86d-4439-8ece-4d8c3dbe986b\") " pod="openshift-console-operator/console-operator-58897d9998-qvwzc" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.789841 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-cw57r"] Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.791617 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/6ece12fd-d439-4104-a7a9-a9d174e29b5a-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-nkbg2\" (UID: \"6ece12fd-d439-4104-a7a9-a9d174e29b5a\") " pod="openshift-authentication/oauth-openshift-558db77b4-nkbg2" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.792054 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/d9aa8f66-da03-49e5-b7b4-d17135ccaa03-audit-dir\") pod \"apiserver-7bbb656c7d-2rrfz\" (UID: \"d9aa8f66-da03-49e5-b7b4-d17135ccaa03\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2rrfz" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.792077 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/8cc01026-2c76-4446-b78f-7d8fb5d9a887-images\") pod \"machine-config-operator-74547568cd-w7pv5\" (UID: \"8cc01026-2c76-4446-b78f-7d8fb5d9a887\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-w7pv5" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.792069 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-8wk54"] Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.792387 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/9bb5eece-3596-485e-a859-b61d201b05f5-available-featuregates\") pod \"openshift-config-operator-7777fb866f-z9zch\" (UID: \"9bb5eece-3596-485e-a859-b61d201b05f5\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-z9zch" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.792840 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/da284312-b98a-48d6-8cff-f8071c334e0b-trusted-ca-bundle\") pod \"console-f9d7485db-8wk54\" (UID: \"da284312-b98a-48d6-8cff-f8071c334e0b\") " pod="openshift-console/console-f9d7485db-8wk54" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.793138 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/da284312-b98a-48d6-8cff-f8071c334e0b-service-ca\") pod \"console-f9d7485db-8wk54\" (UID: \"da284312-b98a-48d6-8cff-f8071c334e0b\") " pod="openshift-console/console-f9d7485db-8wk54" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.793363 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/d9aa8f66-da03-49e5-b7b4-d17135ccaa03-etcd-client\") pod \"apiserver-7bbb656c7d-2rrfz\" (UID: \"d9aa8f66-da03-49e5-b7b4-d17135ccaa03\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2rrfz" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.794532 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-f8qzz"] Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.794834 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/6ece12fd-d439-4104-a7a9-a9d174e29b5a-audit-policies\") pod \"oauth-openshift-558db77b4-nkbg2\" (UID: \"6ece12fd-d439-4104-a7a9-a9d174e29b5a\") " pod="openshift-authentication/oauth-openshift-558db77b4-nkbg2" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.795038 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6ece12fd-d439-4104-a7a9-a9d174e29b5a-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-nkbg2\" (UID: \"6ece12fd-d439-4104-a7a9-a9d174e29b5a\") " pod="openshift-authentication/oauth-openshift-558db77b4-nkbg2" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.795147 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/19b65635-c86d-4439-8ece-4d8c3dbe986b-trusted-ca\") pod \"console-operator-58897d9998-qvwzc\" (UID: \"19b65635-c86d-4439-8ece-4d8c3dbe986b\") " pod="openshift-console-operator/console-operator-58897d9998-qvwzc" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.795250 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/d9aa8f66-da03-49e5-b7b4-d17135ccaa03-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-2rrfz\" (UID: \"d9aa8f66-da03-49e5-b7b4-d17135ccaa03\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2rrfz" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.795871 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d9aa8f66-da03-49e5-b7b4-d17135ccaa03-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-2rrfz\" (UID: \"d9aa8f66-da03-49e5-b7b4-d17135ccaa03\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2rrfz" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.796158 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9bb5eece-3596-485e-a859-b61d201b05f5-serving-cert\") pod \"openshift-config-operator-7777fb866f-z9zch\" (UID: \"9bb5eece-3596-485e-a859-b61d201b05f5\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-z9zch" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.796138 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/6ece12fd-d439-4104-a7a9-a9d174e29b5a-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-nkbg2\" (UID: \"6ece12fd-d439-4104-a7a9-a9d174e29b5a\") " pod="openshift-authentication/oauth-openshift-558db77b4-nkbg2" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.796224 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/6ece12fd-d439-4104-a7a9-a9d174e29b5a-audit-dir\") pod \"oauth-openshift-558db77b4-nkbg2\" (UID: \"6ece12fd-d439-4104-a7a9-a9d174e29b5a\") " pod="openshift-authentication/oauth-openshift-558db77b4-nkbg2" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.796386 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/6ece12fd-d439-4104-a7a9-a9d174e29b5a-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-nkbg2\" (UID: \"6ece12fd-d439-4104-a7a9-a9d174e29b5a\") " pod="openshift-authentication/oauth-openshift-558db77b4-nkbg2" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.796968 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/da284312-b98a-48d6-8cff-f8071c334e0b-console-oauth-config\") pod \"console-f9d7485db-8wk54\" (UID: \"da284312-b98a-48d6-8cff-f8071c334e0b\") " pod="openshift-console/console-f9d7485db-8wk54" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.797014 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bk49v"] Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.797034 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-kbtxf"] Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.797173 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/6ece12fd-d439-4104-a7a9-a9d174e29b5a-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-nkbg2\" (UID: \"6ece12fd-d439-4104-a7a9-a9d174e29b5a\") " pod="openshift-authentication/oauth-openshift-558db77b4-nkbg2" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.797496 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/78554bd5-6fc8-4ce8-b1d2-b183dab17495-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-2k66r\" (UID: \"78554bd5-6fc8-4ce8-b1d2-b183dab17495\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2k66r" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.797765 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/42b49d65-d823-45af-b274-380dfbddc9b7-client-ca\") pod \"route-controller-manager-6576b87f9c-spnfb\" (UID: \"42b49d65-d823-45af-b274-380dfbddc9b7\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-spnfb" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.797989 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-z9zch"] Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.798018 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/6ece12fd-d439-4104-a7a9-a9d174e29b5a-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-nkbg2\" (UID: \"6ece12fd-d439-4104-a7a9-a9d174e29b5a\") " pod="openshift-authentication/oauth-openshift-558db77b4-nkbg2" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.798214 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/6ece12fd-d439-4104-a7a9-a9d174e29b5a-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-nkbg2\" (UID: \"6ece12fd-d439-4104-a7a9-a9d174e29b5a\") " pod="openshift-authentication/oauth-openshift-558db77b4-nkbg2" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.798377 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/da284312-b98a-48d6-8cff-f8071c334e0b-console-config\") pod \"console-f9d7485db-8wk54\" (UID: \"da284312-b98a-48d6-8cff-f8071c334e0b\") " pod="openshift-console/console-f9d7485db-8wk54" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.798670 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/42b49d65-d823-45af-b274-380dfbddc9b7-serving-cert\") pod \"route-controller-manager-6576b87f9c-spnfb\" (UID: \"42b49d65-d823-45af-b274-380dfbddc9b7\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-spnfb" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.799338 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-hbsll"] Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.800176 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-hbsll" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.800598 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-hp72h"] Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.801198 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/d9aa8f66-da03-49e5-b7b4-d17135ccaa03-encryption-config\") pod \"apiserver-7bbb656c7d-2rrfz\" (UID: \"d9aa8f66-da03-49e5-b7b4-d17135ccaa03\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2rrfz" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.801664 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-qvwzc"] Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.802149 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/8cc01026-2c76-4446-b78f-7d8fb5d9a887-proxy-tls\") pod \"machine-config-operator-74547568cd-w7pv5\" (UID: \"8cc01026-2c76-4446-b78f-7d8fb5d9a887\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-w7pv5" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.802625 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-kv4nc"] Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.803607 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-kv4nc" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.803616 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-tcqv8"] Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.804712 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-42tqr"] Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.804764 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-tcqv8" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.804887 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/da284312-b98a-48d6-8cff-f8071c334e0b-console-serving-cert\") pod \"console-f9d7485db-8wk54\" (UID: \"da284312-b98a-48d6-8cff-f8071c334e0b\") " pod="openshift-console/console-f9d7485db-8wk54" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.805616 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-g9hq6"] Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.807143 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-spnfb"] Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.808093 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-sxt8c"] Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.809138 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6cz6v"] Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.809275 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/6ece12fd-d439-4104-a7a9-a9d174e29b5a-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-nkbg2\" (UID: \"6ece12fd-d439-4104-a7a9-a9d174e29b5a\") " pod="openshift-authentication/oauth-openshift-558db77b4-nkbg2" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.810129 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zpshm"] Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.810162 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.812063 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-gcfmk"] Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.812089 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-jzd5b"] Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.813544 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-c6sdx"] Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.814828 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-w7pv5"] Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.814870 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-pdjlb"] Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.816822 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d9aa8f66-da03-49e5-b7b4-d17135ccaa03-serving-cert\") pod \"apiserver-7bbb656c7d-2rrfz\" (UID: \"d9aa8f66-da03-49e5-b7b4-d17135ccaa03\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2rrfz" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.816872 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-5wvhc"] Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.818302 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-4lpn2"] Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.820222 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/cni-sysctl-allowlist-ds-v589k"] Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.820960 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2k66r"] Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.821029 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/cni-sysctl-allowlist-ds-v589k" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.821547 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-8l852"] Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.824681 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-ks964"] Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.824727 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-6q4q2"] Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.826302 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-kgkd4"] Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.827469 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.827468 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/6ece12fd-d439-4104-a7a9-a9d174e29b5a-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-nkbg2\" (UID: \"6ece12fd-d439-4104-a7a9-a9d174e29b5a\") " pod="openshift-authentication/oauth-openshift-558db77b4-nkbg2" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.828055 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-kv4nc"] Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.829225 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-gvv4x"] Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.830178 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-ghtdk"] Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.831108 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29536305-vmtnx"] Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.832459 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-tcqv8"] Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.832980 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-n5jbm"] Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.833947 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-n5jbm"] Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.834025 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-n5jbm" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.848200 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.867807 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.887308 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.907985 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.928727 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.938626 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/6d061f42-06de-43e5-b4d0-dbca09e48630-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-hp72h\" (UID: \"6d061f42-06de-43e5-b4d0-dbca09e48630\") " pod="openshift-marketplace/marketplace-operator-79b997595-hp72h" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.954971 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.957330 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/6d061f42-06de-43e5-b4d0-dbca09e48630-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-hp72h\" (UID: \"6d061f42-06de-43e5-b4d0-dbca09e48630\") " pod="openshift-marketplace/marketplace-operator-79b997595-hp72h" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.968152 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Feb 27 07:50:58 crc kubenswrapper[4612]: I0227 07:50:58.987502 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Feb 27 07:50:59 crc kubenswrapper[4612]: I0227 07:50:59.007824 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Feb 27 07:50:59 crc kubenswrapper[4612]: I0227 07:50:59.028640 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Feb 27 07:50:59 crc kubenswrapper[4612]: I0227 07:50:59.037397 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/faae1773-bf96-4127-8060-a40da07930ce-metrics-certs\") pod \"router-default-5444994796-2d7kh\" (UID: \"faae1773-bf96-4127-8060-a40da07930ce\") " pod="openshift-ingress/router-default-5444994796-2d7kh" Feb 27 07:50:59 crc kubenswrapper[4612]: I0227 07:50:59.048374 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Feb 27 07:50:59 crc kubenswrapper[4612]: I0227 07:50:59.055881 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/faae1773-bf96-4127-8060-a40da07930ce-default-certificate\") pod \"router-default-5444994796-2d7kh\" (UID: \"faae1773-bf96-4127-8060-a40da07930ce\") " pod="openshift-ingress/router-default-5444994796-2d7kh" Feb 27 07:50:59 crc kubenswrapper[4612]: I0227 07:50:59.068562 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Feb 27 07:50:59 crc kubenswrapper[4612]: I0227 07:50:59.075815 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/faae1773-bf96-4127-8060-a40da07930ce-stats-auth\") pod \"router-default-5444994796-2d7kh\" (UID: \"faae1773-bf96-4127-8060-a40da07930ce\") " pod="openshift-ingress/router-default-5444994796-2d7kh" Feb 27 07:50:59 crc kubenswrapper[4612]: I0227 07:50:59.088409 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Feb 27 07:50:59 crc kubenswrapper[4612]: I0227 07:50:59.108659 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Feb 27 07:50:59 crc kubenswrapper[4612]: I0227 07:50:59.128908 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Feb 27 07:50:59 crc kubenswrapper[4612]: I0227 07:50:59.134252 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/faae1773-bf96-4127-8060-a40da07930ce-service-ca-bundle\") pod \"router-default-5444994796-2d7kh\" (UID: \"faae1773-bf96-4127-8060-a40da07930ce\") " pod="openshift-ingress/router-default-5444994796-2d7kh" Feb 27 07:50:59 crc kubenswrapper[4612]: I0227 07:50:59.148304 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Feb 27 07:50:59 crc kubenswrapper[4612]: I0227 07:50:59.168253 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Feb 27 07:50:59 crc kubenswrapper[4612]: I0227 07:50:59.187957 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Feb 27 07:50:59 crc kubenswrapper[4612]: I0227 07:50:59.207923 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Feb 27 07:50:59 crc kubenswrapper[4612]: I0227 07:50:59.217289 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/20e2f07c-3a73-4d78-88e9-c12b093743bb-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-jzd5b\" (UID: \"20e2f07c-3a73-4d78-88e9-c12b093743bb\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-jzd5b" Feb 27 07:50:59 crc kubenswrapper[4612]: I0227 07:50:59.228283 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Feb 27 07:50:59 crc kubenswrapper[4612]: I0227 07:50:59.233830 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/20e2f07c-3a73-4d78-88e9-c12b093743bb-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-jzd5b\" (UID: \"20e2f07c-3a73-4d78-88e9-c12b093743bb\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-jzd5b" Feb 27 07:50:59 crc kubenswrapper[4612]: I0227 07:50:59.248327 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Feb 27 07:50:59 crc kubenswrapper[4612]: I0227 07:50:59.267736 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Feb 27 07:50:59 crc kubenswrapper[4612]: I0227 07:50:59.288288 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Feb 27 07:50:59 crc kubenswrapper[4612]: I0227 07:50:59.307924 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Feb 27 07:50:59 crc kubenswrapper[4612]: I0227 07:50:59.328760 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Feb 27 07:50:59 crc kubenswrapper[4612]: I0227 07:50:59.348489 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Feb 27 07:50:59 crc kubenswrapper[4612]: I0227 07:50:59.369048 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Feb 27 07:50:59 crc kubenswrapper[4612]: I0227 07:50:59.389121 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Feb 27 07:50:59 crc kubenswrapper[4612]: I0227 07:50:59.397911 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/3b5b5090-45a4-4acc-aa94-84c892bcf306-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-f8qzz\" (UID: \"3b5b5090-45a4-4acc-aa94-84c892bcf306\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-f8qzz" Feb 27 07:50:59 crc kubenswrapper[4612]: I0227 07:50:59.428410 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Feb 27 07:50:59 crc kubenswrapper[4612]: I0227 07:50:59.448573 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Feb 27 07:50:59 crc kubenswrapper[4612]: I0227 07:50:59.469122 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Feb 27 07:50:59 crc kubenswrapper[4612]: I0227 07:50:59.488287 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Feb 27 07:50:59 crc kubenswrapper[4612]: I0227 07:50:59.508675 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Feb 27 07:50:59 crc kubenswrapper[4612]: I0227 07:50:59.528184 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Feb 27 07:50:59 crc kubenswrapper[4612]: I0227 07:50:59.548178 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Feb 27 07:50:59 crc kubenswrapper[4612]: I0227 07:50:59.569354 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Feb 27 07:50:59 crc kubenswrapper[4612]: I0227 07:50:59.588614 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Feb 27 07:50:59 crc kubenswrapper[4612]: I0227 07:50:59.600511 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/ff5fcfe1-eb01-4f01-857d-a67712db3b87-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-5wvhc\" (UID: \"ff5fcfe1-eb01-4f01-857d-a67712db3b87\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-5wvhc" Feb 27 07:50:59 crc kubenswrapper[4612]: I0227 07:50:59.609169 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Feb 27 07:50:59 crc kubenswrapper[4612]: I0227 07:50:59.629104 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Feb 27 07:50:59 crc kubenswrapper[4612]: I0227 07:50:59.648554 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Feb 27 07:50:59 crc kubenswrapper[4612]: I0227 07:50:59.669667 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Feb 27 07:50:59 crc kubenswrapper[4612]: E0227 07:50:59.684763 4612 configmap.go:193] Couldn't get configMap openshift-cluster-machine-approver/kube-rbac-proxy: failed to sync configmap cache: timed out waiting for the condition Feb 27 07:50:59 crc kubenswrapper[4612]: E0227 07:50:59.685115 4612 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/29ab8b79-a05e-4b94-a2bb-2c6b412fd297-auth-proxy-config podName:29ab8b79-a05e-4b94-a2bb-2c6b412fd297 nodeName:}" failed. No retries permitted until 2026-02-27 07:51:00.18507875 +0000 UTC m=+118.039008788 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "auth-proxy-config" (UniqueName: "kubernetes.io/configmap/29ab8b79-a05e-4b94-a2bb-2c6b412fd297-auth-proxy-config") pod "machine-approver-56656f9798-k6s2g" (UID: "29ab8b79-a05e-4b94-a2bb-2c6b412fd297") : failed to sync configmap cache: timed out waiting for the condition Feb 27 07:50:59 crc kubenswrapper[4612]: E0227 07:50:59.684804 4612 secret.go:188] Couldn't get secret openshift-cluster-machine-approver/machine-approver-tls: failed to sync secret cache: timed out waiting for the condition Feb 27 07:50:59 crc kubenswrapper[4612]: E0227 07:50:59.684844 4612 configmap.go:193] Couldn't get configMap openshift-cluster-machine-approver/machine-approver-config: failed to sync configmap cache: timed out waiting for the condition Feb 27 07:50:59 crc kubenswrapper[4612]: E0227 07:50:59.685660 4612 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/29ab8b79-a05e-4b94-a2bb-2c6b412fd297-config podName:29ab8b79-a05e-4b94-a2bb-2c6b412fd297 nodeName:}" failed. No retries permitted until 2026-02-27 07:51:00.185629315 +0000 UTC m=+118.039559363 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config" (UniqueName: "kubernetes.io/configmap/29ab8b79-a05e-4b94-a2bb-2c6b412fd297-config") pod "machine-approver-56656f9798-k6s2g" (UID: "29ab8b79-a05e-4b94-a2bb-2c6b412fd297") : failed to sync configmap cache: timed out waiting for the condition Feb 27 07:50:59 crc kubenswrapper[4612]: E0227 07:50:59.685771 4612 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/29ab8b79-a05e-4b94-a2bb-2c6b412fd297-machine-approver-tls podName:29ab8b79-a05e-4b94-a2bb-2c6b412fd297 nodeName:}" failed. No retries permitted until 2026-02-27 07:51:00.185747648 +0000 UTC m=+118.039677686 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "machine-approver-tls" (UniqueName: "kubernetes.io/secret/29ab8b79-a05e-4b94-a2bb-2c6b412fd297-machine-approver-tls") pod "machine-approver-56656f9798-k6s2g" (UID: "29ab8b79-a05e-4b94-a2bb-2c6b412fd297") : failed to sync secret cache: timed out waiting for the condition Feb 27 07:50:59 crc kubenswrapper[4612]: I0227 07:50:59.686970 4612 request.go:700] Waited for 1.0102851s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-ingress-operator/secrets?fieldSelector=metadata.name%3Dingress-operator-dockercfg-7lnqk&limit=500&resourceVersion=0 Feb 27 07:50:59 crc kubenswrapper[4612]: I0227 07:50:59.689183 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Feb 27 07:50:59 crc kubenswrapper[4612]: I0227 07:50:59.709583 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Feb 27 07:50:59 crc kubenswrapper[4612]: I0227 07:50:59.741184 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Feb 27 07:50:59 crc kubenswrapper[4612]: I0227 07:50:59.749173 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Feb 27 07:50:59 crc kubenswrapper[4612]: I0227 07:50:59.768031 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Feb 27 07:50:59 crc kubenswrapper[4612]: I0227 07:50:59.788786 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Feb 27 07:50:59 crc kubenswrapper[4612]: I0227 07:50:59.808516 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Feb 27 07:50:59 crc kubenswrapper[4612]: I0227 07:50:59.828776 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Feb 27 07:50:59 crc kubenswrapper[4612]: I0227 07:50:59.851906 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 07:50:59 crc kubenswrapper[4612]: I0227 07:50:59.852609 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 07:50:59 crc kubenswrapper[4612]: I0227 07:50:59.852861 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-bqh6n" Feb 27 07:50:59 crc kubenswrapper[4612]: I0227 07:50:59.853118 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 07:50:59 crc kubenswrapper[4612]: I0227 07:50:59.863603 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v7d4b\" (UniqueName: \"kubernetes.io/projected/eb060975-a3f2-43f4-b4eb-9456a81c2f1f-kube-api-access-v7d4b\") pod \"openshift-controller-manager-operator-756b6f6bc6-m5b65\" (UID: \"eb060975-a3f2-43f4-b4eb-9456a81c2f1f\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-m5b65" Feb 27 07:50:59 crc kubenswrapper[4612]: I0227 07:50:59.941760 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c8t7j\" (UniqueName: \"kubernetes.io/projected/a5bbba10-16da-4fce-ae5f-311fc5d0d6c1-kube-api-access-c8t7j\") pod \"downloads-7954f5f757-zbccb\" (UID: \"a5bbba10-16da-4fce-ae5f-311fc5d0d6c1\") " pod="openshift-console/downloads-7954f5f757-zbccb" Feb 27 07:50:59 crc kubenswrapper[4612]: I0227 07:50:59.949598 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Feb 27 07:50:59 crc kubenswrapper[4612]: I0227 07:50:59.967720 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Feb 27 07:50:59 crc kubenswrapper[4612]: I0227 07:50:59.988097 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Feb 27 07:51:00 crc kubenswrapper[4612]: I0227 07:51:00.007271 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-m5b65" Feb 27 07:51:00 crc kubenswrapper[4612]: I0227 07:51:00.008404 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Feb 27 07:51:00 crc kubenswrapper[4612]: I0227 07:51:00.025454 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-zbccb" Feb 27 07:51:00 crc kubenswrapper[4612]: I0227 07:51:00.030514 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Feb 27 07:51:00 crc kubenswrapper[4612]: I0227 07:51:00.049120 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Feb 27 07:51:00 crc kubenswrapper[4612]: I0227 07:51:00.068506 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Feb 27 07:51:00 crc kubenswrapper[4612]: I0227 07:51:00.089369 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Feb 27 07:51:00 crc kubenswrapper[4612]: I0227 07:51:00.108595 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Feb 27 07:51:00 crc kubenswrapper[4612]: I0227 07:51:00.128863 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Feb 27 07:51:00 crc kubenswrapper[4612]: I0227 07:51:00.166425 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Feb 27 07:51:00 crc kubenswrapper[4612]: I0227 07:51:00.167276 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Feb 27 07:51:00 crc kubenswrapper[4612]: I0227 07:51:00.189104 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Feb 27 07:51:00 crc kubenswrapper[4612]: I0227 07:51:00.206976 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/29ab8b79-a05e-4b94-a2bb-2c6b412fd297-machine-approver-tls\") pod \"machine-approver-56656f9798-k6s2g\" (UID: \"29ab8b79-a05e-4b94-a2bb-2c6b412fd297\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-k6s2g" Feb 27 07:51:00 crc kubenswrapper[4612]: I0227 07:51:00.207027 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/29ab8b79-a05e-4b94-a2bb-2c6b412fd297-config\") pod \"machine-approver-56656f9798-k6s2g\" (UID: \"29ab8b79-a05e-4b94-a2bb-2c6b412fd297\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-k6s2g" Feb 27 07:51:00 crc kubenswrapper[4612]: I0227 07:51:00.207051 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/29ab8b79-a05e-4b94-a2bb-2c6b412fd297-auth-proxy-config\") pod \"machine-approver-56656f9798-k6s2g\" (UID: \"29ab8b79-a05e-4b94-a2bb-2c6b412fd297\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-k6s2g" Feb 27 07:51:00 crc kubenswrapper[4612]: I0227 07:51:00.249503 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Feb 27 07:51:00 crc kubenswrapper[4612]: I0227 07:51:00.251164 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dpltn\" (UniqueName: \"kubernetes.io/projected/ca4c97ff-9032-48de-b286-89da26f5a822-kube-api-access-dpltn\") pod \"authentication-operator-69f744f599-jdng9\" (UID: \"ca4c97ff-9032-48de-b286-89da26f5a822\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-jdng9" Feb 27 07:51:00 crc kubenswrapper[4612]: I0227 07:51:00.258402 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-zbccb"] Feb 27 07:51:00 crc kubenswrapper[4612]: W0227 07:51:00.267702 4612 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda5bbba10_16da_4fce_ae5f_311fc5d0d6c1.slice/crio-fa88fb80c55eade2fe160566bab6cbecf609a7e40e978cb956bcc7626aa2c3ee WatchSource:0}: Error finding container fa88fb80c55eade2fe160566bab6cbecf609a7e40e978cb956bcc7626aa2c3ee: Status 404 returned error can't find the container with id fa88fb80c55eade2fe160566bab6cbecf609a7e40e978cb956bcc7626aa2c3ee Feb 27 07:51:00 crc kubenswrapper[4612]: I0227 07:51:00.271780 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Feb 27 07:51:00 crc kubenswrapper[4612]: I0227 07:51:00.273533 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-m5b65"] Feb 27 07:51:00 crc kubenswrapper[4612]: W0227 07:51:00.281557 4612 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podeb060975_a3f2_43f4_b4eb_9456a81c2f1f.slice/crio-d47c89fc3f730f29e5f37b4689213fb34a12bbaffad1d7016a09a759cdadf329 WatchSource:0}: Error finding container d47c89fc3f730f29e5f37b4689213fb34a12bbaffad1d7016a09a759cdadf329: Status 404 returned error can't find the container with id d47c89fc3f730f29e5f37b4689213fb34a12bbaffad1d7016a09a759cdadf329 Feb 27 07:51:00 crc kubenswrapper[4612]: I0227 07:51:00.288277 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Feb 27 07:51:00 crc kubenswrapper[4612]: I0227 07:51:00.307877 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Feb 27 07:51:00 crc kubenswrapper[4612]: I0227 07:51:00.316832 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-jdng9" Feb 27 07:51:00 crc kubenswrapper[4612]: I0227 07:51:00.328808 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Feb 27 07:51:00 crc kubenswrapper[4612]: I0227 07:51:00.348271 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Feb 27 07:51:00 crc kubenswrapper[4612]: I0227 07:51:00.368191 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Feb 27 07:51:00 crc kubenswrapper[4612]: I0227 07:51:00.403789 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5wddd\" (UniqueName: \"kubernetes.io/projected/20e2f07c-3a73-4d78-88e9-c12b093743bb-kube-api-access-5wddd\") pod \"kube-storage-version-migrator-operator-b67b599dd-jzd5b\" (UID: \"20e2f07c-3a73-4d78-88e9-c12b093743bb\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-jzd5b" Feb 27 07:51:00 crc kubenswrapper[4612]: I0227 07:51:00.412381 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-zbccb" event={"ID":"a5bbba10-16da-4fce-ae5f-311fc5d0d6c1","Type":"ContainerStarted","Data":"2c814b65d3e326212218ab610ff7c97560147804cade7350fa070ea5112594c1"} Feb 27 07:51:00 crc kubenswrapper[4612]: I0227 07:51:00.412414 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-zbccb" event={"ID":"a5bbba10-16da-4fce-ae5f-311fc5d0d6c1","Type":"ContainerStarted","Data":"fa88fb80c55eade2fe160566bab6cbecf609a7e40e978cb956bcc7626aa2c3ee"} Feb 27 07:51:00 crc kubenswrapper[4612]: I0227 07:51:00.413060 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-zbccb" Feb 27 07:51:00 crc kubenswrapper[4612]: I0227 07:51:00.414189 4612 patch_prober.go:28] interesting pod/downloads-7954f5f757-zbccb container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.9:8080/\": dial tcp 10.217.0.9:8080: connect: connection refused" start-of-body= Feb 27 07:51:00 crc kubenswrapper[4612]: I0227 07:51:00.414222 4612 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-zbccb" podUID="a5bbba10-16da-4fce-ae5f-311fc5d0d6c1" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.9:8080/\": dial tcp 10.217.0.9:8080: connect: connection refused" Feb 27 07:51:00 crc kubenswrapper[4612]: I0227 07:51:00.414767 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-m5b65" event={"ID":"eb060975-a3f2-43f4-b4eb-9456a81c2f1f","Type":"ContainerStarted","Data":"5f07c67e4a1a1f9af5be23511e39a76ca7428532fd14e8a9edbeaae3e78ebb28"} Feb 27 07:51:00 crc kubenswrapper[4612]: I0227 07:51:00.414791 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-m5b65" event={"ID":"eb060975-a3f2-43f4-b4eb-9456a81c2f1f","Type":"ContainerStarted","Data":"d47c89fc3f730f29e5f37b4689213fb34a12bbaffad1d7016a09a759cdadf329"} Feb 27 07:51:00 crc kubenswrapper[4612]: I0227 07:51:00.421163 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lp2b5\" (UniqueName: \"kubernetes.io/projected/8cc01026-2c76-4446-b78f-7d8fb5d9a887-kube-api-access-lp2b5\") pod \"machine-config-operator-74547568cd-w7pv5\" (UID: \"8cc01026-2c76-4446-b78f-7d8fb5d9a887\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-w7pv5" Feb 27 07:51:00 crc kubenswrapper[4612]: I0227 07:51:00.445411 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gf2ls\" (UniqueName: \"kubernetes.io/projected/6ece12fd-d439-4104-a7a9-a9d174e29b5a-kube-api-access-gf2ls\") pod \"oauth-openshift-558db77b4-nkbg2\" (UID: \"6ece12fd-d439-4104-a7a9-a9d174e29b5a\") " pod="openshift-authentication/oauth-openshift-558db77b4-nkbg2" Feb 27 07:51:00 crc kubenswrapper[4612]: I0227 07:51:00.463345 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/3b5b5090-45a4-4acc-aa94-84c892bcf306-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-f8qzz\" (UID: \"3b5b5090-45a4-4acc-aa94-84c892bcf306\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-f8qzz" Feb 27 07:51:00 crc kubenswrapper[4612]: I0227 07:51:00.480973 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tcxcq\" (UniqueName: \"kubernetes.io/projected/faae1773-bf96-4127-8060-a40da07930ce-kube-api-access-tcxcq\") pod \"router-default-5444994796-2d7kh\" (UID: \"faae1773-bf96-4127-8060-a40da07930ce\") " pod="openshift-ingress/router-default-5444994796-2d7kh" Feb 27 07:51:00 crc kubenswrapper[4612]: I0227 07:51:00.499795 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cpp6b\" (UniqueName: \"kubernetes.io/projected/ff5fcfe1-eb01-4f01-857d-a67712db3b87-kube-api-access-cpp6b\") pod \"multus-admission-controller-857f4d67dd-5wvhc\" (UID: \"ff5fcfe1-eb01-4f01-857d-a67712db3b87\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-5wvhc" Feb 27 07:51:00 crc kubenswrapper[4612]: I0227 07:51:00.527001 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zx4pb\" (UniqueName: \"kubernetes.io/projected/78554bd5-6fc8-4ce8-b1d2-b183dab17495-kube-api-access-zx4pb\") pod \"openshift-apiserver-operator-796bbdcf4f-2k66r\" (UID: \"78554bd5-6fc8-4ce8-b1d2-b183dab17495\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2k66r" Feb 27 07:51:00 crc kubenswrapper[4612]: I0227 07:51:00.549393 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2k66r" Feb 27 07:51:00 crc kubenswrapper[4612]: I0227 07:51:00.552441 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2l85j\" (UniqueName: \"kubernetes.io/projected/9bb5eece-3596-485e-a859-b61d201b05f5-kube-api-access-2l85j\") pod \"openshift-config-operator-7777fb866f-z9zch\" (UID: \"9bb5eece-3596-485e-a859-b61d201b05f5\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-z9zch" Feb 27 07:51:00 crc kubenswrapper[4612]: I0227 07:51:00.562880 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l9hk2\" (UniqueName: \"kubernetes.io/projected/3b5b5090-45a4-4acc-aa94-84c892bcf306-kube-api-access-l9hk2\") pod \"cluster-image-registry-operator-dc59b4c8b-f8qzz\" (UID: \"3b5b5090-45a4-4acc-aa94-84c892bcf306\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-f8qzz" Feb 27 07:51:00 crc kubenswrapper[4612]: I0227 07:51:00.584819 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-w7pv5" Feb 27 07:51:00 crc kubenswrapper[4612]: I0227 07:51:00.593102 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f566h\" (UniqueName: \"kubernetes.io/projected/42b49d65-d823-45af-b274-380dfbddc9b7-kube-api-access-f566h\") pod \"route-controller-manager-6576b87f9c-spnfb\" (UID: \"42b49d65-d823-45af-b274-380dfbddc9b7\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-spnfb" Feb 27 07:51:00 crc kubenswrapper[4612]: I0227 07:51:00.619723 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cxrc9\" (UniqueName: \"kubernetes.io/projected/d9aa8f66-da03-49e5-b7b4-d17135ccaa03-kube-api-access-cxrc9\") pod \"apiserver-7bbb656c7d-2rrfz\" (UID: \"d9aa8f66-da03-49e5-b7b4-d17135ccaa03\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2rrfz" Feb 27 07:51:00 crc kubenswrapper[4612]: I0227 07:51:00.629361 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ctqts\" (UniqueName: \"kubernetes.io/projected/6d061f42-06de-43e5-b4d0-dbca09e48630-kube-api-access-ctqts\") pod \"marketplace-operator-79b997595-hp72h\" (UID: \"6d061f42-06de-43e5-b4d0-dbca09e48630\") " pod="openshift-marketplace/marketplace-operator-79b997595-hp72h" Feb 27 07:51:00 crc kubenswrapper[4612]: I0227 07:51:00.634105 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-2d7kh" Feb 27 07:51:00 crc kubenswrapper[4612]: I0227 07:51:00.642429 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c876r\" (UniqueName: \"kubernetes.io/projected/da284312-b98a-48d6-8cff-f8071c334e0b-kube-api-access-c876r\") pod \"console-f9d7485db-8wk54\" (UID: \"da284312-b98a-48d6-8cff-f8071c334e0b\") " pod="openshift-console/console-f9d7485db-8wk54" Feb 27 07:51:00 crc kubenswrapper[4612]: I0227 07:51:00.654020 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-hp72h" Feb 27 07:51:00 crc kubenswrapper[4612]: I0227 07:51:00.658529 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-8wk54" Feb 27 07:51:00 crc kubenswrapper[4612]: I0227 07:51:00.662968 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-jzd5b" Feb 27 07:51:00 crc kubenswrapper[4612]: I0227 07:51:00.666245 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w8p5f\" (UniqueName: \"kubernetes.io/projected/19b65635-c86d-4439-8ece-4d8c3dbe986b-kube-api-access-w8p5f\") pod \"console-operator-58897d9998-qvwzc\" (UID: \"19b65635-c86d-4439-8ece-4d8c3dbe986b\") " pod="openshift-console-operator/console-operator-58897d9998-qvwzc" Feb 27 07:51:00 crc kubenswrapper[4612]: I0227 07:51:00.676762 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2rrfz" Feb 27 07:51:00 crc kubenswrapper[4612]: I0227 07:51:00.684555 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-f8qzz" Feb 27 07:51:00 crc kubenswrapper[4612]: I0227 07:51:00.685608 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-nkbg2" Feb 27 07:51:00 crc kubenswrapper[4612]: I0227 07:51:00.687720 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Feb 27 07:51:00 crc kubenswrapper[4612]: I0227 07:51:00.694816 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-z9zch" Feb 27 07:51:00 crc kubenswrapper[4612]: I0227 07:51:00.700659 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-jdng9"] Feb 27 07:51:00 crc kubenswrapper[4612]: I0227 07:51:00.705161 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-spnfb" Feb 27 07:51:00 crc kubenswrapper[4612]: I0227 07:51:00.706575 4612 request.go:700] Waited for 1.905998042s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-machine-config-operator/secrets?fieldSelector=metadata.name%3Dnode-bootstrapper-token&limit=500&resourceVersion=0 Feb 27 07:51:00 crc kubenswrapper[4612]: I0227 07:51:00.708303 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Feb 27 07:51:00 crc kubenswrapper[4612]: I0227 07:51:00.711461 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-5wvhc" Feb 27 07:51:00 crc kubenswrapper[4612]: I0227 07:51:00.726513 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-69bw7\" (UniqueName: \"kubernetes.io/projected/cc3f4dca-b740-4294-93e5-73906f8dc82c-kube-api-access-69bw7\") pod \"migrator-59844c95c7-g9hq6\" (UID: \"cc3f4dca-b740-4294-93e5-73906f8dc82c\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-g9hq6" Feb 27 07:51:00 crc kubenswrapper[4612]: I0227 07:51:00.733090 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Feb 27 07:51:00 crc kubenswrapper[4612]: I0227 07:51:00.754948 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Feb 27 07:51:00 crc kubenswrapper[4612]: I0227 07:51:00.770275 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Feb 27 07:51:00 crc kubenswrapper[4612]: I0227 07:51:00.771161 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2k66r"] Feb 27 07:51:00 crc kubenswrapper[4612]: I0227 07:51:00.792440 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Feb 27 07:51:00 crc kubenswrapper[4612]: W0227 07:51:00.796886 4612 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podca4c97ff_9032_48de_b286_89da26f5a822.slice/crio-f8d847978854c04209852aaaf38afa6493d30de62070076d1e6da80974d018a7 WatchSource:0}: Error finding container f8d847978854c04209852aaaf38afa6493d30de62070076d1e6da80974d018a7: Status 404 returned error can't find the container with id f8d847978854c04209852aaaf38afa6493d30de62070076d1e6da80974d018a7 Feb 27 07:51:00 crc kubenswrapper[4612]: I0227 07:51:00.808953 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Feb 27 07:51:00 crc kubenswrapper[4612]: I0227 07:51:00.829563 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Feb 27 07:51:00 crc kubenswrapper[4612]: I0227 07:51:00.849961 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Feb 27 07:51:00 crc kubenswrapper[4612]: I0227 07:51:00.852224 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-qvwzc" Feb 27 07:51:00 crc kubenswrapper[4612]: I0227 07:51:00.868516 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Feb 27 07:51:00 crc kubenswrapper[4612]: I0227 07:51:00.873125 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-w7pv5"] Feb 27 07:51:00 crc kubenswrapper[4612]: E0227 07:51:00.886324 4612 projected.go:288] Couldn't get configMap openshift-machine-api/openshift-service-ca.crt: failed to sync configmap cache: timed out waiting for the condition Feb 27 07:51:00 crc kubenswrapper[4612]: E0227 07:51:00.886355 4612 projected.go:194] Error preparing data for projected volume kube-api-access-4l9jx for pod openshift-machine-api/machine-api-operator-5694c8668f-q6h7h: failed to sync configmap cache: timed out waiting for the condition Feb 27 07:51:00 crc kubenswrapper[4612]: E0227 07:51:00.886415 4612 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/094b001c-e80c-4d07-a161-e00514667462-kube-api-access-4l9jx podName:094b001c-e80c-4d07-a161-e00514667462 nodeName:}" failed. No retries permitted until 2026-02-27 07:51:01.386394854 +0000 UTC m=+119.240324852 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-4l9jx" (UniqueName: "kubernetes.io/projected/094b001c-e80c-4d07-a161-e00514667462-kube-api-access-4l9jx") pod "machine-api-operator-5694c8668f-q6h7h" (UID: "094b001c-e80c-4d07-a161-e00514667462") : failed to sync configmap cache: timed out waiting for the condition Feb 27 07:51:00 crc kubenswrapper[4612]: I0227 07:51:00.891465 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-sysctl-allowlist" Feb 27 07:51:00 crc kubenswrapper[4612]: I0227 07:51:00.914222 4612 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Feb 27 07:51:00 crc kubenswrapper[4612]: W0227 07:51:00.914407 4612 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8cc01026_2c76_4446_b78f_7d8fb5d9a887.slice/crio-d38018a2cdc4959d6c2fd1d93a059974a4521fa72f5f32193f2c330f26308dcb WatchSource:0}: Error finding container d38018a2cdc4959d6c2fd1d93a059974a4521fa72f5f32193f2c330f26308dcb: Status 404 returned error can't find the container with id d38018a2cdc4959d6c2fd1d93a059974a4521fa72f5f32193f2c330f26308dcb Feb 27 07:51:00 crc kubenswrapper[4612]: E0227 07:51:00.919799 4612 projected.go:288] Couldn't get configMap openshift-cluster-samples-operator/openshift-service-ca.crt: failed to sync configmap cache: timed out waiting for the condition Feb 27 07:51:00 crc kubenswrapper[4612]: E0227 07:51:00.919818 4612 projected.go:194] Error preparing data for projected volume kube-api-access-k7t8s for pod openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-96vl6: failed to sync configmap cache: timed out waiting for the condition Feb 27 07:51:00 crc kubenswrapper[4612]: E0227 07:51:00.919871 4612 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/620a7a2f-92fc-4a76-900b-021d0dd78d79-kube-api-access-k7t8s podName:620a7a2f-92fc-4a76-900b-021d0dd78d79 nodeName:}" failed. No retries permitted until 2026-02-27 07:51:01.419852156 +0000 UTC m=+119.273782154 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-k7t8s" (UniqueName: "kubernetes.io/projected/620a7a2f-92fc-4a76-900b-021d0dd78d79-kube-api-access-k7t8s") pod "cluster-samples-operator-665b6dd947-96vl6" (UID: "620a7a2f-92fc-4a76-900b-021d0dd78d79") : failed to sync configmap cache: timed out waiting for the condition Feb 27 07:51:00 crc kubenswrapper[4612]: I0227 07:51:00.929479 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Feb 27 07:51:00 crc kubenswrapper[4612]: I0227 07:51:00.950238 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Feb 27 07:51:00 crc kubenswrapper[4612]: I0227 07:51:00.954066 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-g9hq6" Feb 27 07:51:00 crc kubenswrapper[4612]: I0227 07:51:00.972039 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.008377 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.028532 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/a6e73baa-3a12-462c-b85c-1b964a52e38f-tmpfs\") pod \"packageserver-d55dfcdfc-c6sdx\" (UID: \"a6e73baa-3a12-462c-b85c-1b964a52e38f\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-c6sdx" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.028575 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/25a0e1aa-86e2-4acf-885b-16c26efc337a-trusted-ca-bundle\") pod \"apiserver-76f77b778f-cw57r\" (UID: \"25a0e1aa-86e2-4acf-885b-16c26efc337a\") " pod="openshift-apiserver/apiserver-76f77b778f-cw57r" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.028593 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rt89w\" (UniqueName: \"kubernetes.io/projected/fb198883-74b8-43a2-a2db-390e951fda9a-kube-api-access-rt89w\") pod \"olm-operator-6b444d44fb-bk49v\" (UID: \"fb198883-74b8-43a2-a2db-390e951fda9a\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bk49v" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.028620 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/db4d6d91-61f5-4beb-809c-63ae9028a911-ca-trust-extracted\") pod \"image-registry-697d97f7c8-pdjlb\" (UID: \"db4d6d91-61f5-4beb-809c-63ae9028a911\") " pod="openshift-image-registry/image-registry-697d97f7c8-pdjlb" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.028637 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-68bjk\" (UniqueName: \"kubernetes.io/projected/cc5b15e4-6c84-4474-bf4d-f652d2e8000c-kube-api-access-68bjk\") pod \"service-ca-9c57cc56f-8l852\" (UID: \"cc5b15e4-6c84-4474-bf4d-f652d2e8000c\") " pod="openshift-service-ca/service-ca-9c57cc56f-8l852" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.028662 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/25a0e1aa-86e2-4acf-885b-16c26efc337a-audit\") pod \"apiserver-76f77b778f-cw57r\" (UID: \"25a0e1aa-86e2-4acf-885b-16c26efc337a\") " pod="openshift-apiserver/apiserver-76f77b778f-cw57r" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.028687 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/25a0e1aa-86e2-4acf-885b-16c26efc337a-audit-dir\") pod \"apiserver-76f77b778f-cw57r\" (UID: \"25a0e1aa-86e2-4acf-885b-16c26efc337a\") " pod="openshift-apiserver/apiserver-76f77b778f-cw57r" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.028726 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gcbww\" (UniqueName: \"kubernetes.io/projected/a6e73baa-3a12-462c-b85c-1b964a52e38f-kube-api-access-gcbww\") pod \"packageserver-d55dfcdfc-c6sdx\" (UID: \"a6e73baa-3a12-462c-b85c-1b964a52e38f\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-c6sdx" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.028742 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hp9f5\" (UniqueName: \"kubernetes.io/projected/738d2b6e-2b7c-461d-b51e-f35eb5d41363-kube-api-access-hp9f5\") pod \"controller-manager-879f6c89f-sxt8c\" (UID: \"738d2b6e-2b7c-461d-b51e-f35eb5d41363\") " pod="openshift-controller-manager/controller-manager-879f6c89f-sxt8c" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.028763 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pdjlb\" (UID: \"db4d6d91-61f5-4beb-809c-63ae9028a911\") " pod="openshift-image-registry/image-registry-697d97f7c8-pdjlb" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.028787 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/25a0e1aa-86e2-4acf-885b-16c26efc337a-etcd-serving-ca\") pod \"apiserver-76f77b778f-cw57r\" (UID: \"25a0e1aa-86e2-4acf-885b-16c26efc337a\") " pod="openshift-apiserver/apiserver-76f77b778f-cw57r" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.028803 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/f0d8e4b4-aec4-433c-adff-091c8944e8a3-metrics-tls\") pod \"dns-operator-744455d44c-ghtdk\" (UID: \"f0d8e4b4-aec4-433c-adff-091c8944e8a3\") " pod="openshift-dns-operator/dns-operator-744455d44c-ghtdk" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.028819 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/db4d6d91-61f5-4beb-809c-63ae9028a911-trusted-ca\") pod \"image-registry-697d97f7c8-pdjlb\" (UID: \"db4d6d91-61f5-4beb-809c-63ae9028a911\") " pod="openshift-image-registry/image-registry-697d97f7c8-pdjlb" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.028861 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/98e7f73e-ab59-4e4d-af52-4794f9e28abb-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-kbtxf\" (UID: \"98e7f73e-ab59-4e4d-af52-4794f9e28abb\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-kbtxf" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.028882 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/738d2b6e-2b7c-461d-b51e-f35eb5d41363-client-ca\") pod \"controller-manager-879f6c89f-sxt8c\" (UID: \"738d2b6e-2b7c-461d-b51e-f35eb5d41363\") " pod="openshift-controller-manager/controller-manager-879f6c89f-sxt8c" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.028897 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/25a0e1aa-86e2-4acf-885b-16c26efc337a-config\") pod \"apiserver-76f77b778f-cw57r\" (UID: \"25a0e1aa-86e2-4acf-885b-16c26efc337a\") " pod="openshift-apiserver/apiserver-76f77b778f-cw57r" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.028934 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/25a0e1aa-86e2-4acf-885b-16c26efc337a-etcd-client\") pod \"apiserver-76f77b778f-cw57r\" (UID: \"25a0e1aa-86e2-4acf-885b-16c26efc337a\") " pod="openshift-apiserver/apiserver-76f77b778f-cw57r" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.028947 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/25a0e1aa-86e2-4acf-885b-16c26efc337a-image-import-ca\") pod \"apiserver-76f77b778f-cw57r\" (UID: \"25a0e1aa-86e2-4acf-885b-16c26efc337a\") " pod="openshift-apiserver/apiserver-76f77b778f-cw57r" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.028962 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/a6e73baa-3a12-462c-b85c-1b964a52e38f-webhook-cert\") pod \"packageserver-d55dfcdfc-c6sdx\" (UID: \"a6e73baa-3a12-462c-b85c-1b964a52e38f\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-c6sdx" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.028997 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2bbj5\" (UniqueName: \"kubernetes.io/projected/db4d6d91-61f5-4beb-809c-63ae9028a911-kube-api-access-2bbj5\") pod \"image-registry-697d97f7c8-pdjlb\" (UID: \"db4d6d91-61f5-4beb-809c-63ae9028a911\") " pod="openshift-image-registry/image-registry-697d97f7c8-pdjlb" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.029027 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/25a0e1aa-86e2-4acf-885b-16c26efc337a-node-pullsecrets\") pod \"apiserver-76f77b778f-cw57r\" (UID: \"25a0e1aa-86e2-4acf-885b-16c26efc337a\") " pod="openshift-apiserver/apiserver-76f77b778f-cw57r" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.029041 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/db4d6d91-61f5-4beb-809c-63ae9028a911-installation-pull-secrets\") pod \"image-registry-697d97f7c8-pdjlb\" (UID: \"db4d6d91-61f5-4beb-809c-63ae9028a911\") " pod="openshift-image-registry/image-registry-697d97f7c8-pdjlb" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.029057 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/db4d6d91-61f5-4beb-809c-63ae9028a911-registry-tls\") pod \"image-registry-697d97f7c8-pdjlb\" (UID: \"db4d6d91-61f5-4beb-809c-63ae9028a911\") " pod="openshift-image-registry/image-registry-697d97f7c8-pdjlb" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.029076 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/fb198883-74b8-43a2-a2db-390e951fda9a-profile-collector-cert\") pod \"olm-operator-6b444d44fb-bk49v\" (UID: \"fb198883-74b8-43a2-a2db-390e951fda9a\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bk49v" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.029112 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/98e7f73e-ab59-4e4d-af52-4794f9e28abb-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-kbtxf\" (UID: \"98e7f73e-ab59-4e4d-af52-4794f9e28abb\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-kbtxf" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.029135 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/cc5b15e4-6c84-4474-bf4d-f652d2e8000c-signing-key\") pod \"service-ca-9c57cc56f-8l852\" (UID: \"cc5b15e4-6c84-4474-bf4d-f652d2e8000c\") " pod="openshift-service-ca/service-ca-9c57cc56f-8l852" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.029152 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/db4d6d91-61f5-4beb-809c-63ae9028a911-bound-sa-token\") pod \"image-registry-697d97f7c8-pdjlb\" (UID: \"db4d6d91-61f5-4beb-809c-63ae9028a911\") " pod="openshift-image-registry/image-registry-697d97f7c8-pdjlb" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.029183 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/a6e73baa-3a12-462c-b85c-1b964a52e38f-apiservice-cert\") pod \"packageserver-d55dfcdfc-c6sdx\" (UID: \"a6e73baa-3a12-462c-b85c-1b964a52e38f\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-c6sdx" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.029206 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/738d2b6e-2b7c-461d-b51e-f35eb5d41363-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-sxt8c\" (UID: \"738d2b6e-2b7c-461d-b51e-f35eb5d41363\") " pod="openshift-controller-manager/controller-manager-879f6c89f-sxt8c" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.029236 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/738d2b6e-2b7c-461d-b51e-f35eb5d41363-config\") pod \"controller-manager-879f6c89f-sxt8c\" (UID: \"738d2b6e-2b7c-461d-b51e-f35eb5d41363\") " pod="openshift-controller-manager/controller-manager-879f6c89f-sxt8c" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.029249 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/738d2b6e-2b7c-461d-b51e-f35eb5d41363-serving-cert\") pod \"controller-manager-879f6c89f-sxt8c\" (UID: \"738d2b6e-2b7c-461d-b51e-f35eb5d41363\") " pod="openshift-controller-manager/controller-manager-879f6c89f-sxt8c" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.029265 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2hm7j\" (UniqueName: \"kubernetes.io/projected/25a0e1aa-86e2-4acf-885b-16c26efc337a-kube-api-access-2hm7j\") pod \"apiserver-76f77b778f-cw57r\" (UID: \"25a0e1aa-86e2-4acf-885b-16c26efc337a\") " pod="openshift-apiserver/apiserver-76f77b778f-cw57r" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.029279 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8tqbm\" (UniqueName: \"kubernetes.io/projected/f0d8e4b4-aec4-433c-adff-091c8944e8a3-kube-api-access-8tqbm\") pod \"dns-operator-744455d44c-ghtdk\" (UID: \"f0d8e4b4-aec4-433c-adff-091c8944e8a3\") " pod="openshift-dns-operator/dns-operator-744455d44c-ghtdk" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.029319 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/25a0e1aa-86e2-4acf-885b-16c26efc337a-encryption-config\") pod \"apiserver-76f77b778f-cw57r\" (UID: \"25a0e1aa-86e2-4acf-885b-16c26efc337a\") " pod="openshift-apiserver/apiserver-76f77b778f-cw57r" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.029340 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/98e7f73e-ab59-4e4d-af52-4794f9e28abb-config\") pod \"kube-apiserver-operator-766d6c64bb-kbtxf\" (UID: \"98e7f73e-ab59-4e4d-af52-4794f9e28abb\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-kbtxf" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.029368 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/db4d6d91-61f5-4beb-809c-63ae9028a911-registry-certificates\") pod \"image-registry-697d97f7c8-pdjlb\" (UID: \"db4d6d91-61f5-4beb-809c-63ae9028a911\") " pod="openshift-image-registry/image-registry-697d97f7c8-pdjlb" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.029385 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/25a0e1aa-86e2-4acf-885b-16c26efc337a-serving-cert\") pod \"apiserver-76f77b778f-cw57r\" (UID: \"25a0e1aa-86e2-4acf-885b-16c26efc337a\") " pod="openshift-apiserver/apiserver-76f77b778f-cw57r" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.029423 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/cc5b15e4-6c84-4474-bf4d-f652d2e8000c-signing-cabundle\") pod \"service-ca-9c57cc56f-8l852\" (UID: \"cc5b15e4-6c84-4474-bf4d-f652d2e8000c\") " pod="openshift-service-ca/service-ca-9c57cc56f-8l852" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.029455 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/fb198883-74b8-43a2-a2db-390e951fda9a-srv-cert\") pod \"olm-operator-6b444d44fb-bk49v\" (UID: \"fb198883-74b8-43a2-a2db-390e951fda9a\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bk49v" Feb 27 07:51:01 crc kubenswrapper[4612]: E0227 07:51:01.038726 4612 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 07:51:01.538675511 +0000 UTC m=+119.392605509 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pdjlb" (UID: "db4d6d91-61f5-4beb-809c-63ae9028a911") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.042021 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/29ab8b79-a05e-4b94-a2bb-2c6b412fd297-machine-approver-tls\") pod \"machine-approver-56656f9798-k6s2g\" (UID: \"29ab8b79-a05e-4b94-a2bb-2c6b412fd297\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-k6s2g" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.043021 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.043653 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/29ab8b79-a05e-4b94-a2bb-2c6b412fd297-auth-proxy-config\") pod \"machine-approver-56656f9798-k6s2g\" (UID: \"29ab8b79-a05e-4b94-a2bb-2c6b412fd297\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-k6s2g" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.048838 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.049180 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-hp72h"] Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.049868 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/29ab8b79-a05e-4b94-a2bb-2c6b412fd297-config\") pod \"machine-approver-56656f9798-k6s2g\" (UID: \"29ab8b79-a05e-4b94-a2bb-2c6b412fd297\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-k6s2g" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.054495 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-nkbg2"] Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.069407 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.085160 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jplnj\" (UniqueName: \"kubernetes.io/projected/29ab8b79-a05e-4b94-a2bb-2c6b412fd297-kube-api-access-jplnj\") pod \"machine-approver-56656f9798-k6s2g\" (UID: \"29ab8b79-a05e-4b94-a2bb-2c6b412fd297\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-k6s2g" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.088190 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.113183 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.127610 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.138077 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.138311 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2bbj5\" (UniqueName: \"kubernetes.io/projected/db4d6d91-61f5-4beb-809c-63ae9028a911-kube-api-access-2bbj5\") pod \"image-registry-697d97f7c8-pdjlb\" (UID: \"db4d6d91-61f5-4beb-809c-63ae9028a911\") " pod="openshift-image-registry/image-registry-697d97f7c8-pdjlb" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.138340 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/4138a6e8-dd4e-44d3-9103-8f328b794227-bound-sa-token\") pod \"ingress-operator-5b745b69d9-4lpn2\" (UID: \"4138a6e8-dd4e-44d3-9103-8f328b794227\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-4lpn2" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.138366 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-87xs7\" (UniqueName: \"kubernetes.io/projected/3dbda067-50b0-4d11-bb83-b7162d0cebde-kube-api-access-87xs7\") pod \"machine-config-server-hbsll\" (UID: \"3dbda067-50b0-4d11-bb83-b7162d0cebde\") " pod="openshift-machine-config-operator/machine-config-server-hbsll" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.138383 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/25a0e1aa-86e2-4acf-885b-16c26efc337a-node-pullsecrets\") pod \"apiserver-76f77b778f-cw57r\" (UID: \"25a0e1aa-86e2-4acf-885b-16c26efc337a\") " pod="openshift-apiserver/apiserver-76f77b778f-cw57r" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.138400 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/db4d6d91-61f5-4beb-809c-63ae9028a911-installation-pull-secrets\") pod \"image-registry-697d97f7c8-pdjlb\" (UID: \"db4d6d91-61f5-4beb-809c-63ae9028a911\") " pod="openshift-image-registry/image-registry-697d97f7c8-pdjlb" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.138416 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/db4d6d91-61f5-4beb-809c-63ae9028a911-registry-tls\") pod \"image-registry-697d97f7c8-pdjlb\" (UID: \"db4d6d91-61f5-4beb-809c-63ae9028a911\") " pod="openshift-image-registry/image-registry-697d97f7c8-pdjlb" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.138432 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/c28eca4e-4d90-4bf0-95bf-880f3a9037e3-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-gcfmk\" (UID: \"c28eca4e-4d90-4bf0-95bf-880f3a9037e3\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-gcfmk" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.138457 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/fb198883-74b8-43a2-a2db-390e951fda9a-profile-collector-cert\") pod \"olm-operator-6b444d44fb-bk49v\" (UID: \"fb198883-74b8-43a2-a2db-390e951fda9a\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bk49v" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.138472 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-89nrv\" (UniqueName: \"kubernetes.io/projected/c28eca4e-4d90-4bf0-95bf-880f3a9037e3-kube-api-access-89nrv\") pod \"machine-config-controller-84d6567774-gcfmk\" (UID: \"c28eca4e-4d90-4bf0-95bf-880f3a9037e3\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-gcfmk" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.138499 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ready\" (UniqueName: \"kubernetes.io/empty-dir/2522fcbd-fe8e-46cf-95ba-620316513df1-ready\") pod \"cni-sysctl-allowlist-ds-v589k\" (UID: \"2522fcbd-fe8e-46cf-95ba-620316513df1\") " pod="openshift-multus/cni-sysctl-allowlist-ds-v589k" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.138532 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/ad4c7d8c-98d7-40c1-ae6d-21ee5655387f-mountpoint-dir\") pod \"csi-hostpathplugin-n5jbm\" (UID: \"ad4c7d8c-98d7-40c1-ae6d-21ee5655387f\") " pod="hostpath-provisioner/csi-hostpathplugin-n5jbm" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.138567 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/98e7f73e-ab59-4e4d-af52-4794f9e28abb-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-kbtxf\" (UID: \"98e7f73e-ab59-4e4d-af52-4794f9e28abb\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-kbtxf" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.138585 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/cc5b15e4-6c84-4474-bf4d-f652d2e8000c-signing-key\") pod \"service-ca-9c57cc56f-8l852\" (UID: \"cc5b15e4-6c84-4474-bf4d-f652d2e8000c\") " pod="openshift-service-ca/service-ca-9c57cc56f-8l852" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.138608 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/db4d6d91-61f5-4beb-809c-63ae9028a911-bound-sa-token\") pod \"image-registry-697d97f7c8-pdjlb\" (UID: \"db4d6d91-61f5-4beb-809c-63ae9028a911\") " pod="openshift-image-registry/image-registry-697d97f7c8-pdjlb" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.138624 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/ad4c7d8c-98d7-40c1-ae6d-21ee5655387f-csi-data-dir\") pod \"csi-hostpathplugin-n5jbm\" (UID: \"ad4c7d8c-98d7-40c1-ae6d-21ee5655387f\") " pod="hostpath-provisioner/csi-hostpathplugin-n5jbm" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.138640 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b1512ff-906a-4df2-b0ca-a09ba21f4875-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-ks964\" (UID: \"0b1512ff-906a-4df2-b0ca-a09ba21f4875\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-ks964" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.138669 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/e24e170c-7ee2-4d40-badf-21bc59723d1c-etcd-service-ca\") pod \"etcd-operator-b45778765-6q4q2\" (UID: \"e24e170c-7ee2-4d40-badf-21bc59723d1c\") " pod="openshift-etcd-operator/etcd-operator-b45778765-6q4q2" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.138685 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/a6e73baa-3a12-462c-b85c-1b964a52e38f-apiservice-cert\") pod \"packageserver-d55dfcdfc-c6sdx\" (UID: \"a6e73baa-3a12-462c-b85c-1b964a52e38f\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-c6sdx" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.138733 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/738d2b6e-2b7c-461d-b51e-f35eb5d41363-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-sxt8c\" (UID: \"738d2b6e-2b7c-461d-b51e-f35eb5d41363\") " pod="openshift-controller-manager/controller-manager-879f6c89f-sxt8c" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.138758 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/ad4c7d8c-98d7-40c1-ae6d-21ee5655387f-socket-dir\") pod \"csi-hostpathplugin-n5jbm\" (UID: \"ad4c7d8c-98d7-40c1-ae6d-21ee5655387f\") " pod="hostpath-provisioner/csi-hostpathplugin-n5jbm" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.138776 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ea3d0df0-517f-4865-964d-3e12a313696b-secret-volume\") pod \"collect-profiles-29536305-vmtnx\" (UID: \"ea3d0df0-517f-4865-964d-3e12a313696b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29536305-vmtnx" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.138802 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/738d2b6e-2b7c-461d-b51e-f35eb5d41363-config\") pod \"controller-manager-879f6c89f-sxt8c\" (UID: \"738d2b6e-2b7c-461d-b51e-f35eb5d41363\") " pod="openshift-controller-manager/controller-manager-879f6c89f-sxt8c" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.138820 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/738d2b6e-2b7c-461d-b51e-f35eb5d41363-serving-cert\") pod \"controller-manager-879f6c89f-sxt8c\" (UID: \"738d2b6e-2b7c-461d-b51e-f35eb5d41363\") " pod="openshift-controller-manager/controller-manager-879f6c89f-sxt8c" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.138836 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/e24e170c-7ee2-4d40-badf-21bc59723d1c-etcd-client\") pod \"etcd-operator-b45778765-6q4q2\" (UID: \"e24e170c-7ee2-4d40-badf-21bc59723d1c\") " pod="openshift-etcd-operator/etcd-operator-b45778765-6q4q2" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.138930 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/ad4c7d8c-98d7-40c1-ae6d-21ee5655387f-plugins-dir\") pod \"csi-hostpathplugin-n5jbm\" (UID: \"ad4c7d8c-98d7-40c1-ae6d-21ee5655387f\") " pod="hostpath-provisioner/csi-hostpathplugin-n5jbm" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.138979 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2hm7j\" (UniqueName: \"kubernetes.io/projected/25a0e1aa-86e2-4acf-885b-16c26efc337a-kube-api-access-2hm7j\") pod \"apiserver-76f77b778f-cw57r\" (UID: \"25a0e1aa-86e2-4acf-885b-16c26efc337a\") " pod="openshift-apiserver/apiserver-76f77b778f-cw57r" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.139008 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8tqbm\" (UniqueName: \"kubernetes.io/projected/f0d8e4b4-aec4-433c-adff-091c8944e8a3-kube-api-access-8tqbm\") pod \"dns-operator-744455d44c-ghtdk\" (UID: \"f0d8e4b4-aec4-433c-adff-091c8944e8a3\") " pod="openshift-dns-operator/dns-operator-744455d44c-ghtdk" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.139023 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/efc3c589-c379-45ca-a974-75483d90c5a4-metrics-tls\") pod \"dns-default-tcqv8\" (UID: \"efc3c589-c379-45ca-a974-75483d90c5a4\") " pod="openshift-dns/dns-default-tcqv8" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.139049 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/4138a6e8-dd4e-44d3-9103-8f328b794227-metrics-tls\") pod \"ingress-operator-5b745b69d9-4lpn2\" (UID: \"4138a6e8-dd4e-44d3-9103-8f328b794227\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-4lpn2" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.139068 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8ncld\" (UniqueName: \"kubernetes.io/projected/e24e170c-7ee2-4d40-badf-21bc59723d1c-kube-api-access-8ncld\") pod \"etcd-operator-b45778765-6q4q2\" (UID: \"e24e170c-7ee2-4d40-badf-21bc59723d1c\") " pod="openshift-etcd-operator/etcd-operator-b45778765-6q4q2" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.139085 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/98e7f73e-ab59-4e4d-af52-4794f9e28abb-config\") pod \"kube-apiserver-operator-766d6c64bb-kbtxf\" (UID: \"98e7f73e-ab59-4e4d-af52-4794f9e28abb\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-kbtxf" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.139100 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/25a0e1aa-86e2-4acf-885b-16c26efc337a-encryption-config\") pod \"apiserver-76f77b778f-cw57r\" (UID: \"25a0e1aa-86e2-4acf-885b-16c26efc337a\") " pod="openshift-apiserver/apiserver-76f77b778f-cw57r" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.139117 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/90d5c002-a228-4c36-b395-5ba0b01aac8e-srv-cert\") pod \"catalog-operator-68c6474976-6cz6v\" (UID: \"90d5c002-a228-4c36-b395-5ba0b01aac8e\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6cz6v" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.139137 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/db4d6d91-61f5-4beb-809c-63ae9028a911-registry-certificates\") pod \"image-registry-697d97f7c8-pdjlb\" (UID: \"db4d6d91-61f5-4beb-809c-63ae9028a911\") " pod="openshift-image-registry/image-registry-697d97f7c8-pdjlb" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.139157 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a61e7e5a-48f6-4b26-9b1e-1858222ff35f-config\") pod \"service-ca-operator-777779d784-42tqr\" (UID: \"a61e7e5a-48f6-4b26-9b1e-1858222ff35f\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-42tqr" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.139175 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/d951f066-9e39-4e0a-9fa3-da9c868552d1-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-zpshm\" (UID: \"d951f066-9e39-4e0a-9fa3-da9c868552d1\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zpshm" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.139213 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4138a6e8-dd4e-44d3-9103-8f328b794227-trusted-ca\") pod \"ingress-operator-5b745b69d9-4lpn2\" (UID: \"4138a6e8-dd4e-44d3-9103-8f328b794227\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-4lpn2" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.139231 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5rp9l\" (UniqueName: \"kubernetes.io/projected/c02d88e5-7055-4864-8b83-d69fdccad379-kube-api-access-5rp9l\") pod \"package-server-manager-789f6589d5-gvv4x\" (UID: \"c02d88e5-7055-4864-8b83-d69fdccad379\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-gvv4x" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.139255 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/25a0e1aa-86e2-4acf-885b-16c26efc337a-serving-cert\") pod \"apiserver-76f77b778f-cw57r\" (UID: \"25a0e1aa-86e2-4acf-885b-16c26efc337a\") " pod="openshift-apiserver/apiserver-76f77b778f-cw57r" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.139276 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/e24e170c-7ee2-4d40-badf-21bc59723d1c-etcd-ca\") pod \"etcd-operator-b45778765-6q4q2\" (UID: \"e24e170c-7ee2-4d40-badf-21bc59723d1c\") " pod="openshift-etcd-operator/etcd-operator-b45778765-6q4q2" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.139292 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e24e170c-7ee2-4d40-badf-21bc59723d1c-serving-cert\") pod \"etcd-operator-b45778765-6q4q2\" (UID: \"e24e170c-7ee2-4d40-badf-21bc59723d1c\") " pod="openshift-etcd-operator/etcd-operator-b45778765-6q4q2" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.139306 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/3dbda067-50b0-4d11-bb83-b7162d0cebde-node-bootstrap-token\") pod \"machine-config-server-hbsll\" (UID: \"3dbda067-50b0-4d11-bb83-b7162d0cebde\") " pod="openshift-machine-config-operator/machine-config-server-hbsll" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.139347 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8c4mq\" (UniqueName: \"kubernetes.io/projected/90d5c002-a228-4c36-b395-5ba0b01aac8e-kube-api-access-8c4mq\") pod \"catalog-operator-68c6474976-6cz6v\" (UID: \"90d5c002-a228-4c36-b395-5ba0b01aac8e\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6cz6v" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.139362 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/c02d88e5-7055-4864-8b83-d69fdccad379-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-gvv4x\" (UID: \"c02d88e5-7055-4864-8b83-d69fdccad379\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-gvv4x" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.139383 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/cc5b15e4-6c84-4474-bf4d-f652d2e8000c-signing-cabundle\") pod \"service-ca-9c57cc56f-8l852\" (UID: \"cc5b15e4-6c84-4474-bf4d-f652d2e8000c\") " pod="openshift-service-ca/service-ca-9c57cc56f-8l852" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.139400 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/ad4c7d8c-98d7-40c1-ae6d-21ee5655387f-registration-dir\") pod \"csi-hostpathplugin-n5jbm\" (UID: \"ad4c7d8c-98d7-40c1-ae6d-21ee5655387f\") " pod="hostpath-provisioner/csi-hostpathplugin-n5jbm" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.139426 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/90d5c002-a228-4c36-b395-5ba0b01aac8e-profile-collector-cert\") pod \"catalog-operator-68c6474976-6cz6v\" (UID: \"90d5c002-a228-4c36-b395-5ba0b01aac8e\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6cz6v" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.139461 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/fb198883-74b8-43a2-a2db-390e951fda9a-srv-cert\") pod \"olm-operator-6b444d44fb-bk49v\" (UID: \"fb198883-74b8-43a2-a2db-390e951fda9a\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bk49v" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.139477 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/2522fcbd-fe8e-46cf-95ba-620316513df1-cni-sysctl-allowlist\") pod \"cni-sysctl-allowlist-ds-v589k\" (UID: \"2522fcbd-fe8e-46cf-95ba-620316513df1\") " pod="openshift-multus/cni-sysctl-allowlist-ds-v589k" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.139493 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b7dnk\" (UniqueName: \"kubernetes.io/projected/efc3c589-c379-45ca-a974-75483d90c5a4-kube-api-access-b7dnk\") pod \"dns-default-tcqv8\" (UID: \"efc3c589-c379-45ca-a974-75483d90c5a4\") " pod="openshift-dns/dns-default-tcqv8" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.139519 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/a6e73baa-3a12-462c-b85c-1b964a52e38f-tmpfs\") pod \"packageserver-d55dfcdfc-c6sdx\" (UID: \"a6e73baa-3a12-462c-b85c-1b964a52e38f\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-c6sdx" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.139535 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/25a0e1aa-86e2-4acf-885b-16c26efc337a-trusted-ca-bundle\") pod \"apiserver-76f77b778f-cw57r\" (UID: \"25a0e1aa-86e2-4acf-885b-16c26efc337a\") " pod="openshift-apiserver/apiserver-76f77b778f-cw57r" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.139550 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rt89w\" (UniqueName: \"kubernetes.io/projected/fb198883-74b8-43a2-a2db-390e951fda9a-kube-api-access-rt89w\") pod \"olm-operator-6b444d44fb-bk49v\" (UID: \"fb198883-74b8-43a2-a2db-390e951fda9a\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bk49v" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.139574 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vmjmq\" (UniqueName: \"kubernetes.io/projected/a61e7e5a-48f6-4b26-9b1e-1858222ff35f-kube-api-access-vmjmq\") pod \"service-ca-operator-777779d784-42tqr\" (UID: \"a61e7e5a-48f6-4b26-9b1e-1858222ff35f\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-42tqr" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.139590 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4rxck\" (UniqueName: \"kubernetes.io/projected/ea3d0df0-517f-4865-964d-3e12a313696b-kube-api-access-4rxck\") pod \"collect-profiles-29536305-vmtnx\" (UID: \"ea3d0df0-517f-4865-964d-3e12a313696b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29536305-vmtnx" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.139606 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/efc3c589-c379-45ca-a974-75483d90c5a4-config-volume\") pod \"dns-default-tcqv8\" (UID: \"efc3c589-c379-45ca-a974-75483d90c5a4\") " pod="openshift-dns/dns-default-tcqv8" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.139622 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/db4d6d91-61f5-4beb-809c-63ae9028a911-ca-trust-extracted\") pod \"image-registry-697d97f7c8-pdjlb\" (UID: \"db4d6d91-61f5-4beb-809c-63ae9028a911\") " pod="openshift-image-registry/image-registry-697d97f7c8-pdjlb" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.139637 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-68bjk\" (UniqueName: \"kubernetes.io/projected/cc5b15e4-6c84-4474-bf4d-f652d2e8000c-kube-api-access-68bjk\") pod \"service-ca-9c57cc56f-8l852\" (UID: \"cc5b15e4-6c84-4474-bf4d-f652d2e8000c\") " pod="openshift-service-ca/service-ca-9c57cc56f-8l852" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.139653 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pcdhz\" (UniqueName: \"kubernetes.io/projected/2522fcbd-fe8e-46cf-95ba-620316513df1-kube-api-access-pcdhz\") pod \"cni-sysctl-allowlist-ds-v589k\" (UID: \"2522fcbd-fe8e-46cf-95ba-620316513df1\") " pod="openshift-multus/cni-sysctl-allowlist-ds-v589k" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.139668 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b979d24b-0d0c-43c1-b50c-52e6b3801daf-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-kgkd4\" (UID: \"b979d24b-0d0c-43c1-b50c-52e6b3801daf\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-kgkd4" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.139683 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7szw8\" (UniqueName: \"kubernetes.io/projected/4138a6e8-dd4e-44d3-9103-8f328b794227-kube-api-access-7szw8\") pod \"ingress-operator-5b745b69d9-4lpn2\" (UID: \"4138a6e8-dd4e-44d3-9103-8f328b794227\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-4lpn2" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.139712 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/c28eca4e-4d90-4bf0-95bf-880f3a9037e3-proxy-tls\") pod \"machine-config-controller-84d6567774-gcfmk\" (UID: \"c28eca4e-4d90-4bf0-95bf-880f3a9037e3\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-gcfmk" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.139757 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/25a0e1aa-86e2-4acf-885b-16c26efc337a-audit\") pod \"apiserver-76f77b778f-cw57r\" (UID: \"25a0e1aa-86e2-4acf-885b-16c26efc337a\") " pod="openshift-apiserver/apiserver-76f77b778f-cw57r" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.139772 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/2522fcbd-fe8e-46cf-95ba-620316513df1-tuning-conf-dir\") pod \"cni-sysctl-allowlist-ds-v589k\" (UID: \"2522fcbd-fe8e-46cf-95ba-620316513df1\") " pod="openshift-multus/cni-sysctl-allowlist-ds-v589k" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.139805 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/25a0e1aa-86e2-4acf-885b-16c26efc337a-audit-dir\") pod \"apiserver-76f77b778f-cw57r\" (UID: \"25a0e1aa-86e2-4acf-885b-16c26efc337a\") " pod="openshift-apiserver/apiserver-76f77b778f-cw57r" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.139820 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gcbww\" (UniqueName: \"kubernetes.io/projected/a6e73baa-3a12-462c-b85c-1b964a52e38f-kube-api-access-gcbww\") pod \"packageserver-d55dfcdfc-c6sdx\" (UID: \"a6e73baa-3a12-462c-b85c-1b964a52e38f\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-c6sdx" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.139836 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hp9f5\" (UniqueName: \"kubernetes.io/projected/738d2b6e-2b7c-461d-b51e-f35eb5d41363-kube-api-access-hp9f5\") pod \"controller-manager-879f6c89f-sxt8c\" (UID: \"738d2b6e-2b7c-461d-b51e-f35eb5d41363\") " pod="openshift-controller-manager/controller-manager-879f6c89f-sxt8c" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.139853 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0b1512ff-906a-4df2-b0ca-a09ba21f4875-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-ks964\" (UID: \"0b1512ff-906a-4df2-b0ca-a09ba21f4875\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-ks964" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.139867 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/fa91e393-fed8-4f4e-8dbe-7bc51ed35cea-cert\") pod \"ingress-canary-kv4nc\" (UID: \"fa91e393-fed8-4f4e-8dbe-7bc51ed35cea\") " pod="openshift-ingress-canary/ingress-canary-kv4nc" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.139890 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/25a0e1aa-86e2-4acf-885b-16c26efc337a-etcd-serving-ca\") pod \"apiserver-76f77b778f-cw57r\" (UID: \"25a0e1aa-86e2-4acf-885b-16c26efc337a\") " pod="openshift-apiserver/apiserver-76f77b778f-cw57r" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.139905 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/f0d8e4b4-aec4-433c-adff-091c8944e8a3-metrics-tls\") pod \"dns-operator-744455d44c-ghtdk\" (UID: \"f0d8e4b4-aec4-433c-adff-091c8944e8a3\") " pod="openshift-dns-operator/dns-operator-744455d44c-ghtdk" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.139920 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b979d24b-0d0c-43c1-b50c-52e6b3801daf-config\") pod \"kube-controller-manager-operator-78b949d7b-kgkd4\" (UID: \"b979d24b-0d0c-43c1-b50c-52e6b3801daf\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-kgkd4" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.139938 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/db4d6d91-61f5-4beb-809c-63ae9028a911-trusted-ca\") pod \"image-registry-697d97f7c8-pdjlb\" (UID: \"db4d6d91-61f5-4beb-809c-63ae9028a911\") " pod="openshift-image-registry/image-registry-697d97f7c8-pdjlb" Feb 27 07:51:01 crc kubenswrapper[4612]: E0227 07:51:01.141967 4612 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 07:51:01.641948797 +0000 UTC m=+119.495878795 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.142109 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/25a0e1aa-86e2-4acf-885b-16c26efc337a-node-pullsecrets\") pod \"apiserver-76f77b778f-cw57r\" (UID: \"25a0e1aa-86e2-4acf-885b-16c26efc337a\") " pod="openshift-apiserver/apiserver-76f77b778f-cw57r" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.142480 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-78npr\" (UniqueName: \"kubernetes.io/projected/ad4c7d8c-98d7-40c1-ae6d-21ee5655387f-kube-api-access-78npr\") pod \"csi-hostpathplugin-n5jbm\" (UID: \"ad4c7d8c-98d7-40c1-ae6d-21ee5655387f\") " pod="hostpath-provisioner/csi-hostpathplugin-n5jbm" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.142506 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e24e170c-7ee2-4d40-badf-21bc59723d1c-config\") pod \"etcd-operator-b45778765-6q4q2\" (UID: \"e24e170c-7ee2-4d40-badf-21bc59723d1c\") " pod="openshift-etcd-operator/etcd-operator-b45778765-6q4q2" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.142525 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ea3d0df0-517f-4865-964d-3e12a313696b-config-volume\") pod \"collect-profiles-29536305-vmtnx\" (UID: \"ea3d0df0-517f-4865-964d-3e12a313696b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29536305-vmtnx" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.142555 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a61e7e5a-48f6-4b26-9b1e-1858222ff35f-serving-cert\") pod \"service-ca-operator-777779d784-42tqr\" (UID: \"a61e7e5a-48f6-4b26-9b1e-1858222ff35f\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-42tqr" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.142591 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/98e7f73e-ab59-4e4d-af52-4794f9e28abb-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-kbtxf\" (UID: \"98e7f73e-ab59-4e4d-af52-4794f9e28abb\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-kbtxf" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.142608 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/738d2b6e-2b7c-461d-b51e-f35eb5d41363-client-ca\") pod \"controller-manager-879f6c89f-sxt8c\" (UID: \"738d2b6e-2b7c-461d-b51e-f35eb5d41363\") " pod="openshift-controller-manager/controller-manager-879f6c89f-sxt8c" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.142628 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6fg28\" (UniqueName: \"kubernetes.io/projected/d951f066-9e39-4e0a-9fa3-da9c868552d1-kube-api-access-6fg28\") pod \"control-plane-machine-set-operator-78cbb6b69f-zpshm\" (UID: \"d951f066-9e39-4e0a-9fa3-da9c868552d1\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zpshm" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.142645 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4gggr\" (UniqueName: \"kubernetes.io/projected/fa91e393-fed8-4f4e-8dbe-7bc51ed35cea-kube-api-access-4gggr\") pod \"ingress-canary-kv4nc\" (UID: \"fa91e393-fed8-4f4e-8dbe-7bc51ed35cea\") " pod="openshift-ingress-canary/ingress-canary-kv4nc" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.142659 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/3dbda067-50b0-4d11-bb83-b7162d0cebde-certs\") pod \"machine-config-server-hbsll\" (UID: \"3dbda067-50b0-4d11-bb83-b7162d0cebde\") " pod="openshift-machine-config-operator/machine-config-server-hbsll" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.144460 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/25a0e1aa-86e2-4acf-885b-16c26efc337a-config\") pod \"apiserver-76f77b778f-cw57r\" (UID: \"25a0e1aa-86e2-4acf-885b-16c26efc337a\") " pod="openshift-apiserver/apiserver-76f77b778f-cw57r" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.144525 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/25a0e1aa-86e2-4acf-885b-16c26efc337a-etcd-client\") pod \"apiserver-76f77b778f-cw57r\" (UID: \"25a0e1aa-86e2-4acf-885b-16c26efc337a\") " pod="openshift-apiserver/apiserver-76f77b778f-cw57r" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.144555 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/25a0e1aa-86e2-4acf-885b-16c26efc337a-image-import-ca\") pod \"apiserver-76f77b778f-cw57r\" (UID: \"25a0e1aa-86e2-4acf-885b-16c26efc337a\") " pod="openshift-apiserver/apiserver-76f77b778f-cw57r" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.144572 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b979d24b-0d0c-43c1-b50c-52e6b3801daf-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-kgkd4\" (UID: \"b979d24b-0d0c-43c1-b50c-52e6b3801daf\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-kgkd4" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.144592 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/a6e73baa-3a12-462c-b85c-1b964a52e38f-webhook-cert\") pod \"packageserver-d55dfcdfc-c6sdx\" (UID: \"a6e73baa-3a12-462c-b85c-1b964a52e38f\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-c6sdx" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.144610 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b1512ff-906a-4df2-b0ca-a09ba21f4875-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-ks964\" (UID: \"0b1512ff-906a-4df2-b0ca-a09ba21f4875\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-ks964" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.148188 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/db4d6d91-61f5-4beb-809c-63ae9028a911-registry-tls\") pod \"image-registry-697d97f7c8-pdjlb\" (UID: \"db4d6d91-61f5-4beb-809c-63ae9028a911\") " pod="openshift-image-registry/image-registry-697d97f7c8-pdjlb" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.151074 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/738d2b6e-2b7c-461d-b51e-f35eb5d41363-config\") pod \"controller-manager-879f6c89f-sxt8c\" (UID: \"738d2b6e-2b7c-461d-b51e-f35eb5d41363\") " pod="openshift-controller-manager/controller-manager-879f6c89f-sxt8c" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.151810 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.154985 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/cc5b15e4-6c84-4474-bf4d-f652d2e8000c-signing-cabundle\") pod \"service-ca-9c57cc56f-8l852\" (UID: \"cc5b15e4-6c84-4474-bf4d-f652d2e8000c\") " pod="openshift-service-ca/service-ca-9c57cc56f-8l852" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.157513 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/db4d6d91-61f5-4beb-809c-63ae9028a911-installation-pull-secrets\") pod \"image-registry-697d97f7c8-pdjlb\" (UID: \"db4d6d91-61f5-4beb-809c-63ae9028a911\") " pod="openshift-image-registry/image-registry-697d97f7c8-pdjlb" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.157515 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/db4d6d91-61f5-4beb-809c-63ae9028a911-registry-certificates\") pod \"image-registry-697d97f7c8-pdjlb\" (UID: \"db4d6d91-61f5-4beb-809c-63ae9028a911\") " pod="openshift-image-registry/image-registry-697d97f7c8-pdjlb" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.157577 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/a6e73baa-3a12-462c-b85c-1b964a52e38f-tmpfs\") pod \"packageserver-d55dfcdfc-c6sdx\" (UID: \"a6e73baa-3a12-462c-b85c-1b964a52e38f\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-c6sdx" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.157710 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/25a0e1aa-86e2-4acf-885b-16c26efc337a-serving-cert\") pod \"apiserver-76f77b778f-cw57r\" (UID: \"25a0e1aa-86e2-4acf-885b-16c26efc337a\") " pod="openshift-apiserver/apiserver-76f77b778f-cw57r" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.157742 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/25a0e1aa-86e2-4acf-885b-16c26efc337a-trusted-ca-bundle\") pod \"apiserver-76f77b778f-cw57r\" (UID: \"25a0e1aa-86e2-4acf-885b-16c26efc337a\") " pod="openshift-apiserver/apiserver-76f77b778f-cw57r" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.158598 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/db4d6d91-61f5-4beb-809c-63ae9028a911-ca-trust-extracted\") pod \"image-registry-697d97f7c8-pdjlb\" (UID: \"db4d6d91-61f5-4beb-809c-63ae9028a911\") " pod="openshift-image-registry/image-registry-697d97f7c8-pdjlb" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.158952 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/25a0e1aa-86e2-4acf-885b-16c26efc337a-etcd-serving-ca\") pod \"apiserver-76f77b778f-cw57r\" (UID: \"25a0e1aa-86e2-4acf-885b-16c26efc337a\") " pod="openshift-apiserver/apiserver-76f77b778f-cw57r" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.159660 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/25a0e1aa-86e2-4acf-885b-16c26efc337a-config\") pod \"apiserver-76f77b778f-cw57r\" (UID: \"25a0e1aa-86e2-4acf-885b-16c26efc337a\") " pod="openshift-apiserver/apiserver-76f77b778f-cw57r" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.160360 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/db4d6d91-61f5-4beb-809c-63ae9028a911-trusted-ca\") pod \"image-registry-697d97f7c8-pdjlb\" (UID: \"db4d6d91-61f5-4beb-809c-63ae9028a911\") " pod="openshift-image-registry/image-registry-697d97f7c8-pdjlb" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.161250 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/25a0e1aa-86e2-4acf-885b-16c26efc337a-image-import-ca\") pod \"apiserver-76f77b778f-cw57r\" (UID: \"25a0e1aa-86e2-4acf-885b-16c26efc337a\") " pod="openshift-apiserver/apiserver-76f77b778f-cw57r" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.161279 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/25a0e1aa-86e2-4acf-885b-16c26efc337a-audit-dir\") pod \"apiserver-76f77b778f-cw57r\" (UID: \"25a0e1aa-86e2-4acf-885b-16c26efc337a\") " pod="openshift-apiserver/apiserver-76f77b778f-cw57r" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.161996 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/25a0e1aa-86e2-4acf-885b-16c26efc337a-audit\") pod \"apiserver-76f77b778f-cw57r\" (UID: \"25a0e1aa-86e2-4acf-885b-16c26efc337a\") " pod="openshift-apiserver/apiserver-76f77b778f-cw57r" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.162259 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/738d2b6e-2b7c-461d-b51e-f35eb5d41363-client-ca\") pod \"controller-manager-879f6c89f-sxt8c\" (UID: \"738d2b6e-2b7c-461d-b51e-f35eb5d41363\") " pod="openshift-controller-manager/controller-manager-879f6c89f-sxt8c" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.162387 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/738d2b6e-2b7c-461d-b51e-f35eb5d41363-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-sxt8c\" (UID: \"738d2b6e-2b7c-461d-b51e-f35eb5d41363\") " pod="openshift-controller-manager/controller-manager-879f6c89f-sxt8c" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.163530 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/f0d8e4b4-aec4-433c-adff-091c8944e8a3-metrics-tls\") pod \"dns-operator-744455d44c-ghtdk\" (UID: \"f0d8e4b4-aec4-433c-adff-091c8944e8a3\") " pod="openshift-dns-operator/dns-operator-744455d44c-ghtdk" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.165022 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/fb198883-74b8-43a2-a2db-390e951fda9a-srv-cert\") pod \"olm-operator-6b444d44fb-bk49v\" (UID: \"fb198883-74b8-43a2-a2db-390e951fda9a\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bk49v" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.165964 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/738d2b6e-2b7c-461d-b51e-f35eb5d41363-serving-cert\") pod \"controller-manager-879f6c89f-sxt8c\" (UID: \"738d2b6e-2b7c-461d-b51e-f35eb5d41363\") " pod="openshift-controller-manager/controller-manager-879f6c89f-sxt8c" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.168972 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.172617 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/fb198883-74b8-43a2-a2db-390e951fda9a-profile-collector-cert\") pod \"olm-operator-6b444d44fb-bk49v\" (UID: \"fb198883-74b8-43a2-a2db-390e951fda9a\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bk49v" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.172875 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/cc5b15e4-6c84-4474-bf4d-f652d2e8000c-signing-key\") pod \"service-ca-9c57cc56f-8l852\" (UID: \"cc5b15e4-6c84-4474-bf4d-f652d2e8000c\") " pod="openshift-service-ca/service-ca-9c57cc56f-8l852" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.174290 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/98e7f73e-ab59-4e4d-af52-4794f9e28abb-config\") pod \"kube-apiserver-operator-766d6c64bb-kbtxf\" (UID: \"98e7f73e-ab59-4e4d-af52-4794f9e28abb\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-kbtxf" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.177072 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/a6e73baa-3a12-462c-b85c-1b964a52e38f-apiservice-cert\") pod \"packageserver-d55dfcdfc-c6sdx\" (UID: \"a6e73baa-3a12-462c-b85c-1b964a52e38f\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-c6sdx" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.180399 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/98e7f73e-ab59-4e4d-af52-4794f9e28abb-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-kbtxf\" (UID: \"98e7f73e-ab59-4e4d-af52-4794f9e28abb\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-kbtxf" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.181584 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/a6e73baa-3a12-462c-b85c-1b964a52e38f-webhook-cert\") pod \"packageserver-d55dfcdfc-c6sdx\" (UID: \"a6e73baa-3a12-462c-b85c-1b964a52e38f\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-c6sdx" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.189435 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.207178 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/25a0e1aa-86e2-4acf-885b-16c26efc337a-etcd-client\") pod \"apiserver-76f77b778f-cw57r\" (UID: \"25a0e1aa-86e2-4acf-885b-16c26efc337a\") " pod="openshift-apiserver/apiserver-76f77b778f-cw57r" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.211726 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/25a0e1aa-86e2-4acf-885b-16c26efc337a-encryption-config\") pod \"apiserver-76f77b778f-cw57r\" (UID: \"25a0e1aa-86e2-4acf-885b-16c26efc337a\") " pod="openshift-apiserver/apiserver-76f77b778f-cw57r" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.213723 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.232569 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.246131 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-87xs7\" (UniqueName: \"kubernetes.io/projected/3dbda067-50b0-4d11-bb83-b7162d0cebde-kube-api-access-87xs7\") pod \"machine-config-server-hbsll\" (UID: \"3dbda067-50b0-4d11-bb83-b7162d0cebde\") " pod="openshift-machine-config-operator/machine-config-server-hbsll" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.246177 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/c28eca4e-4d90-4bf0-95bf-880f3a9037e3-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-gcfmk\" (UID: \"c28eca4e-4d90-4bf0-95bf-880f3a9037e3\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-gcfmk" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.246205 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-89nrv\" (UniqueName: \"kubernetes.io/projected/c28eca4e-4d90-4bf0-95bf-880f3a9037e3-kube-api-access-89nrv\") pod \"machine-config-controller-84d6567774-gcfmk\" (UID: \"c28eca4e-4d90-4bf0-95bf-880f3a9037e3\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-gcfmk" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.246233 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ready\" (UniqueName: \"kubernetes.io/empty-dir/2522fcbd-fe8e-46cf-95ba-620316513df1-ready\") pod \"cni-sysctl-allowlist-ds-v589k\" (UID: \"2522fcbd-fe8e-46cf-95ba-620316513df1\") " pod="openshift-multus/cni-sysctl-allowlist-ds-v589k" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.246255 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/ad4c7d8c-98d7-40c1-ae6d-21ee5655387f-mountpoint-dir\") pod \"csi-hostpathplugin-n5jbm\" (UID: \"ad4c7d8c-98d7-40c1-ae6d-21ee5655387f\") " pod="hostpath-provisioner/csi-hostpathplugin-n5jbm" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.246291 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/ad4c7d8c-98d7-40c1-ae6d-21ee5655387f-csi-data-dir\") pod \"csi-hostpathplugin-n5jbm\" (UID: \"ad4c7d8c-98d7-40c1-ae6d-21ee5655387f\") " pod="hostpath-provisioner/csi-hostpathplugin-n5jbm" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.246305 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b1512ff-906a-4df2-b0ca-a09ba21f4875-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-ks964\" (UID: \"0b1512ff-906a-4df2-b0ca-a09ba21f4875\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-ks964" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.246323 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/e24e170c-7ee2-4d40-badf-21bc59723d1c-etcd-service-ca\") pod \"etcd-operator-b45778765-6q4q2\" (UID: \"e24e170c-7ee2-4d40-badf-21bc59723d1c\") " pod="openshift-etcd-operator/etcd-operator-b45778765-6q4q2" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.246340 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/ad4c7d8c-98d7-40c1-ae6d-21ee5655387f-socket-dir\") pod \"csi-hostpathplugin-n5jbm\" (UID: \"ad4c7d8c-98d7-40c1-ae6d-21ee5655387f\") " pod="hostpath-provisioner/csi-hostpathplugin-n5jbm" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.246357 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ea3d0df0-517f-4865-964d-3e12a313696b-secret-volume\") pod \"collect-profiles-29536305-vmtnx\" (UID: \"ea3d0df0-517f-4865-964d-3e12a313696b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29536305-vmtnx" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.246374 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/e24e170c-7ee2-4d40-badf-21bc59723d1c-etcd-client\") pod \"etcd-operator-b45778765-6q4q2\" (UID: \"e24e170c-7ee2-4d40-badf-21bc59723d1c\") " pod="openshift-etcd-operator/etcd-operator-b45778765-6q4q2" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.246392 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/ad4c7d8c-98d7-40c1-ae6d-21ee5655387f-plugins-dir\") pod \"csi-hostpathplugin-n5jbm\" (UID: \"ad4c7d8c-98d7-40c1-ae6d-21ee5655387f\") " pod="hostpath-provisioner/csi-hostpathplugin-n5jbm" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.246431 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/efc3c589-c379-45ca-a974-75483d90c5a4-metrics-tls\") pod \"dns-default-tcqv8\" (UID: \"efc3c589-c379-45ca-a974-75483d90c5a4\") " pod="openshift-dns/dns-default-tcqv8" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.246487 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/4138a6e8-dd4e-44d3-9103-8f328b794227-metrics-tls\") pod \"ingress-operator-5b745b69d9-4lpn2\" (UID: \"4138a6e8-dd4e-44d3-9103-8f328b794227\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-4lpn2" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.246515 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8ncld\" (UniqueName: \"kubernetes.io/projected/e24e170c-7ee2-4d40-badf-21bc59723d1c-kube-api-access-8ncld\") pod \"etcd-operator-b45778765-6q4q2\" (UID: \"e24e170c-7ee2-4d40-badf-21bc59723d1c\") " pod="openshift-etcd-operator/etcd-operator-b45778765-6q4q2" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.246536 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/90d5c002-a228-4c36-b395-5ba0b01aac8e-srv-cert\") pod \"catalog-operator-68c6474976-6cz6v\" (UID: \"90d5c002-a228-4c36-b395-5ba0b01aac8e\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6cz6v" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.246559 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a61e7e5a-48f6-4b26-9b1e-1858222ff35f-config\") pod \"service-ca-operator-777779d784-42tqr\" (UID: \"a61e7e5a-48f6-4b26-9b1e-1858222ff35f\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-42tqr" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.246655 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/d951f066-9e39-4e0a-9fa3-da9c868552d1-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-zpshm\" (UID: \"d951f066-9e39-4e0a-9fa3-da9c868552d1\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zpshm" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.246713 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4138a6e8-dd4e-44d3-9103-8f328b794227-trusted-ca\") pod \"ingress-operator-5b745b69d9-4lpn2\" (UID: \"4138a6e8-dd4e-44d3-9103-8f328b794227\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-4lpn2" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.246742 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5rp9l\" (UniqueName: \"kubernetes.io/projected/c02d88e5-7055-4864-8b83-d69fdccad379-kube-api-access-5rp9l\") pod \"package-server-manager-789f6589d5-gvv4x\" (UID: \"c02d88e5-7055-4864-8b83-d69fdccad379\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-gvv4x" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.246766 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/e24e170c-7ee2-4d40-badf-21bc59723d1c-etcd-ca\") pod \"etcd-operator-b45778765-6q4q2\" (UID: \"e24e170c-7ee2-4d40-badf-21bc59723d1c\") " pod="openshift-etcd-operator/etcd-operator-b45778765-6q4q2" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.246800 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/3dbda067-50b0-4d11-bb83-b7162d0cebde-node-bootstrap-token\") pod \"machine-config-server-hbsll\" (UID: \"3dbda067-50b0-4d11-bb83-b7162d0cebde\") " pod="openshift-machine-config-operator/machine-config-server-hbsll" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.246820 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e24e170c-7ee2-4d40-badf-21bc59723d1c-serving-cert\") pod \"etcd-operator-b45778765-6q4q2\" (UID: \"e24e170c-7ee2-4d40-badf-21bc59723d1c\") " pod="openshift-etcd-operator/etcd-operator-b45778765-6q4q2" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.246843 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8c4mq\" (UniqueName: \"kubernetes.io/projected/90d5c002-a228-4c36-b395-5ba0b01aac8e-kube-api-access-8c4mq\") pod \"catalog-operator-68c6474976-6cz6v\" (UID: \"90d5c002-a228-4c36-b395-5ba0b01aac8e\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6cz6v" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.246890 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ready\" (UniqueName: \"kubernetes.io/empty-dir/2522fcbd-fe8e-46cf-95ba-620316513df1-ready\") pod \"cni-sysctl-allowlist-ds-v589k\" (UID: \"2522fcbd-fe8e-46cf-95ba-620316513df1\") " pod="openshift-multus/cni-sysctl-allowlist-ds-v589k" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.246916 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/ad4c7d8c-98d7-40c1-ae6d-21ee5655387f-socket-dir\") pod \"csi-hostpathplugin-n5jbm\" (UID: \"ad4c7d8c-98d7-40c1-ae6d-21ee5655387f\") " pod="hostpath-provisioner/csi-hostpathplugin-n5jbm" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.247327 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/e24e170c-7ee2-4d40-badf-21bc59723d1c-etcd-service-ca\") pod \"etcd-operator-b45778765-6q4q2\" (UID: \"e24e170c-7ee2-4d40-badf-21bc59723d1c\") " pod="openshift-etcd-operator/etcd-operator-b45778765-6q4q2" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.247342 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/c02d88e5-7055-4864-8b83-d69fdccad379-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-gvv4x\" (UID: \"c02d88e5-7055-4864-8b83-d69fdccad379\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-gvv4x" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.247363 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/ad4c7d8c-98d7-40c1-ae6d-21ee5655387f-registration-dir\") pod \"csi-hostpathplugin-n5jbm\" (UID: \"ad4c7d8c-98d7-40c1-ae6d-21ee5655387f\") " pod="hostpath-provisioner/csi-hostpathplugin-n5jbm" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.247383 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/90d5c002-a228-4c36-b395-5ba0b01aac8e-profile-collector-cert\") pod \"catalog-operator-68c6474976-6cz6v\" (UID: \"90d5c002-a228-4c36-b395-5ba0b01aac8e\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6cz6v" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.247399 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/ad4c7d8c-98d7-40c1-ae6d-21ee5655387f-mountpoint-dir\") pod \"csi-hostpathplugin-n5jbm\" (UID: \"ad4c7d8c-98d7-40c1-ae6d-21ee5655387f\") " pod="hostpath-provisioner/csi-hostpathplugin-n5jbm" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.247402 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/2522fcbd-fe8e-46cf-95ba-620316513df1-cni-sysctl-allowlist\") pod \"cni-sysctl-allowlist-ds-v589k\" (UID: \"2522fcbd-fe8e-46cf-95ba-620316513df1\") " pod="openshift-multus/cni-sysctl-allowlist-ds-v589k" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.247438 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b7dnk\" (UniqueName: \"kubernetes.io/projected/efc3c589-c379-45ca-a974-75483d90c5a4-kube-api-access-b7dnk\") pod \"dns-default-tcqv8\" (UID: \"efc3c589-c379-45ca-a974-75483d90c5a4\") " pod="openshift-dns/dns-default-tcqv8" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.247481 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4rxck\" (UniqueName: \"kubernetes.io/projected/ea3d0df0-517f-4865-964d-3e12a313696b-kube-api-access-4rxck\") pod \"collect-profiles-29536305-vmtnx\" (UID: \"ea3d0df0-517f-4865-964d-3e12a313696b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29536305-vmtnx" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.247499 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/efc3c589-c379-45ca-a974-75483d90c5a4-config-volume\") pod \"dns-default-tcqv8\" (UID: \"efc3c589-c379-45ca-a974-75483d90c5a4\") " pod="openshift-dns/dns-default-tcqv8" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.247519 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vmjmq\" (UniqueName: \"kubernetes.io/projected/a61e7e5a-48f6-4b26-9b1e-1858222ff35f-kube-api-access-vmjmq\") pod \"service-ca-operator-777779d784-42tqr\" (UID: \"a61e7e5a-48f6-4b26-9b1e-1858222ff35f\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-42tqr" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.247546 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pcdhz\" (UniqueName: \"kubernetes.io/projected/2522fcbd-fe8e-46cf-95ba-620316513df1-kube-api-access-pcdhz\") pod \"cni-sysctl-allowlist-ds-v589k\" (UID: \"2522fcbd-fe8e-46cf-95ba-620316513df1\") " pod="openshift-multus/cni-sysctl-allowlist-ds-v589k" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.247563 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b979d24b-0d0c-43c1-b50c-52e6b3801daf-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-kgkd4\" (UID: \"b979d24b-0d0c-43c1-b50c-52e6b3801daf\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-kgkd4" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.247580 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7szw8\" (UniqueName: \"kubernetes.io/projected/4138a6e8-dd4e-44d3-9103-8f328b794227-kube-api-access-7szw8\") pod \"ingress-operator-5b745b69d9-4lpn2\" (UID: \"4138a6e8-dd4e-44d3-9103-8f328b794227\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-4lpn2" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.247598 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/c28eca4e-4d90-4bf0-95bf-880f3a9037e3-proxy-tls\") pod \"machine-config-controller-84d6567774-gcfmk\" (UID: \"c28eca4e-4d90-4bf0-95bf-880f3a9037e3\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-gcfmk" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.247627 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/2522fcbd-fe8e-46cf-95ba-620316513df1-tuning-conf-dir\") pod \"cni-sysctl-allowlist-ds-v589k\" (UID: \"2522fcbd-fe8e-46cf-95ba-620316513df1\") " pod="openshift-multus/cni-sysctl-allowlist-ds-v589k" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.247658 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0b1512ff-906a-4df2-b0ca-a09ba21f4875-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-ks964\" (UID: \"0b1512ff-906a-4df2-b0ca-a09ba21f4875\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-ks964" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.247676 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/fa91e393-fed8-4f4e-8dbe-7bc51ed35cea-cert\") pod \"ingress-canary-kv4nc\" (UID: \"fa91e393-fed8-4f4e-8dbe-7bc51ed35cea\") " pod="openshift-ingress-canary/ingress-canary-kv4nc" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.247724 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b979d24b-0d0c-43c1-b50c-52e6b3801daf-config\") pod \"kube-controller-manager-operator-78b949d7b-kgkd4\" (UID: \"b979d24b-0d0c-43c1-b50c-52e6b3801daf\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-kgkd4" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.247751 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pdjlb\" (UID: \"db4d6d91-61f5-4beb-809c-63ae9028a911\") " pod="openshift-image-registry/image-registry-697d97f7c8-pdjlb" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.247772 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-78npr\" (UniqueName: \"kubernetes.io/projected/ad4c7d8c-98d7-40c1-ae6d-21ee5655387f-kube-api-access-78npr\") pod \"csi-hostpathplugin-n5jbm\" (UID: \"ad4c7d8c-98d7-40c1-ae6d-21ee5655387f\") " pod="hostpath-provisioner/csi-hostpathplugin-n5jbm" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.247789 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e24e170c-7ee2-4d40-badf-21bc59723d1c-config\") pod \"etcd-operator-b45778765-6q4q2\" (UID: \"e24e170c-7ee2-4d40-badf-21bc59723d1c\") " pod="openshift-etcd-operator/etcd-operator-b45778765-6q4q2" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.247807 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ea3d0df0-517f-4865-964d-3e12a313696b-config-volume\") pod \"collect-profiles-29536305-vmtnx\" (UID: \"ea3d0df0-517f-4865-964d-3e12a313696b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29536305-vmtnx" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.247825 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a61e7e5a-48f6-4b26-9b1e-1858222ff35f-serving-cert\") pod \"service-ca-operator-777779d784-42tqr\" (UID: \"a61e7e5a-48f6-4b26-9b1e-1858222ff35f\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-42tqr" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.247836 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/ad4c7d8c-98d7-40c1-ae6d-21ee5655387f-csi-data-dir\") pod \"csi-hostpathplugin-n5jbm\" (UID: \"ad4c7d8c-98d7-40c1-ae6d-21ee5655387f\") " pod="hostpath-provisioner/csi-hostpathplugin-n5jbm" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.247855 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6fg28\" (UniqueName: \"kubernetes.io/projected/d951f066-9e39-4e0a-9fa3-da9c868552d1-kube-api-access-6fg28\") pod \"control-plane-machine-set-operator-78cbb6b69f-zpshm\" (UID: \"d951f066-9e39-4e0a-9fa3-da9c868552d1\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zpshm" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.247872 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4gggr\" (UniqueName: \"kubernetes.io/projected/fa91e393-fed8-4f4e-8dbe-7bc51ed35cea-kube-api-access-4gggr\") pod \"ingress-canary-kv4nc\" (UID: \"fa91e393-fed8-4f4e-8dbe-7bc51ed35cea\") " pod="openshift-ingress-canary/ingress-canary-kv4nc" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.247886 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/3dbda067-50b0-4d11-bb83-b7162d0cebde-certs\") pod \"machine-config-server-hbsll\" (UID: \"3dbda067-50b0-4d11-bb83-b7162d0cebde\") " pod="openshift-machine-config-operator/machine-config-server-hbsll" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.247930 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b979d24b-0d0c-43c1-b50c-52e6b3801daf-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-kgkd4\" (UID: \"b979d24b-0d0c-43c1-b50c-52e6b3801daf\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-kgkd4" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.247959 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b1512ff-906a-4df2-b0ca-a09ba21f4875-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-ks964\" (UID: \"0b1512ff-906a-4df2-b0ca-a09ba21f4875\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-ks964" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.247990 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/4138a6e8-dd4e-44d3-9103-8f328b794227-bound-sa-token\") pod \"ingress-operator-5b745b69d9-4lpn2\" (UID: \"4138a6e8-dd4e-44d3-9103-8f328b794227\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-4lpn2" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.248954 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/efc3c589-c379-45ca-a974-75483d90c5a4-config-volume\") pod \"dns-default-tcqv8\" (UID: \"efc3c589-c379-45ca-a974-75483d90c5a4\") " pod="openshift-dns/dns-default-tcqv8" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.249994 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/ad4c7d8c-98d7-40c1-ae6d-21ee5655387f-plugins-dir\") pod \"csi-hostpathplugin-n5jbm\" (UID: \"ad4c7d8c-98d7-40c1-ae6d-21ee5655387f\") " pod="hostpath-provisioner/csi-hostpathplugin-n5jbm" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.252116 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e24e170c-7ee2-4d40-badf-21bc59723d1c-config\") pod \"etcd-operator-b45778765-6q4q2\" (UID: \"e24e170c-7ee2-4d40-badf-21bc59723d1c\") " pod="openshift-etcd-operator/etcd-operator-b45778765-6q4q2" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.252211 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/2522fcbd-fe8e-46cf-95ba-620316513df1-tuning-conf-dir\") pod \"cni-sysctl-allowlist-ds-v589k\" (UID: \"2522fcbd-fe8e-46cf-95ba-620316513df1\") " pod="openshift-multus/cni-sysctl-allowlist-ds-v589k" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.252788 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0b1512ff-906a-4df2-b0ca-a09ba21f4875-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-ks964\" (UID: \"0b1512ff-906a-4df2-b0ca-a09ba21f4875\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-ks964" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.253521 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a61e7e5a-48f6-4b26-9b1e-1858222ff35f-config\") pod \"service-ca-operator-777779d784-42tqr\" (UID: \"a61e7e5a-48f6-4b26-9b1e-1858222ff35f\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-42tqr" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.253605 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ea3d0df0-517f-4865-964d-3e12a313696b-config-volume\") pod \"collect-profiles-29536305-vmtnx\" (UID: \"ea3d0df0-517f-4865-964d-3e12a313696b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29536305-vmtnx" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.247673 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/c28eca4e-4d90-4bf0-95bf-880f3a9037e3-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-gcfmk\" (UID: \"c28eca4e-4d90-4bf0-95bf-880f3a9037e3\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-gcfmk" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.257290 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b979d24b-0d0c-43c1-b50c-52e6b3801daf-config\") pod \"kube-controller-manager-operator-78b949d7b-kgkd4\" (UID: \"b979d24b-0d0c-43c1-b50c-52e6b3801daf\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-kgkd4" Feb 27 07:51:01 crc kubenswrapper[4612]: E0227 07:51:01.257531 4612 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 07:51:01.757517681 +0000 UTC m=+119.611447679 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pdjlb" (UID: "db4d6d91-61f5-4beb-809c-63ae9028a911") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.260228 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/4138a6e8-dd4e-44d3-9103-8f328b794227-metrics-tls\") pod \"ingress-operator-5b745b69d9-4lpn2\" (UID: \"4138a6e8-dd4e-44d3-9103-8f328b794227\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-4lpn2" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.260313 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/ad4c7d8c-98d7-40c1-ae6d-21ee5655387f-registration-dir\") pod \"csi-hostpathplugin-n5jbm\" (UID: \"ad4c7d8c-98d7-40c1-ae6d-21ee5655387f\") " pod="hostpath-provisioner/csi-hostpathplugin-n5jbm" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.261419 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4138a6e8-dd4e-44d3-9103-8f328b794227-trusted-ca\") pod \"ingress-operator-5b745b69d9-4lpn2\" (UID: \"4138a6e8-dd4e-44d3-9103-8f328b794227\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-4lpn2" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.261496 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.262050 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/e24e170c-7ee2-4d40-badf-21bc59723d1c-etcd-ca\") pod \"etcd-operator-b45778765-6q4q2\" (UID: \"e24e170c-7ee2-4d40-badf-21bc59723d1c\") " pod="openshift-etcd-operator/etcd-operator-b45778765-6q4q2" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.265926 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/2522fcbd-fe8e-46cf-95ba-620316513df1-cni-sysctl-allowlist\") pod \"cni-sysctl-allowlist-ds-v589k\" (UID: \"2522fcbd-fe8e-46cf-95ba-620316513df1\") " pod="openshift-multus/cni-sysctl-allowlist-ds-v589k" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.266408 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/c28eca4e-4d90-4bf0-95bf-880f3a9037e3-proxy-tls\") pod \"machine-config-controller-84d6567774-gcfmk\" (UID: \"c28eca4e-4d90-4bf0-95bf-880f3a9037e3\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-gcfmk" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.271521 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a61e7e5a-48f6-4b26-9b1e-1858222ff35f-serving-cert\") pod \"service-ca-operator-777779d784-42tqr\" (UID: \"a61e7e5a-48f6-4b26-9b1e-1858222ff35f\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-42tqr" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.273153 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ea3d0df0-517f-4865-964d-3e12a313696b-secret-volume\") pod \"collect-profiles-29536305-vmtnx\" (UID: \"ea3d0df0-517f-4865-964d-3e12a313696b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29536305-vmtnx" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.273636 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/90d5c002-a228-4c36-b395-5ba0b01aac8e-srv-cert\") pod \"catalog-operator-68c6474976-6cz6v\" (UID: \"90d5c002-a228-4c36-b395-5ba0b01aac8e\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6cz6v" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.274459 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/fa91e393-fed8-4f4e-8dbe-7bc51ed35cea-cert\") pod \"ingress-canary-kv4nc\" (UID: \"fa91e393-fed8-4f4e-8dbe-7bc51ed35cea\") " pod="openshift-ingress-canary/ingress-canary-kv4nc" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.274521 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b1512ff-906a-4df2-b0ca-a09ba21f4875-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-ks964\" (UID: \"0b1512ff-906a-4df2-b0ca-a09ba21f4875\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-ks964" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.274629 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b979d24b-0d0c-43c1-b50c-52e6b3801daf-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-kgkd4\" (UID: \"b979d24b-0d0c-43c1-b50c-52e6b3801daf\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-kgkd4" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.274806 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/efc3c589-c379-45ca-a974-75483d90c5a4-metrics-tls\") pod \"dns-default-tcqv8\" (UID: \"efc3c589-c379-45ca-a974-75483d90c5a4\") " pod="openshift-dns/dns-default-tcqv8" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.276417 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/c02d88e5-7055-4864-8b83-d69fdccad379-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-gvv4x\" (UID: \"c02d88e5-7055-4864-8b83-d69fdccad379\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-gvv4x" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.278579 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/3dbda067-50b0-4d11-bb83-b7162d0cebde-node-bootstrap-token\") pod \"machine-config-server-hbsll\" (UID: \"3dbda067-50b0-4d11-bb83-b7162d0cebde\") " pod="openshift-machine-config-operator/machine-config-server-hbsll" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.279826 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/d951f066-9e39-4e0a-9fa3-da9c868552d1-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-zpshm\" (UID: \"d951f066-9e39-4e0a-9fa3-da9c868552d1\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zpshm" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.283411 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e24e170c-7ee2-4d40-badf-21bc59723d1c-serving-cert\") pod \"etcd-operator-b45778765-6q4q2\" (UID: \"e24e170c-7ee2-4d40-badf-21bc59723d1c\") " pod="openshift-etcd-operator/etcd-operator-b45778765-6q4q2" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.290086 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.290176 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/e24e170c-7ee2-4d40-badf-21bc59723d1c-etcd-client\") pod \"etcd-operator-b45778765-6q4q2\" (UID: \"e24e170c-7ee2-4d40-badf-21bc59723d1c\") " pod="openshift-etcd-operator/etcd-operator-b45778765-6q4q2" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.290180 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/90d5c002-a228-4c36-b395-5ba0b01aac8e-profile-collector-cert\") pod \"catalog-operator-68c6474976-6cz6v\" (UID: \"90d5c002-a228-4c36-b395-5ba0b01aac8e\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6cz6v" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.295820 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/3dbda067-50b0-4d11-bb83-b7162d0cebde-certs\") pod \"machine-config-server-hbsll\" (UID: \"3dbda067-50b0-4d11-bb83-b7162d0cebde\") " pod="openshift-machine-config-operator/machine-config-server-hbsll" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.298358 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-k6s2g" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.349658 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8tqbm\" (UniqueName: \"kubernetes.io/projected/f0d8e4b4-aec4-433c-adff-091c8944e8a3-kube-api-access-8tqbm\") pod \"dns-operator-744455d44c-ghtdk\" (UID: \"f0d8e4b4-aec4-433c-adff-091c8944e8a3\") " pod="openshift-dns-operator/dns-operator-744455d44c-ghtdk" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.350107 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 07:51:01 crc kubenswrapper[4612]: E0227 07:51:01.350805 4612 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 07:51:01.850789912 +0000 UTC m=+119.704719910 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.350903 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2bbj5\" (UniqueName: \"kubernetes.io/projected/db4d6d91-61f5-4beb-809c-63ae9028a911-kube-api-access-2bbj5\") pod \"image-registry-697d97f7c8-pdjlb\" (UID: \"db4d6d91-61f5-4beb-809c-63ae9028a911\") " pod="openshift-image-registry/image-registry-697d97f7c8-pdjlb" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.351134 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pdjlb\" (UID: \"db4d6d91-61f5-4beb-809c-63ae9028a911\") " pod="openshift-image-registry/image-registry-697d97f7c8-pdjlb" Feb 27 07:51:01 crc kubenswrapper[4612]: E0227 07:51:01.351638 4612 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 07:51:01.851626815 +0000 UTC m=+119.705556813 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pdjlb" (UID: "db4d6d91-61f5-4beb-809c-63ae9028a911") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.363273 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2hm7j\" (UniqueName: \"kubernetes.io/projected/25a0e1aa-86e2-4acf-885b-16c26efc337a-kube-api-access-2hm7j\") pod \"apiserver-76f77b778f-cw57r\" (UID: \"25a0e1aa-86e2-4acf-885b-16c26efc337a\") " pod="openshift-apiserver/apiserver-76f77b778f-cw57r" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.367532 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-8wk54"] Feb 27 07:51:01 crc kubenswrapper[4612]: W0227 07:51:01.397965 4612 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podda284312_b98a_48d6_8cff_f8071c334e0b.slice/crio-535ccf5b8854efb31e4b3373f23422349a74e17fc98ec63a67a191a8691cb543 WatchSource:0}: Error finding container 535ccf5b8854efb31e4b3373f23422349a74e17fc98ec63a67a191a8691cb543: Status 404 returned error can't find the container with id 535ccf5b8854efb31e4b3373f23422349a74e17fc98ec63a67a191a8691cb543 Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.401490 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hp9f5\" (UniqueName: \"kubernetes.io/projected/738d2b6e-2b7c-461d-b51e-f35eb5d41363-kube-api-access-hp9f5\") pod \"controller-manager-879f6c89f-sxt8c\" (UID: \"738d2b6e-2b7c-461d-b51e-f35eb5d41363\") " pod="openshift-controller-manager/controller-manager-879f6c89f-sxt8c" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.417359 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/98e7f73e-ab59-4e4d-af52-4794f9e28abb-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-kbtxf\" (UID: \"98e7f73e-ab59-4e4d-af52-4794f9e28abb\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-kbtxf" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.423850 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-cw57r" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.435465 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rt89w\" (UniqueName: \"kubernetes.io/projected/fb198883-74b8-43a2-a2db-390e951fda9a-kube-api-access-rt89w\") pod \"olm-operator-6b444d44fb-bk49v\" (UID: \"fb198883-74b8-43a2-a2db-390e951fda9a\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bk49v" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.440461 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-spnfb"] Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.440512 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-f8qzz"] Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.451820 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.452011 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4l9jx\" (UniqueName: \"kubernetes.io/projected/094b001c-e80c-4d07-a161-e00514667462-kube-api-access-4l9jx\") pod \"machine-api-operator-5694c8668f-q6h7h\" (UID: \"094b001c-e80c-4d07-a161-e00514667462\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-q6h7h" Feb 27 07:51:01 crc kubenswrapper[4612]: E0227 07:51:01.452094 4612 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 07:51:01.952036112 +0000 UTC m=+119.805966110 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.452180 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pdjlb\" (UID: \"db4d6d91-61f5-4beb-809c-63ae9028a911\") " pod="openshift-image-registry/image-registry-697d97f7c8-pdjlb" Feb 27 07:51:01 crc kubenswrapper[4612]: E0227 07:51:01.452997 4612 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 07:51:01.952988048 +0000 UTC m=+119.806918046 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pdjlb" (UID: "db4d6d91-61f5-4beb-809c-63ae9028a911") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.453418 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/db4d6d91-61f5-4beb-809c-63ae9028a911-bound-sa-token\") pod \"image-registry-697d97f7c8-pdjlb\" (UID: \"db4d6d91-61f5-4beb-809c-63ae9028a911\") " pod="openshift-image-registry/image-registry-697d97f7c8-pdjlb" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.453544 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k7t8s\" (UniqueName: \"kubernetes.io/projected/620a7a2f-92fc-4a76-900b-021d0dd78d79-kube-api-access-k7t8s\") pod \"cluster-samples-operator-665b6dd947-96vl6\" (UID: \"620a7a2f-92fc-4a76-900b-021d0dd78d79\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-96vl6" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.457511 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4l9jx\" (UniqueName: \"kubernetes.io/projected/094b001c-e80c-4d07-a161-e00514667462-kube-api-access-4l9jx\") pod \"machine-api-operator-5694c8668f-q6h7h\" (UID: \"094b001c-e80c-4d07-a161-e00514667462\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-q6h7h" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.463766 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k7t8s\" (UniqueName: \"kubernetes.io/projected/620a7a2f-92fc-4a76-900b-021d0dd78d79-kube-api-access-k7t8s\") pod \"cluster-samples-operator-665b6dd947-96vl6\" (UID: \"620a7a2f-92fc-4a76-900b-021d0dd78d79\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-96vl6" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.466302 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-2rrfz"] Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.472798 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gcbww\" (UniqueName: \"kubernetes.io/projected/a6e73baa-3a12-462c-b85c-1b964a52e38f-kube-api-access-gcbww\") pod \"packageserver-d55dfcdfc-c6sdx\" (UID: \"a6e73baa-3a12-462c-b85c-1b964a52e38f\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-c6sdx" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.477731 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-2d7kh" event={"ID":"faae1773-bf96-4127-8060-a40da07930ce","Type":"ContainerStarted","Data":"6add95c3db714e2135588dca5862abc35f5829a751945d207ce470cb59bb14f6"} Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.477766 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-2d7kh" event={"ID":"faae1773-bf96-4127-8060-a40da07930ce","Type":"ContainerStarted","Data":"e2143dc0f85f93832d9fc88670e1fb8f061a8e94f60d6c5b15a42482ddbad518"} Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.481213 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2k66r" event={"ID":"78554bd5-6fc8-4ce8-b1d2-b183dab17495","Type":"ContainerStarted","Data":"11788a8a411d750a3355b51df9d01a590bf7b802feacb0e048f02b50afea5f43"} Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.481255 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2k66r" event={"ID":"78554bd5-6fc8-4ce8-b1d2-b183dab17495","Type":"ContainerStarted","Data":"29151a452b771df2fbe425d081880a0036426bd95595bfd97bae8ddedb1ea10d"} Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.486874 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-q6h7h" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.489466 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-k6s2g" event={"ID":"29ab8b79-a05e-4b94-a2bb-2c6b412fd297","Type":"ContainerStarted","Data":"30338e32ae49470362aefe206908fd916e4239cc980660ec278ab344270cab23"} Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.491634 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-kbtxf" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.493240 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-68bjk\" (UniqueName: \"kubernetes.io/projected/cc5b15e4-6c84-4474-bf4d-f652d2e8000c-kube-api-access-68bjk\") pod \"service-ca-9c57cc56f-8l852\" (UID: \"cc5b15e4-6c84-4474-bf4d-f652d2e8000c\") " pod="openshift-service-ca/service-ca-9c57cc56f-8l852" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.494769 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-g9hq6"] Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.495130 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-nkbg2" event={"ID":"6ece12fd-d439-4104-a7a9-a9d174e29b5a","Type":"ContainerStarted","Data":"2142c9ee66879280306a4cce8c69235d9b9acac78defb5151ef4f7e8961cd22e"} Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.496571 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-96vl6" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.498893 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-w7pv5" event={"ID":"8cc01026-2c76-4446-b78f-7d8fb5d9a887","Type":"ContainerStarted","Data":"735daf2bc412039544dcb5924176f4897e5b30cbcb2c6d96bf9ead022e1bac29"} Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.498946 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-w7pv5" event={"ID":"8cc01026-2c76-4446-b78f-7d8fb5d9a887","Type":"ContainerStarted","Data":"c9c86c47f82f486b150b35502363b595f99eece631f760a97c233b436f8ff156"} Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.498956 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-w7pv5" event={"ID":"8cc01026-2c76-4446-b78f-7d8fb5d9a887","Type":"ContainerStarted","Data":"d38018a2cdc4959d6c2fd1d93a059974a4521fa72f5f32193f2c330f26308dcb"} Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.509020 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-jdng9" event={"ID":"ca4c97ff-9032-48de-b286-89da26f5a822","Type":"ContainerStarted","Data":"5ce511858e764855d0c051bf47d19c5043442eabe793281c6df726d8d0164d72"} Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.509056 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-jdng9" event={"ID":"ca4c97ff-9032-48de-b286-89da26f5a822","Type":"ContainerStarted","Data":"f8d847978854c04209852aaaf38afa6493d30de62070076d1e6da80974d018a7"} Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.510787 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-87xs7\" (UniqueName: \"kubernetes.io/projected/3dbda067-50b0-4d11-bb83-b7162d0cebde-kube-api-access-87xs7\") pod \"machine-config-server-hbsll\" (UID: \"3dbda067-50b0-4d11-bb83-b7162d0cebde\") " pod="openshift-machine-config-operator/machine-config-server-hbsll" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.513984 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-8wk54" event={"ID":"da284312-b98a-48d6-8cff-f8071c334e0b","Type":"ContainerStarted","Data":"535ccf5b8854efb31e4b3373f23422349a74e17fc98ec63a67a191a8691cb543"} Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.516401 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-hp72h" event={"ID":"6d061f42-06de-43e5-b4d0-dbca09e48630","Type":"ContainerStarted","Data":"872f6d6ba2cff41e30dff82c7bfa288c746d738f5f9784ceca408e98c3730656"} Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.516426 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-hp72h" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.516435 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-hp72h" event={"ID":"6d061f42-06de-43e5-b4d0-dbca09e48630","Type":"ContainerStarted","Data":"b00ebd717acd7c4583cb33e8ef55fa92362ffe216db3eda89fe149afef661031"} Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.516934 4612 patch_prober.go:28] interesting pod/downloads-7954f5f757-zbccb container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.9:8080/\": dial tcp 10.217.0.9:8080: connect: connection refused" start-of-body= Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.516958 4612 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-zbccb" podUID="a5bbba10-16da-4fce-ae5f-311fc5d0d6c1" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.9:8080/\": dial tcp 10.217.0.9:8080: connect: connection refused" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.520429 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-89nrv\" (UniqueName: \"kubernetes.io/projected/c28eca4e-4d90-4bf0-95bf-880f3a9037e3-kube-api-access-89nrv\") pod \"machine-config-controller-84d6567774-gcfmk\" (UID: \"c28eca4e-4d90-4bf0-95bf-880f3a9037e3\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-gcfmk" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.525827 4612 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-hp72h container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.19:8080/healthz\": dial tcp 10.217.0.19:8080: connect: connection refused" start-of-body= Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.525857 4612 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-hp72h" podUID="6d061f42-06de-43e5-b4d0-dbca09e48630" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.19:8080/healthz\": dial tcp 10.217.0.19:8080: connect: connection refused" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.539867 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bk49v" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.549029 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b979d24b-0d0c-43c1-b50c-52e6b3801daf-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-kgkd4\" (UID: \"b979d24b-0d0c-43c1-b50c-52e6b3801daf\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-kgkd4" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.550510 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-5wvhc"] Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.555223 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 07:51:01 crc kubenswrapper[4612]: E0227 07:51:01.558293 4612 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 07:51:02.058275749 +0000 UTC m=+119.912205747 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.563201 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-jzd5b"] Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.567607 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-c6sdx" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.589128 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b1512ff-906a-4df2-b0ca-a09ba21f4875-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-ks964\" (UID: \"0b1512ff-906a-4df2-b0ca-a09ba21f4875\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-ks964" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.595346 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-ghtdk" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.603085 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-qvwzc"] Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.603140 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-z9zch"] Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.603650 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b7dnk\" (UniqueName: \"kubernetes.io/projected/efc3c589-c379-45ca-a974-75483d90c5a4-kube-api-access-b7dnk\") pod \"dns-default-tcqv8\" (UID: \"efc3c589-c379-45ca-a974-75483d90c5a4\") " pod="openshift-dns/dns-default-tcqv8" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.603928 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-8l852" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.622450 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-gcfmk" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.623895 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4rxck\" (UniqueName: \"kubernetes.io/projected/ea3d0df0-517f-4865-964d-3e12a313696b-kube-api-access-4rxck\") pod \"collect-profiles-29536305-vmtnx\" (UID: \"ea3d0df0-517f-4865-964d-3e12a313696b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29536305-vmtnx" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.632170 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/4138a6e8-dd4e-44d3-9103-8f328b794227-bound-sa-token\") pod \"ingress-operator-5b745b69d9-4lpn2\" (UID: \"4138a6e8-dd4e-44d3-9103-8f328b794227\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-4lpn2" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.635990 4612 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-2d7kh" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.636401 4612 patch_prober.go:28] interesting pod/router-default-5444994796-2d7kh container/router namespace/openshift-ingress: Startup probe status=failure output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" start-of-body= Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.636438 4612 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-2d7kh" podUID="faae1773-bf96-4127-8060-a40da07930ce" containerName="router" probeResult="failure" output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.645778 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7szw8\" (UniqueName: \"kubernetes.io/projected/4138a6e8-dd4e-44d3-9103-8f328b794227-kube-api-access-7szw8\") pod \"ingress-operator-5b745b69d9-4lpn2\" (UID: \"4138a6e8-dd4e-44d3-9103-8f328b794227\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-4lpn2" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.647230 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-kgkd4" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.648570 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-sxt8c" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.654898 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-ks964" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.657618 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pdjlb\" (UID: \"db4d6d91-61f5-4beb-809c-63ae9028a911\") " pod="openshift-image-registry/image-registry-697d97f7c8-pdjlb" Feb 27 07:51:01 crc kubenswrapper[4612]: E0227 07:51:01.663125 4612 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 07:51:02.163100068 +0000 UTC m=+120.017030066 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pdjlb" (UID: "db4d6d91-61f5-4beb-809c-63ae9028a911") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.675929 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-hbsll" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.690935 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-tcqv8" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.695677 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vmjmq\" (UniqueName: \"kubernetes.io/projected/a61e7e5a-48f6-4b26-9b1e-1858222ff35f-kube-api-access-vmjmq\") pod \"service-ca-operator-777779d784-42tqr\" (UID: \"a61e7e5a-48f6-4b26-9b1e-1858222ff35f\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-42tqr" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.725148 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pcdhz\" (UniqueName: \"kubernetes.io/projected/2522fcbd-fe8e-46cf-95ba-620316513df1-kube-api-access-pcdhz\") pod \"cni-sysctl-allowlist-ds-v589k\" (UID: \"2522fcbd-fe8e-46cf-95ba-620316513df1\") " pod="openshift-multus/cni-sysctl-allowlist-ds-v589k" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.737176 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4gggr\" (UniqueName: \"kubernetes.io/projected/fa91e393-fed8-4f4e-8dbe-7bc51ed35cea-kube-api-access-4gggr\") pod \"ingress-canary-kv4nc\" (UID: \"fa91e393-fed8-4f4e-8dbe-7bc51ed35cea\") " pod="openshift-ingress-canary/ingress-canary-kv4nc" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.744941 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8ncld\" (UniqueName: \"kubernetes.io/projected/e24e170c-7ee2-4d40-badf-21bc59723d1c-kube-api-access-8ncld\") pod \"etcd-operator-b45778765-6q4q2\" (UID: \"e24e170c-7ee2-4d40-badf-21bc59723d1c\") " pod="openshift-etcd-operator/etcd-operator-b45778765-6q4q2" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.759844 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 07:51:01 crc kubenswrapper[4612]: E0227 07:51:01.760653 4612 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 07:51:02.260634796 +0000 UTC m=+120.114564784 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.766812 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-78npr\" (UniqueName: \"kubernetes.io/projected/ad4c7d8c-98d7-40c1-ae6d-21ee5655387f-kube-api-access-78npr\") pod \"csi-hostpathplugin-n5jbm\" (UID: \"ad4c7d8c-98d7-40c1-ae6d-21ee5655387f\") " pod="hostpath-provisioner/csi-hostpathplugin-n5jbm" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.791415 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6fg28\" (UniqueName: \"kubernetes.io/projected/d951f066-9e39-4e0a-9fa3-da9c868552d1-kube-api-access-6fg28\") pod \"control-plane-machine-set-operator-78cbb6b69f-zpshm\" (UID: \"d951f066-9e39-4e0a-9fa3-da9c868552d1\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zpshm" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.802317 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8c4mq\" (UniqueName: \"kubernetes.io/projected/90d5c002-a228-4c36-b395-5ba0b01aac8e-kube-api-access-8c4mq\") pod \"catalog-operator-68c6474976-6cz6v\" (UID: \"90d5c002-a228-4c36-b395-5ba0b01aac8e\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6cz6v" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.813842 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5rp9l\" (UniqueName: \"kubernetes.io/projected/c02d88e5-7055-4864-8b83-d69fdccad379-kube-api-access-5rp9l\") pod \"package-server-manager-789f6589d5-gvv4x\" (UID: \"c02d88e5-7055-4864-8b83-d69fdccad379\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-gvv4x" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.863615 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pdjlb\" (UID: \"db4d6d91-61f5-4beb-809c-63ae9028a911\") " pod="openshift-image-registry/image-registry-697d97f7c8-pdjlb" Feb 27 07:51:01 crc kubenswrapper[4612]: E0227 07:51:01.864179 4612 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 07:51:02.364162468 +0000 UTC m=+120.218092466 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pdjlb" (UID: "db4d6d91-61f5-4beb-809c-63ae9028a911") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.886488 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-cw57r"] Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.892628 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29536305-vmtnx" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.926040 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-4lpn2" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.935054 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-42tqr" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.940426 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-6q4q2" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.963391 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-gvv4x" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.964053 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.964228 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9421f529-9b1e-4bd0-8032-7baa608981b4-metrics-certs\") pod \"network-metrics-daemon-bqh6n\" (UID: \"9421f529-9b1e-4bd0-8032-7baa608981b4\") " pod="openshift-multus/network-metrics-daemon-bqh6n" Feb 27 07:51:01 crc kubenswrapper[4612]: E0227 07:51:01.965011 4612 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 07:51:02.464948976 +0000 UTC m=+120.318878974 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.969607 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zpshm" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.971866 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9421f529-9b1e-4bd0-8032-7baa608981b4-metrics-certs\") pod \"network-metrics-daemon-bqh6n\" (UID: \"9421f529-9b1e-4bd0-8032-7baa608981b4\") " pod="openshift-multus/network-metrics-daemon-bqh6n" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.985114 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-q6h7h"] Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.987041 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-kv4nc" Feb 27 07:51:01 crc kubenswrapper[4612]: I0227 07:51:01.997665 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-bqh6n" Feb 27 07:51:02 crc kubenswrapper[4612]: I0227 07:51:02.009660 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/cni-sysctl-allowlist-ds-v589k" Feb 27 07:51:02 crc kubenswrapper[4612]: I0227 07:51:02.011458 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-c6sdx"] Feb 27 07:51:02 crc kubenswrapper[4612]: I0227 07:51:02.027860 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-n5jbm" Feb 27 07:51:02 crc kubenswrapper[4612]: I0227 07:51:02.045045 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6cz6v" Feb 27 07:51:02 crc kubenswrapper[4612]: I0227 07:51:02.070798 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pdjlb\" (UID: \"db4d6d91-61f5-4beb-809c-63ae9028a911\") " pod="openshift-image-registry/image-registry-697d97f7c8-pdjlb" Feb 27 07:51:02 crc kubenswrapper[4612]: E0227 07:51:02.071112 4612 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 07:51:02.571100131 +0000 UTC m=+120.425030129 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pdjlb" (UID: "db4d6d91-61f5-4beb-809c-63ae9028a911") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 07:51:02 crc kubenswrapper[4612]: I0227 07:51:02.120262 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-kbtxf"] Feb 27 07:51:02 crc kubenswrapper[4612]: I0227 07:51:02.175428 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 07:51:02 crc kubenswrapper[4612]: E0227 07:51:02.176095 4612 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 07:51:02.676080734 +0000 UTC m=+120.530010732 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 07:51:02 crc kubenswrapper[4612]: I0227 07:51:02.217958 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bk49v"] Feb 27 07:51:02 crc kubenswrapper[4612]: I0227 07:51:02.280656 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pdjlb\" (UID: \"db4d6d91-61f5-4beb-809c-63ae9028a911\") " pod="openshift-image-registry/image-registry-697d97f7c8-pdjlb" Feb 27 07:51:02 crc kubenswrapper[4612]: E0227 07:51:02.280945 4612 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 07:51:02.780934854 +0000 UTC m=+120.634864852 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pdjlb" (UID: "db4d6d91-61f5-4beb-809c-63ae9028a911") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 07:51:02 crc kubenswrapper[4612]: I0227 07:51:02.354627 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-sxt8c"] Feb 27 07:51:02 crc kubenswrapper[4612]: I0227 07:51:02.390405 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 07:51:02 crc kubenswrapper[4612]: E0227 07:51:02.390867 4612 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 07:51:02.890838192 +0000 UTC m=+120.744768180 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 07:51:02 crc kubenswrapper[4612]: W0227 07:51:02.402146 4612 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfb198883_74b8_43a2_a2db_390e951fda9a.slice/crio-3d21c3631a46e51b61430f6da4d748ed54db8d1f3289c024aada20919a0182a4 WatchSource:0}: Error finding container 3d21c3631a46e51b61430f6da4d748ed54db8d1f3289c024aada20919a0182a4: Status 404 returned error can't find the container with id 3d21c3631a46e51b61430f6da4d748ed54db8d1f3289c024aada20919a0182a4 Feb 27 07:51:02 crc kubenswrapper[4612]: I0227 07:51:02.440780 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-spnfb"] Feb 27 07:51:02 crc kubenswrapper[4612]: I0227 07:51:02.469840 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-jdng9" podStartSLOduration=59.469819029 podStartE2EDuration="59.469819029s" podCreationTimestamp="2026-02-27 07:50:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 07:51:02.466854357 +0000 UTC m=+120.320784355" watchObservedRunningTime="2026-02-27 07:51:02.469819029 +0000 UTC m=+120.323749027" Feb 27 07:51:02 crc kubenswrapper[4612]: I0227 07:51:02.507800 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pdjlb\" (UID: \"db4d6d91-61f5-4beb-809c-63ae9028a911\") " pod="openshift-image-registry/image-registry-697d97f7c8-pdjlb" Feb 27 07:51:02 crc kubenswrapper[4612]: E0227 07:51:02.508269 4612 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 07:51:03.008256338 +0000 UTC m=+120.862186336 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pdjlb" (UID: "db4d6d91-61f5-4beb-809c-63ae9028a911") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 07:51:02 crc kubenswrapper[4612]: I0227 07:51:02.503669 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-ghtdk"] Feb 27 07:51:02 crc kubenswrapper[4612]: I0227 07:51:02.539374 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-tcqv8"] Feb 27 07:51:02 crc kubenswrapper[4612]: I0227 07:51:02.541980 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/cni-sysctl-allowlist-ds-v589k" event={"ID":"2522fcbd-fe8e-46cf-95ba-620316513df1","Type":"ContainerStarted","Data":"baf8c9a7cabecfe6817cfdfd39ba317bbbf806b6e12b1d396640418d85825448"} Feb 27 07:51:02 crc kubenswrapper[4612]: I0227 07:51:02.551298 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-hbsll" event={"ID":"3dbda067-50b0-4d11-bb83-b7162d0cebde","Type":"ContainerStarted","Data":"62cc3f31f052a81f7811baacbcef6eae95e018544be1a1a6e6614fbab68d60aa"} Feb 27 07:51:02 crc kubenswrapper[4612]: I0227 07:51:02.567529 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-5wvhc" event={"ID":"ff5fcfe1-eb01-4f01-857d-a67712db3b87","Type":"ContainerStarted","Data":"f66f57a8356a8e83dd1ba236dfe7812ef080397e1c16446155696a63b8444f93"} Feb 27 07:51:02 crc kubenswrapper[4612]: I0227 07:51:02.591440 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-f8qzz" event={"ID":"3b5b5090-45a4-4acc-aa94-84c892bcf306","Type":"ContainerStarted","Data":"09d216dc3cdb2e564be1cf1f496892a6b9fcd7577f9b92dacebbd7439842ce7d"} Feb 27 07:51:02 crc kubenswrapper[4612]: I0227 07:51:02.591478 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-f8qzz" event={"ID":"3b5b5090-45a4-4acc-aa94-84c892bcf306","Type":"ContainerStarted","Data":"716c8106fdd6c7eec4097ef44470fc71de31e6789e6a7412b14ae979b74910eb"} Feb 27 07:51:02 crc kubenswrapper[4612]: I0227 07:51:02.610097 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 07:51:02 crc kubenswrapper[4612]: E0227 07:51:02.610530 4612 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 07:51:03.110513615 +0000 UTC m=+120.964443613 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 07:51:02 crc kubenswrapper[4612]: I0227 07:51:02.629662 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-k6s2g" event={"ID":"29ab8b79-a05e-4b94-a2bb-2c6b412fd297","Type":"ContainerStarted","Data":"639457bfe52f62fc1c84f36a67a4eb725db43f3f9c6d30e8f0a27b4c219dfe96"} Feb 27 07:51:02 crc kubenswrapper[4612]: I0227 07:51:02.638473 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-c6sdx" event={"ID":"a6e73baa-3a12-462c-b85c-1b964a52e38f","Type":"ContainerStarted","Data":"9b098cf1a52144fc8a3c77ed75d97aaec02e76d1d094ff2e7d8ca0cd6be63133"} Feb 27 07:51:02 crc kubenswrapper[4612]: I0227 07:51:02.650219 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-2d7kh" podStartSLOduration=59.650198918 podStartE2EDuration="59.650198918s" podCreationTimestamp="2026-02-27 07:50:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 07:51:02.608801807 +0000 UTC m=+120.462731805" watchObservedRunningTime="2026-02-27 07:51:02.650198918 +0000 UTC m=+120.504128916" Feb 27 07:51:02 crc kubenswrapper[4612]: I0227 07:51:02.658602 4612 patch_prober.go:28] interesting pod/router-default-5444994796-2d7kh container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 27 07:51:02 crc kubenswrapper[4612]: [-]has-synced failed: reason withheld Feb 27 07:51:02 crc kubenswrapper[4612]: [+]process-running ok Feb 27 07:51:02 crc kubenswrapper[4612]: healthz check failed Feb 27 07:51:02 crc kubenswrapper[4612]: I0227 07:51:02.658636 4612 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-2d7kh" podUID="faae1773-bf96-4127-8060-a40da07930ce" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 27 07:51:02 crc kubenswrapper[4612]: I0227 07:51:02.661102 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-8wk54" event={"ID":"da284312-b98a-48d6-8cff-f8071c334e0b","Type":"ContainerStarted","Data":"f7fc77f7a6fb23f2d47be4c83b3939f6ce2f3662d142d7a4bd154ce263f0f715"} Feb 27 07:51:02 crc kubenswrapper[4612]: I0227 07:51:02.708789 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-cw57r" event={"ID":"25a0e1aa-86e2-4acf-885b-16c26efc337a","Type":"ContainerStarted","Data":"beff58bb672ac4b7caf9b3a33f6d6cf5c3ed5bc7ae237925e7e0244b6857620a"} Feb 27 07:51:02 crc kubenswrapper[4612]: I0227 07:51:02.711375 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pdjlb\" (UID: \"db4d6d91-61f5-4beb-809c-63ae9028a911\") " pod="openshift-image-registry/image-registry-697d97f7c8-pdjlb" Feb 27 07:51:02 crc kubenswrapper[4612]: E0227 07:51:02.712919 4612 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 07:51:03.212905406 +0000 UTC m=+121.066835404 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pdjlb" (UID: "db4d6d91-61f5-4beb-809c-63ae9028a911") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 07:51:02 crc kubenswrapper[4612]: I0227 07:51:02.768057 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-z9zch" event={"ID":"9bb5eece-3596-485e-a859-b61d201b05f5","Type":"ContainerStarted","Data":"fc4fe9a756496d498b0cc8e97eb045fa841e77c917edf125fed3710ed176e178"} Feb 27 07:51:02 crc kubenswrapper[4612]: I0227 07:51:02.798176 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-spnfb" event={"ID":"42b49d65-d823-45af-b274-380dfbddc9b7","Type":"ContainerStarted","Data":"1e8c7d56859a3f57ec388590f4f8362847e655b83c149dcef9376df449351658"} Feb 27 07:51:02 crc kubenswrapper[4612]: I0227 07:51:02.798299 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-spnfb" event={"ID":"42b49d65-d823-45af-b274-380dfbddc9b7","Type":"ContainerStarted","Data":"5802342342fd2e28bb97f45d7fe121c64b9fb737dd9987195385f7c56553f64b"} Feb 27 07:51:02 crc kubenswrapper[4612]: I0227 07:51:02.814907 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 07:51:02 crc kubenswrapper[4612]: E0227 07:51:02.815510 4612 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 07:51:03.315497523 +0000 UTC m=+121.169427521 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 07:51:02 crc kubenswrapper[4612]: I0227 07:51:02.843254 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-q6h7h" event={"ID":"094b001c-e80c-4d07-a161-e00514667462","Type":"ContainerStarted","Data":"53d6179d502aaed67ab7b33f6f5a4ebbaa10167386c4a5cb4a4e01b0436241e4"} Feb 27 07:51:02 crc kubenswrapper[4612]: I0227 07:51:02.941461 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-g9hq6" event={"ID":"cc3f4dca-b740-4294-93e5-73906f8dc82c","Type":"ContainerStarted","Data":"f71305c2ac0a51de4a304c0667eaadcec8adf683497c552d5911fc8e5885f644"} Feb 27 07:51:02 crc kubenswrapper[4612]: I0227 07:51:02.942369 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-g9hq6" event={"ID":"cc3f4dca-b740-4294-93e5-73906f8dc82c","Type":"ContainerStarted","Data":"0e6a099224652b1023d89e8c24b9594f2cdc31cec4c46fff1c5e8a37b144f385"} Feb 27 07:51:02 crc kubenswrapper[4612]: I0227 07:51:02.952225 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pdjlb\" (UID: \"db4d6d91-61f5-4beb-809c-63ae9028a911\") " pod="openshift-image-registry/image-registry-697d97f7c8-pdjlb" Feb 27 07:51:02 crc kubenswrapper[4612]: E0227 07:51:02.955875 4612 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 07:51:03.455852651 +0000 UTC m=+121.309782649 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pdjlb" (UID: "db4d6d91-61f5-4beb-809c-63ae9028a911") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 07:51:02 crc kubenswrapper[4612]: I0227 07:51:02.978492 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2rrfz" event={"ID":"d9aa8f66-da03-49e5-b7b4-d17135ccaa03","Type":"ContainerStarted","Data":"3ab029f4c976c77717e65922928926220bb94693d66bac6400f01d8be947c572"} Feb 27 07:51:03 crc kubenswrapper[4612]: I0227 07:51:03.047890 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-96vl6"] Feb 27 07:51:03 crc kubenswrapper[4612]: I0227 07:51:03.052682 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-8l852"] Feb 27 07:51:03 crc kubenswrapper[4612]: I0227 07:51:03.054401 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 07:51:03 crc kubenswrapper[4612]: E0227 07:51:03.055967 4612 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 07:51:03.555948899 +0000 UTC m=+121.409878897 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 07:51:03 crc kubenswrapper[4612]: I0227 07:51:03.065745 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-sxt8c"] Feb 27 07:51:03 crc kubenswrapper[4612]: I0227 07:51:03.066092 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-qvwzc" event={"ID":"19b65635-c86d-4439-8ece-4d8c3dbe986b","Type":"ContainerStarted","Data":"8318f9c65488d67a52adc45c3ab5f9789f5e2e573b74ee45829788f759e13722"} Feb 27 07:51:03 crc kubenswrapper[4612]: I0227 07:51:03.066958 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-qvwzc" Feb 27 07:51:03 crc kubenswrapper[4612]: I0227 07:51:03.086117 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-kbtxf" event={"ID":"98e7f73e-ab59-4e4d-af52-4794f9e28abb","Type":"ContainerStarted","Data":"8f6bb6810dc034c5d765afd2a79f681dc57bfe775bd046d091aa91d7a261875d"} Feb 27 07:51:03 crc kubenswrapper[4612]: I0227 07:51:03.086408 4612 patch_prober.go:28] interesting pod/console-operator-58897d9998-qvwzc container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.17:8443/readyz\": dial tcp 10.217.0.17:8443: connect: connection refused" start-of-body= Feb 27 07:51:03 crc kubenswrapper[4612]: I0227 07:51:03.086445 4612 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-qvwzc" podUID="19b65635-c86d-4439-8ece-4d8c3dbe986b" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.17:8443/readyz\": dial tcp 10.217.0.17:8443: connect: connection refused" Feb 27 07:51:03 crc kubenswrapper[4612]: I0227 07:51:03.115831 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bk49v" event={"ID":"fb198883-74b8-43a2-a2db-390e951fda9a","Type":"ContainerStarted","Data":"3d21c3631a46e51b61430f6da4d748ed54db8d1f3289c024aada20919a0182a4"} Feb 27 07:51:03 crc kubenswrapper[4612]: I0227 07:51:03.139887 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-nkbg2" event={"ID":"6ece12fd-d439-4104-a7a9-a9d174e29b5a","Type":"ContainerStarted","Data":"337d6f75cb558f81002522a58efc2f5296e04ef640a355c38af791aaf38647a3"} Feb 27 07:51:03 crc kubenswrapper[4612]: I0227 07:51:03.144111 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-nkbg2" Feb 27 07:51:03 crc kubenswrapper[4612]: I0227 07:51:03.157644 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pdjlb\" (UID: \"db4d6d91-61f5-4beb-809c-63ae9028a911\") " pod="openshift-image-registry/image-registry-697d97f7c8-pdjlb" Feb 27 07:51:03 crc kubenswrapper[4612]: E0227 07:51:03.160326 4612 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 07:51:03.660308855 +0000 UTC m=+121.514238933 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pdjlb" (UID: "db4d6d91-61f5-4beb-809c-63ae9028a911") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 07:51:03 crc kubenswrapper[4612]: I0227 07:51:03.162173 4612 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-nkbg2 container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.11:6443/healthz\": dial tcp 10.217.0.11:6443: connect: connection refused" start-of-body= Feb 27 07:51:03 crc kubenswrapper[4612]: I0227 07:51:03.162218 4612 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-nkbg2" podUID="6ece12fd-d439-4104-a7a9-a9d174e29b5a" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.11:6443/healthz\": dial tcp 10.217.0.11:6443: connect: connection refused" Feb 27 07:51:03 crc kubenswrapper[4612]: I0227 07:51:03.185290 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-jzd5b" event={"ID":"20e2f07c-3a73-4d78-88e9-c12b093743bb","Type":"ContainerStarted","Data":"44119b098ac158b60ba934b7998c69fb0f31f2fbc49af314005f6dbacdea2ad4"} Feb 27 07:51:03 crc kubenswrapper[4612]: I0227 07:51:03.186970 4612 patch_prober.go:28] interesting pod/downloads-7954f5f757-zbccb container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.9:8080/\": dial tcp 10.217.0.9:8080: connect: connection refused" start-of-body= Feb 27 07:51:03 crc kubenswrapper[4612]: I0227 07:51:03.187017 4612 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-zbccb" podUID="a5bbba10-16da-4fce-ae5f-311fc5d0d6c1" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.9:8080/\": dial tcp 10.217.0.9:8080: connect: connection refused" Feb 27 07:51:03 crc kubenswrapper[4612]: I0227 07:51:03.190980 4612 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-hp72h container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.19:8080/healthz\": dial tcp 10.217.0.19:8080: connect: connection refused" start-of-body= Feb 27 07:51:03 crc kubenswrapper[4612]: I0227 07:51:03.191019 4612 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-hp72h" podUID="6d061f42-06de-43e5-b4d0-dbca09e48630" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.19:8080/healthz\": dial tcp 10.217.0.19:8080: connect: connection refused" Feb 27 07:51:03 crc kubenswrapper[4612]: I0227 07:51:03.216219 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-hp72h" podStartSLOduration=60.216201486 podStartE2EDuration="1m0.216201486s" podCreationTimestamp="2026-02-27 07:50:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 07:51:03.215107495 +0000 UTC m=+121.069037493" watchObservedRunningTime="2026-02-27 07:51:03.216201486 +0000 UTC m=+121.070131484" Feb 27 07:51:03 crc kubenswrapper[4612]: I0227 07:51:03.248899 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2k66r" podStartSLOduration=60.248881886 podStartE2EDuration="1m0.248881886s" podCreationTimestamp="2026-02-27 07:50:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 07:51:03.244184477 +0000 UTC m=+121.098114475" watchObservedRunningTime="2026-02-27 07:51:03.248881886 +0000 UTC m=+121.102811884" Feb 27 07:51:03 crc kubenswrapper[4612]: I0227 07:51:03.254220 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-kgkd4"] Feb 27 07:51:03 crc kubenswrapper[4612]: I0227 07:51:03.260781 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 07:51:03 crc kubenswrapper[4612]: E0227 07:51:03.261735 4612 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 07:51:03.76172054 +0000 UTC m=+121.615650538 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 07:51:03 crc kubenswrapper[4612]: I0227 07:51:03.265725 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-w7pv5" podStartSLOduration=60.26570945 podStartE2EDuration="1m0.26570945s" podCreationTimestamp="2026-02-27 07:50:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 07:51:03.264232279 +0000 UTC m=+121.118162277" watchObservedRunningTime="2026-02-27 07:51:03.26570945 +0000 UTC m=+121.119639468" Feb 27 07:51:03 crc kubenswrapper[4612]: I0227 07:51:03.362087 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pdjlb\" (UID: \"db4d6d91-61f5-4beb-809c-63ae9028a911\") " pod="openshift-image-registry/image-registry-697d97f7c8-pdjlb" Feb 27 07:51:03 crc kubenswrapper[4612]: E0227 07:51:03.371342 4612 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 07:51:03.87132547 +0000 UTC m=+121.725255548 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pdjlb" (UID: "db4d6d91-61f5-4beb-809c-63ae9028a911") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 07:51:03 crc kubenswrapper[4612]: I0227 07:51:03.497625 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 07:51:03 crc kubenswrapper[4612]: E0227 07:51:03.507864 4612 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 07:51:04.007844462 +0000 UTC m=+121.861774460 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 07:51:03 crc kubenswrapper[4612]: I0227 07:51:03.540638 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-zbccb" podStartSLOduration=60.540614545 podStartE2EDuration="1m0.540614545s" podCreationTimestamp="2026-02-27 07:50:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 07:51:03.494549216 +0000 UTC m=+121.348479214" watchObservedRunningTime="2026-02-27 07:51:03.540614545 +0000 UTC m=+121.394544543" Feb 27 07:51:03 crc kubenswrapper[4612]: I0227 07:51:03.543409 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pdjlb\" (UID: \"db4d6d91-61f5-4beb-809c-63ae9028a911\") " pod="openshift-image-registry/image-registry-697d97f7c8-pdjlb" Feb 27 07:51:03 crc kubenswrapper[4612]: E0227 07:51:03.543846 4612 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 07:51:04.043835574 +0000 UTC m=+121.897765572 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pdjlb" (UID: "db4d6d91-61f5-4beb-809c-63ae9028a911") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 07:51:03 crc kubenswrapper[4612]: I0227 07:51:03.571951 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-m5b65" podStartSLOduration=60.571932858 podStartE2EDuration="1m0.571932858s" podCreationTimestamp="2026-02-27 07:50:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 07:51:03.570260332 +0000 UTC m=+121.424190330" watchObservedRunningTime="2026-02-27 07:51:03.571932858 +0000 UTC m=+121.425862856" Feb 27 07:51:03 crc kubenswrapper[4612]: I0227 07:51:03.650128 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 07:51:03 crc kubenswrapper[4612]: E0227 07:51:03.650513 4612 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 07:51:04.150498623 +0000 UTC m=+122.004428611 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 07:51:03 crc kubenswrapper[4612]: I0227 07:51:03.658883 4612 patch_prober.go:28] interesting pod/router-default-5444994796-2d7kh container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 27 07:51:03 crc kubenswrapper[4612]: [-]has-synced failed: reason withheld Feb 27 07:51:03 crc kubenswrapper[4612]: [+]process-running ok Feb 27 07:51:03 crc kubenswrapper[4612]: healthz check failed Feb 27 07:51:03 crc kubenswrapper[4612]: I0227 07:51:03.658933 4612 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-2d7kh" podUID="faae1773-bf96-4127-8060-a40da07930ce" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 27 07:51:03 crc kubenswrapper[4612]: I0227 07:51:03.673576 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-ks964"] Feb 27 07:51:03 crc kubenswrapper[4612]: I0227 07:51:03.720185 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-n5jbm"] Feb 27 07:51:03 crc kubenswrapper[4612]: I0227 07:51:03.754245 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pdjlb\" (UID: \"db4d6d91-61f5-4beb-809c-63ae9028a911\") " pod="openshift-image-registry/image-registry-697d97f7c8-pdjlb" Feb 27 07:51:03 crc kubenswrapper[4612]: E0227 07:51:03.755118 4612 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 07:51:04.255104886 +0000 UTC m=+122.109034884 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pdjlb" (UID: "db4d6d91-61f5-4beb-809c-63ae9028a911") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 07:51:03 crc kubenswrapper[4612]: I0227 07:51:03.781160 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-nkbg2" podStartSLOduration=60.781145084 podStartE2EDuration="1m0.781145084s" podCreationTimestamp="2026-02-27 07:50:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 07:51:03.71933845 +0000 UTC m=+121.573268448" watchObservedRunningTime="2026-02-27 07:51:03.781145084 +0000 UTC m=+121.635075082" Feb 27 07:51:03 crc kubenswrapper[4612]: I0227 07:51:03.788455 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-kv4nc"] Feb 27 07:51:03 crc kubenswrapper[4612]: I0227 07:51:03.831324 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29536305-vmtnx"] Feb 27 07:51:03 crc kubenswrapper[4612]: I0227 07:51:03.865240 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 07:51:03 crc kubenswrapper[4612]: E0227 07:51:03.865848 4612 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 07:51:04.365797526 +0000 UTC m=+122.219727524 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 07:51:03 crc kubenswrapper[4612]: I0227 07:51:03.868767 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-gcfmk"] Feb 27 07:51:03 crc kubenswrapper[4612]: I0227 07:51:03.875357 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-spnfb" podStartSLOduration=60.875340279 podStartE2EDuration="1m0.875340279s" podCreationTimestamp="2026-02-27 07:50:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 07:51:03.778774818 +0000 UTC m=+121.632704816" watchObservedRunningTime="2026-02-27 07:51:03.875340279 +0000 UTC m=+121.729270277" Feb 27 07:51:03 crc kubenswrapper[4612]: I0227 07:51:03.876835 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-f8qzz" podStartSLOduration=60.87682861 podStartE2EDuration="1m0.87682861s" podCreationTimestamp="2026-02-27 07:50:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 07:51:03.811988284 +0000 UTC m=+121.665918282" watchObservedRunningTime="2026-02-27 07:51:03.87682861 +0000 UTC m=+121.730758608" Feb 27 07:51:03 crc kubenswrapper[4612]: I0227 07:51:03.919884 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-qvwzc" podStartSLOduration=60.919866746 podStartE2EDuration="1m0.919866746s" podCreationTimestamp="2026-02-27 07:50:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 07:51:03.878059824 +0000 UTC m=+121.731989822" watchObservedRunningTime="2026-02-27 07:51:03.919866746 +0000 UTC m=+121.773796744" Feb 27 07:51:03 crc kubenswrapper[4612]: E0227 07:51:03.922046 4612 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9bb5eece_3596_485e_a859_b61d201b05f5.slice/crio-conmon-e03fde18a73e3e017f836d1b17730864651f827b7e1bb9e31de738a820bfc924.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9bb5eece_3596_485e_a859_b61d201b05f5.slice/crio-e03fde18a73e3e017f836d1b17730864651f827b7e1bb9e31de738a820bfc924.scope\": RecentStats: unable to find data in memory cache]" Feb 27 07:51:03 crc kubenswrapper[4612]: I0227 07:51:03.924606 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-8wk54" podStartSLOduration=60.924588936 podStartE2EDuration="1m0.924588936s" podCreationTimestamp="2026-02-27 07:50:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 07:51:03.90365967 +0000 UTC m=+121.757589668" watchObservedRunningTime="2026-02-27 07:51:03.924588936 +0000 UTC m=+121.778518934" Feb 27 07:51:03 crc kubenswrapper[4612]: I0227 07:51:03.929765 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-jzd5b" podStartSLOduration=60.929751869 podStartE2EDuration="1m0.929751869s" podCreationTimestamp="2026-02-27 07:50:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 07:51:03.928314249 +0000 UTC m=+121.782244237" watchObservedRunningTime="2026-02-27 07:51:03.929751869 +0000 UTC m=+121.783681867" Feb 27 07:51:03 crc kubenswrapper[4612]: I0227 07:51:03.966147 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pdjlb\" (UID: \"db4d6d91-61f5-4beb-809c-63ae9028a911\") " pod="openshift-image-registry/image-registry-697d97f7c8-pdjlb" Feb 27 07:51:03 crc kubenswrapper[4612]: E0227 07:51:03.968228 4612 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 07:51:04.468213479 +0000 UTC m=+122.322143477 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pdjlb" (UID: "db4d6d91-61f5-4beb-809c-63ae9028a911") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 07:51:04 crc kubenswrapper[4612]: I0227 07:51:04.047422 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zpshm"] Feb 27 07:51:04 crc kubenswrapper[4612]: I0227 07:51:04.073503 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 07:51:04 crc kubenswrapper[4612]: E0227 07:51:04.074343 4612 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 07:51:04.574326723 +0000 UTC m=+122.428256721 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 07:51:04 crc kubenswrapper[4612]: I0227 07:51:04.074375 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-42tqr"] Feb 27 07:51:04 crc kubenswrapper[4612]: I0227 07:51:04.103264 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-6q4q2"] Feb 27 07:51:04 crc kubenswrapper[4612]: W0227 07:51:04.108931 4612 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda61e7e5a_48f6_4b26_9b1e_1858222ff35f.slice/crio-2514a5958648bd41447cb003cd1548f24b0d93cb1dd3170f26d5232c3aa7bef6 WatchSource:0}: Error finding container 2514a5958648bd41447cb003cd1548f24b0d93cb1dd3170f26d5232c3aa7bef6: Status 404 returned error can't find the container with id 2514a5958648bd41447cb003cd1548f24b0d93cb1dd3170f26d5232c3aa7bef6 Feb 27 07:51:04 crc kubenswrapper[4612]: I0227 07:51:04.115931 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-4lpn2"] Feb 27 07:51:04 crc kubenswrapper[4612]: I0227 07:51:04.166778 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-gvv4x"] Feb 27 07:51:04 crc kubenswrapper[4612]: I0227 07:51:04.190562 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pdjlb\" (UID: \"db4d6d91-61f5-4beb-809c-63ae9028a911\") " pod="openshift-image-registry/image-registry-697d97f7c8-pdjlb" Feb 27 07:51:04 crc kubenswrapper[4612]: E0227 07:51:04.190978 4612 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 07:51:04.690967447 +0000 UTC m=+122.544897445 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pdjlb" (UID: "db4d6d91-61f5-4beb-809c-63ae9028a911") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 07:51:04 crc kubenswrapper[4612]: I0227 07:51:04.205175 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6cz6v"] Feb 27 07:51:04 crc kubenswrapper[4612]: I0227 07:51:04.273345 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-42tqr" event={"ID":"a61e7e5a-48f6-4b26-9b1e-1858222ff35f","Type":"ContainerStarted","Data":"2514a5958648bd41447cb003cd1548f24b0d93cb1dd3170f26d5232c3aa7bef6"} Feb 27 07:51:04 crc kubenswrapper[4612]: I0227 07:51:04.291614 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 07:51:04 crc kubenswrapper[4612]: E0227 07:51:04.291999 4612 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 07:51:04.791980451 +0000 UTC m=+122.645910439 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 07:51:04 crc kubenswrapper[4612]: I0227 07:51:04.293379 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-bqh6n"] Feb 27 07:51:04 crc kubenswrapper[4612]: I0227 07:51:04.295364 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-qvwzc" event={"ID":"19b65635-c86d-4439-8ece-4d8c3dbe986b","Type":"ContainerStarted","Data":"b8e5ede809860c90ab14be9d3c1c52fda779847f396b412fa3ef94bc9fac3547"} Feb 27 07:51:04 crc kubenswrapper[4612]: I0227 07:51:04.308210 4612 patch_prober.go:28] interesting pod/console-operator-58897d9998-qvwzc container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.17:8443/readyz\": dial tcp 10.217.0.17:8443: connect: connection refused" start-of-body= Feb 27 07:51:04 crc kubenswrapper[4612]: I0227 07:51:04.308499 4612 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-qvwzc" podUID="19b65635-c86d-4439-8ece-4d8c3dbe986b" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.17:8443/readyz\": dial tcp 10.217.0.17:8443: connect: connection refused" Feb 27 07:51:04 crc kubenswrapper[4612]: I0227 07:51:04.316386 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-ks964" event={"ID":"0b1512ff-906a-4df2-b0ca-a09ba21f4875","Type":"ContainerStarted","Data":"50eccc3e9eba66ce924b4d74688666ec812ab0f0532981d0263419d6d5c68640"} Feb 27 07:51:04 crc kubenswrapper[4612]: I0227 07:51:04.343257 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-k6s2g" event={"ID":"29ab8b79-a05e-4b94-a2bb-2c6b412fd297","Type":"ContainerStarted","Data":"e676766fe8c1d3700874aeb3b261ffda87b1d40bd73e8c7c38ede29afa605151"} Feb 27 07:51:04 crc kubenswrapper[4612]: I0227 07:51:04.355280 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-4lpn2" event={"ID":"4138a6e8-dd4e-44d3-9103-8f328b794227","Type":"ContainerStarted","Data":"41e9e43e14f5973763aed16cc8eef6c43226b8962d0d3bf799283b2f4ab8e8b7"} Feb 27 07:51:04 crc kubenswrapper[4612]: W0227 07:51:04.368134 4612 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9421f529_9b1e_4bd0_8032_7baa608981b4.slice/crio-c01519a8b2d4b37110f2966d90389d8ece791b2db3b0def6040c6291f38a500c WatchSource:0}: Error finding container c01519a8b2d4b37110f2966d90389d8ece791b2db3b0def6040c6291f38a500c: Status 404 returned error can't find the container with id c01519a8b2d4b37110f2966d90389d8ece791b2db3b0def6040c6291f38a500c Feb 27 07:51:04 crc kubenswrapper[4612]: I0227 07:51:04.369117 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-k6s2g" podStartSLOduration=61.369103716 podStartE2EDuration="1m1.369103716s" podCreationTimestamp="2026-02-27 07:50:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 07:51:04.368475769 +0000 UTC m=+122.222405757" watchObservedRunningTime="2026-02-27 07:51:04.369103716 +0000 UTC m=+122.223033714" Feb 27 07:51:04 crc kubenswrapper[4612]: I0227 07:51:04.388419 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-tcqv8" event={"ID":"efc3c589-c379-45ca-a974-75483d90c5a4","Type":"ContainerStarted","Data":"f299fb0c13a67d10a995c84c9d2ad5eab8e54acab9759ecc8527fc442b8836ed"} Feb 27 07:51:04 crc kubenswrapper[4612]: I0227 07:51:04.397535 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pdjlb\" (UID: \"db4d6d91-61f5-4beb-809c-63ae9028a911\") " pod="openshift-image-registry/image-registry-697d97f7c8-pdjlb" Feb 27 07:51:04 crc kubenswrapper[4612]: E0227 07:51:04.398732 4612 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 07:51:04.898715832 +0000 UTC m=+122.752645830 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pdjlb" (UID: "db4d6d91-61f5-4beb-809c-63ae9028a911") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 07:51:04 crc kubenswrapper[4612]: I0227 07:51:04.407546 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-8l852" event={"ID":"cc5b15e4-6c84-4474-bf4d-f652d2e8000c","Type":"ContainerStarted","Data":"178dbdb38e21b029ae764c825792efc2f15e56450c0ae509c98a5a98f978cd81"} Feb 27 07:51:04 crc kubenswrapper[4612]: I0227 07:51:04.409629 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-96vl6" event={"ID":"620a7a2f-92fc-4a76-900b-021d0dd78d79","Type":"ContainerStarted","Data":"9204fa681c4cb9f5f0ccce6e0d8e261c93ae3e79a4b4207d5fcb0ed72e21efe6"} Feb 27 07:51:04 crc kubenswrapper[4612]: I0227 07:51:04.431864 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-kv4nc" event={"ID":"fa91e393-fed8-4f4e-8dbe-7bc51ed35cea","Type":"ContainerStarted","Data":"d8e6dbdef23e249cd5b295d3b8f03cff959f14317706f0720edfa45618fd9e1e"} Feb 27 07:51:04 crc kubenswrapper[4612]: I0227 07:51:04.474924 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-gvv4x" event={"ID":"c02d88e5-7055-4864-8b83-d69fdccad379","Type":"ContainerStarted","Data":"c4f1e246cad21e9fd74c95d0a26213f3e2ccda99b8d89e0bb40b360984d7be6c"} Feb 27 07:51:04 crc kubenswrapper[4612]: I0227 07:51:04.492846 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-6q4q2" event={"ID":"e24e170c-7ee2-4d40-badf-21bc59723d1c","Type":"ContainerStarted","Data":"691f528c0cb78d62106ccb60e0abee7889489a30a97fa3937b6e2a5d4fd616d2"} Feb 27 07:51:04 crc kubenswrapper[4612]: I0227 07:51:04.497883 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 07:51:04 crc kubenswrapper[4612]: E0227 07:51:04.498134 4612 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 07:51:04.998120561 +0000 UTC m=+122.852050559 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 07:51:04 crc kubenswrapper[4612]: I0227 07:51:04.502462 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-g9hq6" event={"ID":"cc3f4dca-b740-4294-93e5-73906f8dc82c","Type":"ContainerStarted","Data":"5a6c1344195dc2ec8e08c80e0ebce60e8d2dc79755f9e551207c7a03b57c37d3"} Feb 27 07:51:04 crc kubenswrapper[4612]: I0227 07:51:04.506946 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bk49v" Feb 27 07:51:04 crc kubenswrapper[4612]: I0227 07:51:04.517353 4612 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-bk49v container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.18:8443/healthz\": dial tcp 10.217.0.18:8443: connect: connection refused" start-of-body= Feb 27 07:51:04 crc kubenswrapper[4612]: I0227 07:51:04.517407 4612 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bk49v" podUID="fb198883-74b8-43a2-a2db-390e951fda9a" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.18:8443/healthz\": dial tcp 10.217.0.18:8443: connect: connection refused" Feb 27 07:51:04 crc kubenswrapper[4612]: I0227 07:51:04.519339 4612 generic.go:334] "Generic (PLEG): container finished" podID="d9aa8f66-da03-49e5-b7b4-d17135ccaa03" containerID="bb8a601e474672fbf76a0b96cac7d089492a22ed977a39950b8bd98be9a67922" exitCode=0 Feb 27 07:51:04 crc kubenswrapper[4612]: I0227 07:51:04.519399 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2rrfz" event={"ID":"d9aa8f66-da03-49e5-b7b4-d17135ccaa03","Type":"ContainerDied","Data":"bb8a601e474672fbf76a0b96cac7d089492a22ed977a39950b8bd98be9a67922"} Feb 27 07:51:04 crc kubenswrapper[4612]: I0227 07:51:04.530630 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-c6sdx" event={"ID":"a6e73baa-3a12-462c-b85c-1b964a52e38f","Type":"ContainerStarted","Data":"8f0e54f8ae3a537c827797674bd0870c7f12291f551f9b7ad0f46021b3a65cb3"} Feb 27 07:51:04 crc kubenswrapper[4612]: I0227 07:51:04.531147 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-c6sdx" Feb 27 07:51:04 crc kubenswrapper[4612]: I0227 07:51:04.533587 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-gcfmk" event={"ID":"c28eca4e-4d90-4bf0-95bf-880f3a9037e3","Type":"ContainerStarted","Data":"2395971e65a17fd1dbc3541a1c2cf11b45ff8c41f671c95b36de72f6b4e4c3c0"} Feb 27 07:51:04 crc kubenswrapper[4612]: I0227 07:51:04.541858 4612 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-c6sdx container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.28:5443/healthz\": dial tcp 10.217.0.28:5443: connect: connection refused" start-of-body= Feb 27 07:51:04 crc kubenswrapper[4612]: I0227 07:51:04.541917 4612 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-c6sdx" podUID="a6e73baa-3a12-462c-b85c-1b964a52e38f" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.28:5443/healthz\": dial tcp 10.217.0.28:5443: connect: connection refused" Feb 27 07:51:04 crc kubenswrapper[4612]: I0227 07:51:04.542658 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-g9hq6" podStartSLOduration=61.542636728 podStartE2EDuration="1m1.542636728s" podCreationTimestamp="2026-02-27 07:50:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 07:51:04.542080273 +0000 UTC m=+122.396010271" watchObservedRunningTime="2026-02-27 07:51:04.542636728 +0000 UTC m=+122.396566736" Feb 27 07:51:04 crc kubenswrapper[4612]: I0227 07:51:04.543572 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-5wvhc" event={"ID":"ff5fcfe1-eb01-4f01-857d-a67712db3b87","Type":"ContainerStarted","Data":"078090cc7f631ae3cd1a11bc7faaddbb60186085cbbeeea6952b5e865e481340"} Feb 27 07:51:04 crc kubenswrapper[4612]: I0227 07:51:04.561394 4612 generic.go:334] "Generic (PLEG): container finished" podID="9bb5eece-3596-485e-a859-b61d201b05f5" containerID="e03fde18a73e3e017f836d1b17730864651f827b7e1bb9e31de738a820bfc924" exitCode=0 Feb 27 07:51:04 crc kubenswrapper[4612]: I0227 07:51:04.561455 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-z9zch" event={"ID":"9bb5eece-3596-485e-a859-b61d201b05f5","Type":"ContainerDied","Data":"e03fde18a73e3e017f836d1b17730864651f827b7e1bb9e31de738a820bfc924"} Feb 27 07:51:04 crc kubenswrapper[4612]: I0227 07:51:04.567979 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6cz6v" event={"ID":"90d5c002-a228-4c36-b395-5ba0b01aac8e","Type":"ContainerStarted","Data":"755b472df9f3489afbb62b2590f90132473fd92aaa624dc8eb8fdfc033e7bbec"} Feb 27 07:51:04 crc kubenswrapper[4612]: I0227 07:51:04.587100 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-jzd5b" event={"ID":"20e2f07c-3a73-4d78-88e9-c12b093743bb","Type":"ContainerStarted","Data":"74275821b32ae58b7a543379c34201f0ced844badfa05def086ddf9a118d863d"} Feb 27 07:51:04 crc kubenswrapper[4612]: I0227 07:51:04.600749 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pdjlb\" (UID: \"db4d6d91-61f5-4beb-809c-63ae9028a911\") " pod="openshift-image-registry/image-registry-697d97f7c8-pdjlb" Feb 27 07:51:04 crc kubenswrapper[4612]: E0227 07:51:04.601101 4612 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 07:51:05.101084409 +0000 UTC m=+122.955014407 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pdjlb" (UID: "db4d6d91-61f5-4beb-809c-63ae9028a911") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 07:51:04 crc kubenswrapper[4612]: I0227 07:51:04.602005 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-kgkd4" event={"ID":"b979d24b-0d0c-43c1-b50c-52e6b3801daf","Type":"ContainerStarted","Data":"baef74036bc1e99f67533780dbf4986fba5ba1aef24adb46d6676b59672adeaa"} Feb 27 07:51:04 crc kubenswrapper[4612]: I0227 07:51:04.604377 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zpshm" event={"ID":"d951f066-9e39-4e0a-9fa3-da9c868552d1","Type":"ContainerStarted","Data":"cccd5812e66a7963c55fe853a7b8b5e76949bdb4fc55b840cec8e2fed7f9dff1"} Feb 27 07:51:04 crc kubenswrapper[4612]: I0227 07:51:04.619009 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-c6sdx" podStartSLOduration=61.618994292 podStartE2EDuration="1m1.618994292s" podCreationTimestamp="2026-02-27 07:50:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 07:51:04.617613624 +0000 UTC m=+122.471543622" watchObservedRunningTime="2026-02-27 07:51:04.618994292 +0000 UTC m=+122.472924290" Feb 27 07:51:04 crc kubenswrapper[4612]: I0227 07:51:04.620374 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bk49v" podStartSLOduration=61.62036478 podStartE2EDuration="1m1.62036478s" podCreationTimestamp="2026-02-27 07:50:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 07:51:04.583109513 +0000 UTC m=+122.437039511" watchObservedRunningTime="2026-02-27 07:51:04.62036478 +0000 UTC m=+122.474294778" Feb 27 07:51:04 crc kubenswrapper[4612]: I0227 07:51:04.622837 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-ghtdk" event={"ID":"f0d8e4b4-aec4-433c-adff-091c8944e8a3","Type":"ContainerStarted","Data":"519ef2ee7dc63b38a29adb58810eee8d0255ce1e4d231aca85702c469ddff33e"} Feb 27 07:51:04 crc kubenswrapper[4612]: I0227 07:51:04.629949 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29536305-vmtnx" event={"ID":"ea3d0df0-517f-4865-964d-3e12a313696b","Type":"ContainerStarted","Data":"bf05a7f96e03026c6cbda7c4f30a3f4f88ae90748101d570a2eaf5bc3d545924"} Feb 27 07:51:04 crc kubenswrapper[4612]: I0227 07:51:04.633532 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-q6h7h" event={"ID":"094b001c-e80c-4d07-a161-e00514667462","Type":"ContainerStarted","Data":"fc4041afacdac6d1a5857251b668d0028bef980fc2750c7dec54cfadf6f1bd2a"} Feb 27 07:51:04 crc kubenswrapper[4612]: I0227 07:51:04.644339 4612 patch_prober.go:28] interesting pod/router-default-5444994796-2d7kh container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 27 07:51:04 crc kubenswrapper[4612]: [-]has-synced failed: reason withheld Feb 27 07:51:04 crc kubenswrapper[4612]: [+]process-running ok Feb 27 07:51:04 crc kubenswrapper[4612]: healthz check failed Feb 27 07:51:04 crc kubenswrapper[4612]: I0227 07:51:04.644373 4612 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-2d7kh" podUID="faae1773-bf96-4127-8060-a40da07930ce" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 27 07:51:04 crc kubenswrapper[4612]: I0227 07:51:04.690630 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-sxt8c" event={"ID":"738d2b6e-2b7c-461d-b51e-f35eb5d41363","Type":"ContainerStarted","Data":"ffadd5b41c89bb82f7b5c02114da43182b61205b4abd163256d5cbee5663e237"} Feb 27 07:51:04 crc kubenswrapper[4612]: I0227 07:51:04.690829 4612 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-sxt8c" podUID="738d2b6e-2b7c-461d-b51e-f35eb5d41363" containerName="controller-manager" containerID="cri-o://80058f3d8b6825a6aa02bba10ea42f4f51630b27d2fc39c813f4cc5bb5412319" gracePeriod=30 Feb 27 07:51:04 crc kubenswrapper[4612]: I0227 07:51:04.691283 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-sxt8c" Feb 27 07:51:04 crc kubenswrapper[4612]: I0227 07:51:04.693126 4612 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-spnfb" podUID="42b49d65-d823-45af-b274-380dfbddc9b7" containerName="route-controller-manager" containerID="cri-o://1e8c7d56859a3f57ec388590f4f8362847e655b83c149dcef9376df449351658" gracePeriod=30 Feb 27 07:51:04 crc kubenswrapper[4612]: I0227 07:51:04.693181 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-n5jbm" event={"ID":"ad4c7d8c-98d7-40c1-ae6d-21ee5655387f","Type":"ContainerStarted","Data":"5730431eaa30fbe444dd66dd65c1515c7b37ec480d2ff93c6c6ff15776393238"} Feb 27 07:51:04 crc kubenswrapper[4612]: I0227 07:51:04.695540 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-spnfb" Feb 27 07:51:04 crc kubenswrapper[4612]: I0227 07:51:04.702238 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 07:51:04 crc kubenswrapper[4612]: E0227 07:51:04.703920 4612 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 07:51:05.203905812 +0000 UTC m=+123.057835810 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 07:51:04 crc kubenswrapper[4612]: I0227 07:51:04.724138 4612 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-spnfb container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.15:8443/healthz\": read tcp 10.217.0.2:46904->10.217.0.15:8443: read: connection reset by peer" start-of-body= Feb 27 07:51:04 crc kubenswrapper[4612]: I0227 07:51:04.724176 4612 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-spnfb" podUID="42b49d65-d823-45af-b274-380dfbddc9b7" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.15:8443/healthz\": read tcp 10.217.0.2:46904->10.217.0.15:8443: read: connection reset by peer" Feb 27 07:51:04 crc kubenswrapper[4612]: I0227 07:51:04.724232 4612 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-sxt8c container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.16:8443/healthz\": dial tcp 10.217.0.16:8443: connect: connection refused" start-of-body= Feb 27 07:51:04 crc kubenswrapper[4612]: I0227 07:51:04.724244 4612 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-sxt8c" podUID="738d2b6e-2b7c-461d-b51e-f35eb5d41363" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.16:8443/healthz\": dial tcp 10.217.0.16:8443: connect: connection refused" Feb 27 07:51:04 crc kubenswrapper[4612]: I0227 07:51:04.745455 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-hbsll" podStartSLOduration=6.745439757 podStartE2EDuration="6.745439757s" podCreationTimestamp="2026-02-27 07:50:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 07:51:04.74339506 +0000 UTC m=+122.597325058" watchObservedRunningTime="2026-02-27 07:51:04.745439757 +0000 UTC m=+122.599369755" Feb 27 07:51:04 crc kubenswrapper[4612]: I0227 07:51:04.780020 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-sxt8c" podStartSLOduration=61.780001449 podStartE2EDuration="1m1.780001449s" podCreationTimestamp="2026-02-27 07:50:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 07:51:04.777276624 +0000 UTC m=+122.631206632" watchObservedRunningTime="2026-02-27 07:51:04.780001449 +0000 UTC m=+122.633931457" Feb 27 07:51:04 crc kubenswrapper[4612]: I0227 07:51:04.807438 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pdjlb\" (UID: \"db4d6d91-61f5-4beb-809c-63ae9028a911\") " pod="openshift-image-registry/image-registry-697d97f7c8-pdjlb" Feb 27 07:51:04 crc kubenswrapper[4612]: E0227 07:51:04.818176 4612 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 07:51:05.31814788 +0000 UTC m=+123.172077878 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pdjlb" (UID: "db4d6d91-61f5-4beb-809c-63ae9028a911") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 07:51:04 crc kubenswrapper[4612]: I0227 07:51:04.919896 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 07:51:04 crc kubenswrapper[4612]: E0227 07:51:04.920667 4612 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 07:51:05.420647865 +0000 UTC m=+123.274577863 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 07:51:05 crc kubenswrapper[4612]: I0227 07:51:05.023434 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pdjlb\" (UID: \"db4d6d91-61f5-4beb-809c-63ae9028a911\") " pod="openshift-image-registry/image-registry-697d97f7c8-pdjlb" Feb 27 07:51:05 crc kubenswrapper[4612]: E0227 07:51:05.023748 4612 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 07:51:05.523736506 +0000 UTC m=+123.377666504 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pdjlb" (UID: "db4d6d91-61f5-4beb-809c-63ae9028a911") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 07:51:05 crc kubenswrapper[4612]: I0227 07:51:05.126130 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 07:51:05 crc kubenswrapper[4612]: E0227 07:51:05.126398 4612 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 07:51:05.626383884 +0000 UTC m=+123.480313882 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 07:51:05 crc kubenswrapper[4612]: I0227 07:51:05.196115 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-nkbg2" Feb 27 07:51:05 crc kubenswrapper[4612]: I0227 07:51:05.228427 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pdjlb\" (UID: \"db4d6d91-61f5-4beb-809c-63ae9028a911\") " pod="openshift-image-registry/image-registry-697d97f7c8-pdjlb" Feb 27 07:51:05 crc kubenswrapper[4612]: E0227 07:51:05.228977 4612 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 07:51:05.728961081 +0000 UTC m=+123.582891079 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pdjlb" (UID: "db4d6d91-61f5-4beb-809c-63ae9028a911") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 07:51:05 crc kubenswrapper[4612]: I0227 07:51:05.346419 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 07:51:05 crc kubenswrapper[4612]: E0227 07:51:05.346730 4612 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 07:51:05.846712066 +0000 UTC m=+123.700642064 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 07:51:05 crc kubenswrapper[4612]: I0227 07:51:05.447575 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pdjlb\" (UID: \"db4d6d91-61f5-4beb-809c-63ae9028a911\") " pod="openshift-image-registry/image-registry-697d97f7c8-pdjlb" Feb 27 07:51:05 crc kubenswrapper[4612]: E0227 07:51:05.448014 4612 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 07:51:05.947998967 +0000 UTC m=+123.801928965 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pdjlb" (UID: "db4d6d91-61f5-4beb-809c-63ae9028a911") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 07:51:05 crc kubenswrapper[4612]: I0227 07:51:05.457816 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-spnfb" Feb 27 07:51:05 crc kubenswrapper[4612]: I0227 07:51:05.537602 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-65bdd5fb46-f8cwg"] Feb 27 07:51:05 crc kubenswrapper[4612]: E0227 07:51:05.537998 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="42b49d65-d823-45af-b274-380dfbddc9b7" containerName="route-controller-manager" Feb 27 07:51:05 crc kubenswrapper[4612]: I0227 07:51:05.538068 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="42b49d65-d823-45af-b274-380dfbddc9b7" containerName="route-controller-manager" Feb 27 07:51:05 crc kubenswrapper[4612]: I0227 07:51:05.538202 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="42b49d65-d823-45af-b274-380dfbddc9b7" containerName="route-controller-manager" Feb 27 07:51:05 crc kubenswrapper[4612]: I0227 07:51:05.538589 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-65bdd5fb46-f8cwg" Feb 27 07:51:05 crc kubenswrapper[4612]: I0227 07:51:05.550391 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 07:51:05 crc kubenswrapper[4612]: I0227 07:51:05.550489 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/42b49d65-d823-45af-b274-380dfbddc9b7-client-ca\") pod \"42b49d65-d823-45af-b274-380dfbddc9b7\" (UID: \"42b49d65-d823-45af-b274-380dfbddc9b7\") " Feb 27 07:51:05 crc kubenswrapper[4612]: I0227 07:51:05.550522 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f566h\" (UniqueName: \"kubernetes.io/projected/42b49d65-d823-45af-b274-380dfbddc9b7-kube-api-access-f566h\") pod \"42b49d65-d823-45af-b274-380dfbddc9b7\" (UID: \"42b49d65-d823-45af-b274-380dfbddc9b7\") " Feb 27 07:51:05 crc kubenswrapper[4612]: I0227 07:51:05.550540 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/42b49d65-d823-45af-b274-380dfbddc9b7-config\") pod \"42b49d65-d823-45af-b274-380dfbddc9b7\" (UID: \"42b49d65-d823-45af-b274-380dfbddc9b7\") " Feb 27 07:51:05 crc kubenswrapper[4612]: I0227 07:51:05.550618 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/42b49d65-d823-45af-b274-380dfbddc9b7-serving-cert\") pod \"42b49d65-d823-45af-b274-380dfbddc9b7\" (UID: \"42b49d65-d823-45af-b274-380dfbddc9b7\") " Feb 27 07:51:05 crc kubenswrapper[4612]: I0227 07:51:05.550819 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8da140b1-a24b-4c2a-bc20-dbf244d0283e-client-ca\") pod \"route-controller-manager-65bdd5fb46-f8cwg\" (UID: \"8da140b1-a24b-4c2a-bc20-dbf244d0283e\") " pod="openshift-route-controller-manager/route-controller-manager-65bdd5fb46-f8cwg" Feb 27 07:51:05 crc kubenswrapper[4612]: I0227 07:51:05.550863 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8da140b1-a24b-4c2a-bc20-dbf244d0283e-serving-cert\") pod \"route-controller-manager-65bdd5fb46-f8cwg\" (UID: \"8da140b1-a24b-4c2a-bc20-dbf244d0283e\") " pod="openshift-route-controller-manager/route-controller-manager-65bdd5fb46-f8cwg" Feb 27 07:51:05 crc kubenswrapper[4612]: I0227 07:51:05.550881 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8da140b1-a24b-4c2a-bc20-dbf244d0283e-config\") pod \"route-controller-manager-65bdd5fb46-f8cwg\" (UID: \"8da140b1-a24b-4c2a-bc20-dbf244d0283e\") " pod="openshift-route-controller-manager/route-controller-manager-65bdd5fb46-f8cwg" Feb 27 07:51:05 crc kubenswrapper[4612]: I0227 07:51:05.550900 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gmqw9\" (UniqueName: \"kubernetes.io/projected/8da140b1-a24b-4c2a-bc20-dbf244d0283e-kube-api-access-gmqw9\") pod \"route-controller-manager-65bdd5fb46-f8cwg\" (UID: \"8da140b1-a24b-4c2a-bc20-dbf244d0283e\") " pod="openshift-route-controller-manager/route-controller-manager-65bdd5fb46-f8cwg" Feb 27 07:51:05 crc kubenswrapper[4612]: E0227 07:51:05.551009 4612 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 07:51:06.050996265 +0000 UTC m=+123.904926263 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 07:51:05 crc kubenswrapper[4612]: I0227 07:51:05.551642 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/42b49d65-d823-45af-b274-380dfbddc9b7-client-ca" (OuterVolumeSpecName: "client-ca") pod "42b49d65-d823-45af-b274-380dfbddc9b7" (UID: "42b49d65-d823-45af-b274-380dfbddc9b7"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 07:51:05 crc kubenswrapper[4612]: I0227 07:51:05.553574 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/42b49d65-d823-45af-b274-380dfbddc9b7-config" (OuterVolumeSpecName: "config") pod "42b49d65-d823-45af-b274-380dfbddc9b7" (UID: "42b49d65-d823-45af-b274-380dfbddc9b7"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 07:51:05 crc kubenswrapper[4612]: I0227 07:51:05.566840 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-65bdd5fb46-f8cwg"] Feb 27 07:51:05 crc kubenswrapper[4612]: I0227 07:51:05.636364 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/42b49d65-d823-45af-b274-380dfbddc9b7-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "42b49d65-d823-45af-b274-380dfbddc9b7" (UID: "42b49d65-d823-45af-b274-380dfbddc9b7"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 07:51:05 crc kubenswrapper[4612]: I0227 07:51:05.647917 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/42b49d65-d823-45af-b274-380dfbddc9b7-kube-api-access-f566h" (OuterVolumeSpecName: "kube-api-access-f566h") pod "42b49d65-d823-45af-b274-380dfbddc9b7" (UID: "42b49d65-d823-45af-b274-380dfbddc9b7"). InnerVolumeSpecName "kube-api-access-f566h". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 07:51:05 crc kubenswrapper[4612]: I0227 07:51:05.656884 4612 patch_prober.go:28] interesting pod/router-default-5444994796-2d7kh container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 27 07:51:05 crc kubenswrapper[4612]: [-]has-synced failed: reason withheld Feb 27 07:51:05 crc kubenswrapper[4612]: [+]process-running ok Feb 27 07:51:05 crc kubenswrapper[4612]: healthz check failed Feb 27 07:51:05 crc kubenswrapper[4612]: I0227 07:51:05.656941 4612 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-2d7kh" podUID="faae1773-bf96-4127-8060-a40da07930ce" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 27 07:51:05 crc kubenswrapper[4612]: I0227 07:51:05.659355 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pdjlb\" (UID: \"db4d6d91-61f5-4beb-809c-63ae9028a911\") " pod="openshift-image-registry/image-registry-697d97f7c8-pdjlb" Feb 27 07:51:05 crc kubenswrapper[4612]: I0227 07:51:05.659421 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8da140b1-a24b-4c2a-bc20-dbf244d0283e-client-ca\") pod \"route-controller-manager-65bdd5fb46-f8cwg\" (UID: \"8da140b1-a24b-4c2a-bc20-dbf244d0283e\") " pod="openshift-route-controller-manager/route-controller-manager-65bdd5fb46-f8cwg" Feb 27 07:51:05 crc kubenswrapper[4612]: I0227 07:51:05.659462 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8da140b1-a24b-4c2a-bc20-dbf244d0283e-serving-cert\") pod \"route-controller-manager-65bdd5fb46-f8cwg\" (UID: \"8da140b1-a24b-4c2a-bc20-dbf244d0283e\") " pod="openshift-route-controller-manager/route-controller-manager-65bdd5fb46-f8cwg" Feb 27 07:51:05 crc kubenswrapper[4612]: I0227 07:51:05.659480 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8da140b1-a24b-4c2a-bc20-dbf244d0283e-config\") pod \"route-controller-manager-65bdd5fb46-f8cwg\" (UID: \"8da140b1-a24b-4c2a-bc20-dbf244d0283e\") " pod="openshift-route-controller-manager/route-controller-manager-65bdd5fb46-f8cwg" Feb 27 07:51:05 crc kubenswrapper[4612]: I0227 07:51:05.659515 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gmqw9\" (UniqueName: \"kubernetes.io/projected/8da140b1-a24b-4c2a-bc20-dbf244d0283e-kube-api-access-gmqw9\") pod \"route-controller-manager-65bdd5fb46-f8cwg\" (UID: \"8da140b1-a24b-4c2a-bc20-dbf244d0283e\") " pod="openshift-route-controller-manager/route-controller-manager-65bdd5fb46-f8cwg" Feb 27 07:51:05 crc kubenswrapper[4612]: I0227 07:51:05.659563 4612 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/42b49d65-d823-45af-b274-380dfbddc9b7-client-ca\") on node \"crc\" DevicePath \"\"" Feb 27 07:51:05 crc kubenswrapper[4612]: I0227 07:51:05.661732 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8da140b1-a24b-4c2a-bc20-dbf244d0283e-client-ca\") pod \"route-controller-manager-65bdd5fb46-f8cwg\" (UID: \"8da140b1-a24b-4c2a-bc20-dbf244d0283e\") " pod="openshift-route-controller-manager/route-controller-manager-65bdd5fb46-f8cwg" Feb 27 07:51:05 crc kubenswrapper[4612]: E0227 07:51:05.662027 4612 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 07:51:06.162013214 +0000 UTC m=+124.015943212 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pdjlb" (UID: "db4d6d91-61f5-4beb-809c-63ae9028a911") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 07:51:05 crc kubenswrapper[4612]: I0227 07:51:05.659578 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f566h\" (UniqueName: \"kubernetes.io/projected/42b49d65-d823-45af-b274-380dfbddc9b7-kube-api-access-f566h\") on node \"crc\" DevicePath \"\"" Feb 27 07:51:05 crc kubenswrapper[4612]: I0227 07:51:05.666299 4612 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/42b49d65-d823-45af-b274-380dfbddc9b7-config\") on node \"crc\" DevicePath \"\"" Feb 27 07:51:05 crc kubenswrapper[4612]: I0227 07:51:05.666309 4612 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/42b49d65-d823-45af-b274-380dfbddc9b7-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 27 07:51:05 crc kubenswrapper[4612]: I0227 07:51:05.667370 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8da140b1-a24b-4c2a-bc20-dbf244d0283e-config\") pod \"route-controller-manager-65bdd5fb46-f8cwg\" (UID: \"8da140b1-a24b-4c2a-bc20-dbf244d0283e\") " pod="openshift-route-controller-manager/route-controller-manager-65bdd5fb46-f8cwg" Feb 27 07:51:05 crc kubenswrapper[4612]: I0227 07:51:05.707490 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8da140b1-a24b-4c2a-bc20-dbf244d0283e-serving-cert\") pod \"route-controller-manager-65bdd5fb46-f8cwg\" (UID: \"8da140b1-a24b-4c2a-bc20-dbf244d0283e\") " pod="openshift-route-controller-manager/route-controller-manager-65bdd5fb46-f8cwg" Feb 27 07:51:05 crc kubenswrapper[4612]: I0227 07:51:05.708336 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gmqw9\" (UniqueName: \"kubernetes.io/projected/8da140b1-a24b-4c2a-bc20-dbf244d0283e-kube-api-access-gmqw9\") pod \"route-controller-manager-65bdd5fb46-f8cwg\" (UID: \"8da140b1-a24b-4c2a-bc20-dbf244d0283e\") " pod="openshift-route-controller-manager/route-controller-manager-65bdd5fb46-f8cwg" Feb 27 07:51:05 crc kubenswrapper[4612]: I0227 07:51:05.771292 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 07:51:05 crc kubenswrapper[4612]: E0227 07:51:05.771376 4612 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 07:51:06.271359698 +0000 UTC m=+124.125289696 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 07:51:05 crc kubenswrapper[4612]: I0227 07:51:05.771422 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pdjlb\" (UID: \"db4d6d91-61f5-4beb-809c-63ae9028a911\") " pod="openshift-image-registry/image-registry-697d97f7c8-pdjlb" Feb 27 07:51:05 crc kubenswrapper[4612]: E0227 07:51:05.772200 4612 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 07:51:06.27217595 +0000 UTC m=+124.126106028 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pdjlb" (UID: "db4d6d91-61f5-4beb-809c-63ae9028a911") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 07:51:05 crc kubenswrapper[4612]: I0227 07:51:05.799972 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-kbtxf" event={"ID":"98e7f73e-ab59-4e4d-af52-4794f9e28abb","Type":"ContainerStarted","Data":"94b06713348541975a588cb25b64c5e05323af0242349b45260718d53ac6f081"} Feb 27 07:51:05 crc kubenswrapper[4612]: I0227 07:51:05.803172 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-bqh6n" event={"ID":"9421f529-9b1e-4bd0-8032-7baa608981b4","Type":"ContainerStarted","Data":"c01519a8b2d4b37110f2966d90389d8ece791b2db3b0def6040c6291f38a500c"} Feb 27 07:51:05 crc kubenswrapper[4612]: I0227 07:51:05.804823 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-42tqr" event={"ID":"a61e7e5a-48f6-4b26-9b1e-1858222ff35f","Type":"ContainerStarted","Data":"6f0a2d76e114bb11931e3fa7e9273674832d52ad8d9fac066509a0cc4825afd6"} Feb 27 07:51:05 crc kubenswrapper[4612]: I0227 07:51:05.817286 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-kbtxf" podStartSLOduration=62.817267423 podStartE2EDuration="1m2.817267423s" podCreationTimestamp="2026-02-27 07:50:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 07:51:05.816583124 +0000 UTC m=+123.670513112" watchObservedRunningTime="2026-02-27 07:51:05.817267423 +0000 UTC m=+123.671197421" Feb 27 07:51:05 crc kubenswrapper[4612]: I0227 07:51:05.817861 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-kv4nc" event={"ID":"fa91e393-fed8-4f4e-8dbe-7bc51ed35cea","Type":"ContainerStarted","Data":"99e0a038a7c3eb6195b603076de139003b18eb5b10b0bf1a4d381ad3faed9d65"} Feb 27 07:51:05 crc kubenswrapper[4612]: I0227 07:51:05.823606 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-hbsll" event={"ID":"3dbda067-50b0-4d11-bb83-b7162d0cebde","Type":"ContainerStarted","Data":"463936a2c3c8c4a481ca4427640c4cb986a48a45c3b60ec46ebea6ba24890101"} Feb 27 07:51:05 crc kubenswrapper[4612]: I0227 07:51:05.825366 4612 generic.go:334] "Generic (PLEG): container finished" podID="25a0e1aa-86e2-4acf-885b-16c26efc337a" containerID="d5d2564a925ed0e38f2fb57832fd9046692be185fa14db8d71f5574f7b31c7fe" exitCode=0 Feb 27 07:51:05 crc kubenswrapper[4612]: I0227 07:51:05.825433 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-cw57r" event={"ID":"25a0e1aa-86e2-4acf-885b-16c26efc337a","Type":"ContainerDied","Data":"d5d2564a925ed0e38f2fb57832fd9046692be185fa14db8d71f5574f7b31c7fe"} Feb 27 07:51:05 crc kubenswrapper[4612]: I0227 07:51:05.827898 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29536305-vmtnx" event={"ID":"ea3d0df0-517f-4865-964d-3e12a313696b","Type":"ContainerStarted","Data":"9f4a63ddb5dc79f146a7585e5535f6c947de9edebadc177ca8b4362f7aa13188"} Feb 27 07:51:05 crc kubenswrapper[4612]: I0227 07:51:05.829091 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/cni-sysctl-allowlist-ds-v589k" event={"ID":"2522fcbd-fe8e-46cf-95ba-620316513df1","Type":"ContainerStarted","Data":"19989c4604cb1985b76a2c1d2f8276aa3c4c5fb242e7ebf2619f837c570f9715"} Feb 27 07:51:05 crc kubenswrapper[4612]: I0227 07:51:05.829260 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-multus/cni-sysctl-allowlist-ds-v589k" Feb 27 07:51:05 crc kubenswrapper[4612]: I0227 07:51:05.876949 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 07:51:05 crc kubenswrapper[4612]: E0227 07:51:05.877868 4612 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 07:51:06.377847002 +0000 UTC m=+124.231777000 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 07:51:05 crc kubenswrapper[4612]: I0227 07:51:05.896471 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6cz6v" event={"ID":"90d5c002-a228-4c36-b395-5ba0b01aac8e","Type":"ContainerStarted","Data":"743268e4ccef0dc1ddfa020fd97ba2fd3583df7c2054ed7b3a5221b987dabe4c"} Feb 27 07:51:05 crc kubenswrapper[4612]: I0227 07:51:05.899087 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6cz6v" Feb 27 07:51:05 crc kubenswrapper[4612]: I0227 07:51:05.899802 4612 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-6cz6v container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.26:8443/healthz\": dial tcp 10.217.0.26:8443: connect: connection refused" start-of-body= Feb 27 07:51:05 crc kubenswrapper[4612]: I0227 07:51:05.899845 4612 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6cz6v" podUID="90d5c002-a228-4c36-b395-5ba0b01aac8e" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.26:8443/healthz\": dial tcp 10.217.0.26:8443: connect: connection refused" Feb 27 07:51:05 crc kubenswrapper[4612]: I0227 07:51:05.902079 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-42tqr" podStartSLOduration=62.902062679 podStartE2EDuration="1m2.902062679s" podCreationTimestamp="2026-02-27 07:50:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 07:51:05.895632462 +0000 UTC m=+123.749562460" watchObservedRunningTime="2026-02-27 07:51:05.902062679 +0000 UTC m=+123.755992677" Feb 27 07:51:05 crc kubenswrapper[4612]: I0227 07:51:05.903205 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Feb 27 07:51:05 crc kubenswrapper[4612]: I0227 07:51:05.926227 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-65bdd5fb46-f8cwg" Feb 27 07:51:05 crc kubenswrapper[4612]: I0227 07:51:05.933537 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/cni-sysctl-allowlist-ds-v589k" podStartSLOduration=7.933524756 podStartE2EDuration="7.933524756s" podCreationTimestamp="2026-02-27 07:50:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 07:51:05.932903559 +0000 UTC m=+123.786833557" watchObservedRunningTime="2026-02-27 07:51:05.933524756 +0000 UTC m=+123.787454754" Feb 27 07:51:05 crc kubenswrapper[4612]: I0227 07:51:05.949270 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-4lpn2" event={"ID":"4138a6e8-dd4e-44d3-9103-8f328b794227","Type":"ContainerStarted","Data":"efc4c56fb2b8a007fbe87eed95a9207e899b065d8c9f7de07bab15e783f7387e"} Feb 27 07:51:05 crc kubenswrapper[4612]: I0227 07:51:05.949404 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-multus/cni-sysctl-allowlist-ds-v589k" Feb 27 07:51:05 crc kubenswrapper[4612]: I0227 07:51:05.962548 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-kgkd4" event={"ID":"b979d24b-0d0c-43c1-b50c-52e6b3801daf","Type":"ContainerStarted","Data":"c3f748ae8b82e6f322c9cbe371c4ad5da65ddfe2d0f10cb2d4b31fa3c988b6f2"} Feb 27 07:51:05 crc kubenswrapper[4612]: I0227 07:51:05.965234 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-kv4nc" podStartSLOduration=7.96521641 podStartE2EDuration="7.96521641s" podCreationTimestamp="2026-02-27 07:50:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 07:51:05.962742332 +0000 UTC m=+123.816672330" watchObservedRunningTime="2026-02-27 07:51:05.96521641 +0000 UTC m=+123.819146408" Feb 27 07:51:05 crc kubenswrapper[4612]: I0227 07:51:05.978583 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pdjlb\" (UID: \"db4d6d91-61f5-4beb-809c-63ae9028a911\") " pod="openshift-image-registry/image-registry-697d97f7c8-pdjlb" Feb 27 07:51:05 crc kubenswrapper[4612]: E0227 07:51:05.979042 4612 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 07:51:06.47903128 +0000 UTC m=+124.332961278 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pdjlb" (UID: "db4d6d91-61f5-4beb-809c-63ae9028a911") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 07:51:06 crc kubenswrapper[4612]: I0227 07:51:06.029543 4612 ???:1] "http: TLS handshake error from 192.168.126.11:37986: no serving certificate available for the kubelet" Feb 27 07:51:06 crc kubenswrapper[4612]: I0227 07:51:06.031868 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29536305-vmtnx" podStartSLOduration=63.031853326 podStartE2EDuration="1m3.031853326s" podCreationTimestamp="2026-02-27 07:50:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 07:51:06.031238029 +0000 UTC m=+123.885168027" watchObservedRunningTime="2026-02-27 07:51:06.031853326 +0000 UTC m=+123.885783324" Feb 27 07:51:06 crc kubenswrapper[4612]: I0227 07:51:06.077289 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-kgkd4" podStartSLOduration=63.077271608 podStartE2EDuration="1m3.077271608s" podCreationTimestamp="2026-02-27 07:50:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 07:51:06.075985222 +0000 UTC m=+123.929915220" watchObservedRunningTime="2026-02-27 07:51:06.077271608 +0000 UTC m=+123.931201616" Feb 27 07:51:06 crc kubenswrapper[4612]: I0227 07:51:06.081148 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 07:51:06 crc kubenswrapper[4612]: E0227 07:51:06.082175 4612 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 07:51:06.582160412 +0000 UTC m=+124.436090410 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 07:51:06 crc kubenswrapper[4612]: I0227 07:51:06.112204 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-6q4q2" event={"ID":"e24e170c-7ee2-4d40-badf-21bc59723d1c","Type":"ContainerStarted","Data":"92297545df44383340b8df50fed4609b1792798e5c21b88ecaab3b414ef678e9"} Feb 27 07:51:06 crc kubenswrapper[4612]: I0227 07:51:06.124085 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6cz6v" podStartSLOduration=63.124069916 podStartE2EDuration="1m3.124069916s" podCreationTimestamp="2026-02-27 07:50:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 07:51:06.119439499 +0000 UTC m=+123.973369497" watchObservedRunningTime="2026-02-27 07:51:06.124069916 +0000 UTC m=+123.977999914" Feb 27 07:51:06 crc kubenswrapper[4612]: I0227 07:51:06.129332 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-ghtdk" event={"ID":"f0d8e4b4-aec4-433c-adff-091c8944e8a3","Type":"ContainerStarted","Data":"6fd3b0d5ef4fc3f057a6c3853ccbecd68d7fc68dda7ef865da9a2f00acf3d7eb"} Feb 27 07:51:06 crc kubenswrapper[4612]: I0227 07:51:06.143767 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-sxt8c" event={"ID":"738d2b6e-2b7c-461d-b51e-f35eb5d41363","Type":"ContainerStarted","Data":"80058f3d8b6825a6aa02bba10ea42f4f51630b27d2fc39c813f4cc5bb5412319"} Feb 27 07:51:06 crc kubenswrapper[4612]: I0227 07:51:06.150207 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-tcqv8" event={"ID":"efc3c589-c379-45ca-a974-75483d90c5a4","Type":"ContainerStarted","Data":"c7b5e385b9ce70d53d2563f284936ae28b230ba0a5fa2da1a202981307602b38"} Feb 27 07:51:06 crc kubenswrapper[4612]: I0227 07:51:06.154169 4612 generic.go:334] "Generic (PLEG): container finished" podID="42b49d65-d823-45af-b274-380dfbddc9b7" containerID="1e8c7d56859a3f57ec388590f4f8362847e655b83c149dcef9376df449351658" exitCode=0 Feb 27 07:51:06 crc kubenswrapper[4612]: I0227 07:51:06.154266 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-spnfb" event={"ID":"42b49d65-d823-45af-b274-380dfbddc9b7","Type":"ContainerDied","Data":"1e8c7d56859a3f57ec388590f4f8362847e655b83c149dcef9376df449351658"} Feb 27 07:51:06 crc kubenswrapper[4612]: I0227 07:51:06.154294 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-spnfb" event={"ID":"42b49d65-d823-45af-b274-380dfbddc9b7","Type":"ContainerDied","Data":"5802342342fd2e28bb97f45d7fe121c64b9fb737dd9987195385f7c56553f64b"} Feb 27 07:51:06 crc kubenswrapper[4612]: I0227 07:51:06.154312 4612 scope.go:117] "RemoveContainer" containerID="1e8c7d56859a3f57ec388590f4f8362847e655b83c149dcef9376df449351658" Feb 27 07:51:06 crc kubenswrapper[4612]: I0227 07:51:06.154429 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-spnfb" Feb 27 07:51:06 crc kubenswrapper[4612]: I0227 07:51:06.154990 4612 ???:1] "http: TLS handshake error from 192.168.126.11:38002: no serving certificate available for the kubelet" Feb 27 07:51:06 crc kubenswrapper[4612]: I0227 07:51:06.167056 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-8l852" event={"ID":"cc5b15e4-6c84-4474-bf4d-f652d2e8000c","Type":"ContainerStarted","Data":"7b5f8c26e7648096eece09ec8510d3c264f23dbc699e2b8b19b6fd24409a3f58"} Feb 27 07:51:06 crc kubenswrapper[4612]: I0227 07:51:06.182162 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pdjlb\" (UID: \"db4d6d91-61f5-4beb-809c-63ae9028a911\") " pod="openshift-image-registry/image-registry-697d97f7c8-pdjlb" Feb 27 07:51:06 crc kubenswrapper[4612]: E0227 07:51:06.182700 4612 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 07:51:06.682671721 +0000 UTC m=+124.536601719 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pdjlb" (UID: "db4d6d91-61f5-4beb-809c-63ae9028a911") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 07:51:06 crc kubenswrapper[4612]: I0227 07:51:06.188835 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-gcfmk" event={"ID":"c28eca4e-4d90-4bf0-95bf-880f3a9037e3","Type":"ContainerStarted","Data":"ebdd467139167b52a6ae0224bed3942acf7c096020de538b26ec145a47946433"} Feb 27 07:51:06 crc kubenswrapper[4612]: I0227 07:51:06.202738 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-96vl6" event={"ID":"620a7a2f-92fc-4a76-900b-021d0dd78d79","Type":"ContainerStarted","Data":"3cfa93ffeebff5827f2fd5205708025e4fadcd176df02ed556733b4b4b88ef6b"} Feb 27 07:51:06 crc kubenswrapper[4612]: I0227 07:51:06.222603 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bk49v" event={"ID":"fb198883-74b8-43a2-a2db-390e951fda9a","Type":"ContainerStarted","Data":"a63cecd0890b8e08b9018144e91605c88f0a472d92b4a29a1ea66429c4a76e16"} Feb 27 07:51:06 crc kubenswrapper[4612]: I0227 07:51:06.225917 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-8l852" podStartSLOduration=63.225901502 podStartE2EDuration="1m3.225901502s" podCreationTimestamp="2026-02-27 07:50:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 07:51:06.22363343 +0000 UTC m=+124.077563428" watchObservedRunningTime="2026-02-27 07:51:06.225901502 +0000 UTC m=+124.079831500" Feb 27 07:51:06 crc kubenswrapper[4612]: I0227 07:51:06.233935 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=1.233913453 podStartE2EDuration="1.233913453s" podCreationTimestamp="2026-02-27 07:51:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 07:51:06.180866241 +0000 UTC m=+124.034796259" watchObservedRunningTime="2026-02-27 07:51:06.233913453 +0000 UTC m=+124.087843461" Feb 27 07:51:06 crc kubenswrapper[4612]: I0227 07:51:06.239523 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bk49v" Feb 27 07:51:06 crc kubenswrapper[4612]: I0227 07:51:06.273336 4612 scope.go:117] "RemoveContainer" containerID="1e8c7d56859a3f57ec388590f4f8362847e655b83c149dcef9376df449351658" Feb 27 07:51:06 crc kubenswrapper[4612]: I0227 07:51:06.275152 4612 ???:1] "http: TLS handshake error from 192.168.126.11:38008: no serving certificate available for the kubelet" Feb 27 07:51:06 crc kubenswrapper[4612]: E0227 07:51:06.277095 4612 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1e8c7d56859a3f57ec388590f4f8362847e655b83c149dcef9376df449351658\": container with ID starting with 1e8c7d56859a3f57ec388590f4f8362847e655b83c149dcef9376df449351658 not found: ID does not exist" containerID="1e8c7d56859a3f57ec388590f4f8362847e655b83c149dcef9376df449351658" Feb 27 07:51:06 crc kubenswrapper[4612]: I0227 07:51:06.277139 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1e8c7d56859a3f57ec388590f4f8362847e655b83c149dcef9376df449351658"} err="failed to get container status \"1e8c7d56859a3f57ec388590f4f8362847e655b83c149dcef9376df449351658\": rpc error: code = NotFound desc = could not find container \"1e8c7d56859a3f57ec388590f4f8362847e655b83c149dcef9376df449351658\": container with ID starting with 1e8c7d56859a3f57ec388590f4f8362847e655b83c149dcef9376df449351658 not found: ID does not exist" Feb 27 07:51:06 crc kubenswrapper[4612]: I0227 07:51:06.283451 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 07:51:06 crc kubenswrapper[4612]: E0227 07:51:06.284710 4612 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 07:51:06.784678902 +0000 UTC m=+124.638608900 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 07:51:06 crc kubenswrapper[4612]: I0227 07:51:06.297377 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-6q4q2" podStartSLOduration=63.297357442 podStartE2EDuration="1m3.297357442s" podCreationTimestamp="2026-02-27 07:50:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 07:51:06.253898974 +0000 UTC m=+124.107828972" watchObservedRunningTime="2026-02-27 07:51:06.297357442 +0000 UTC m=+124.151287430" Feb 27 07:51:06 crc kubenswrapper[4612]: I0227 07:51:06.315497 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-spnfb"] Feb 27 07:51:06 crc kubenswrapper[4612]: I0227 07:51:06.316964 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-spnfb"] Feb 27 07:51:06 crc kubenswrapper[4612]: I0227 07:51:06.358730 4612 ???:1] "http: TLS handshake error from 192.168.126.11:38016: no serving certificate available for the kubelet" Feb 27 07:51:06 crc kubenswrapper[4612]: I0227 07:51:06.382831 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-qvwzc" Feb 27 07:51:06 crc kubenswrapper[4612]: I0227 07:51:06.385896 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pdjlb\" (UID: \"db4d6d91-61f5-4beb-809c-63ae9028a911\") " pod="openshift-image-registry/image-registry-697d97f7c8-pdjlb" Feb 27 07:51:06 crc kubenswrapper[4612]: E0227 07:51:06.389579 4612 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 07:51:06.889562982 +0000 UTC m=+124.743492980 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pdjlb" (UID: "db4d6d91-61f5-4beb-809c-63ae9028a911") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 07:51:06 crc kubenswrapper[4612]: I0227 07:51:06.491354 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 07:51:06 crc kubenswrapper[4612]: E0227 07:51:06.491736 4612 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 07:51:06.991714877 +0000 UTC m=+124.845644875 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 07:51:06 crc kubenswrapper[4612]: I0227 07:51:06.491900 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pdjlb\" (UID: \"db4d6d91-61f5-4beb-809c-63ae9028a911\") " pod="openshift-image-registry/image-registry-697d97f7c8-pdjlb" Feb 27 07:51:06 crc kubenswrapper[4612]: E0227 07:51:06.492307 4612 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 07:51:06.992281683 +0000 UTC m=+124.846211681 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pdjlb" (UID: "db4d6d91-61f5-4beb-809c-63ae9028a911") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 07:51:06 crc kubenswrapper[4612]: I0227 07:51:06.496081 4612 ???:1] "http: TLS handshake error from 192.168.126.11:38024: no serving certificate available for the kubelet" Feb 27 07:51:06 crc kubenswrapper[4612]: I0227 07:51:06.594512 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 07:51:06 crc kubenswrapper[4612]: E0227 07:51:06.594917 4612 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 07:51:07.094903761 +0000 UTC m=+124.948833759 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 07:51:06 crc kubenswrapper[4612]: I0227 07:51:06.642854 4612 patch_prober.go:28] interesting pod/router-default-5444994796-2d7kh container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 27 07:51:06 crc kubenswrapper[4612]: [-]has-synced failed: reason withheld Feb 27 07:51:06 crc kubenswrapper[4612]: [+]process-running ok Feb 27 07:51:06 crc kubenswrapper[4612]: healthz check failed Feb 27 07:51:06 crc kubenswrapper[4612]: I0227 07:51:06.643219 4612 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-2d7kh" podUID="faae1773-bf96-4127-8060-a40da07930ce" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 27 07:51:06 crc kubenswrapper[4612]: I0227 07:51:06.651239 4612 ???:1] "http: TLS handshake error from 192.168.126.11:38030: no serving certificate available for the kubelet" Feb 27 07:51:06 crc kubenswrapper[4612]: I0227 07:51:06.695634 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pdjlb\" (UID: \"db4d6d91-61f5-4beb-809c-63ae9028a911\") " pod="openshift-image-registry/image-registry-697d97f7c8-pdjlb" Feb 27 07:51:06 crc kubenswrapper[4612]: E0227 07:51:06.696063 4612 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 07:51:07.196050758 +0000 UTC m=+125.049980756 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pdjlb" (UID: "db4d6d91-61f5-4beb-809c-63ae9028a911") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 07:51:06 crc kubenswrapper[4612]: I0227 07:51:06.760107 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-65bdd5fb46-f8cwg"] Feb 27 07:51:06 crc kubenswrapper[4612]: I0227 07:51:06.796615 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 07:51:06 crc kubenswrapper[4612]: E0227 07:51:06.797636 4612 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 07:51:07.297621287 +0000 UTC m=+125.151551285 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 07:51:06 crc kubenswrapper[4612]: W0227 07:51:06.797960 4612 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8da140b1_a24b_4c2a_bc20_dbf244d0283e.slice/crio-a9feef3c425f375dd5a74438f987b5f6944eb30246a99c1477e7b1b83c3cbf82 WatchSource:0}: Error finding container a9feef3c425f375dd5a74438f987b5f6944eb30246a99c1477e7b1b83c3cbf82: Status 404 returned error can't find the container with id a9feef3c425f375dd5a74438f987b5f6944eb30246a99c1477e7b1b83c3cbf82 Feb 27 07:51:06 crc kubenswrapper[4612]: I0227 07:51:06.881074 4612 ???:1] "http: TLS handshake error from 192.168.126.11:38032: no serving certificate available for the kubelet" Feb 27 07:51:06 crc kubenswrapper[4612]: I0227 07:51:06.899303 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pdjlb\" (UID: \"db4d6d91-61f5-4beb-809c-63ae9028a911\") " pod="openshift-image-registry/image-registry-697d97f7c8-pdjlb" Feb 27 07:51:06 crc kubenswrapper[4612]: I0227 07:51:06.902929 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="42b49d65-d823-45af-b274-380dfbddc9b7" path="/var/lib/kubelet/pods/42b49d65-d823-45af-b274-380dfbddc9b7/volumes" Feb 27 07:51:06 crc kubenswrapper[4612]: E0227 07:51:06.903257 4612 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 07:51:07.403243028 +0000 UTC m=+125.257173026 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pdjlb" (UID: "db4d6d91-61f5-4beb-809c-63ae9028a911") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 07:51:06 crc kubenswrapper[4612]: I0227 07:51:06.904917 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-multus/cni-sysctl-allowlist-ds-v589k"] Feb 27 07:51:07 crc kubenswrapper[4612]: I0227 07:51:07.000810 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 07:51:07 crc kubenswrapper[4612]: E0227 07:51:07.000944 4612 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 07:51:07.50092814 +0000 UTC m=+125.354858138 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 07:51:07 crc kubenswrapper[4612]: I0227 07:51:07.000999 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pdjlb\" (UID: \"db4d6d91-61f5-4beb-809c-63ae9028a911\") " pod="openshift-image-registry/image-registry-697d97f7c8-pdjlb" Feb 27 07:51:07 crc kubenswrapper[4612]: E0227 07:51:07.001266 4612 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 07:51:07.501258569 +0000 UTC m=+125.355188567 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pdjlb" (UID: "db4d6d91-61f5-4beb-809c-63ae9028a911") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 07:51:07 crc kubenswrapper[4612]: I0227 07:51:07.102625 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 07:51:07 crc kubenswrapper[4612]: E0227 07:51:07.104858 4612 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 07:51:07.604842543 +0000 UTC m=+125.458772541 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 07:51:07 crc kubenswrapper[4612]: I0227 07:51:07.198734 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-c6sdx" Feb 27 07:51:07 crc kubenswrapper[4612]: I0227 07:51:07.203618 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pdjlb\" (UID: \"db4d6d91-61f5-4beb-809c-63ae9028a911\") " pod="openshift-image-registry/image-registry-697d97f7c8-pdjlb" Feb 27 07:51:07 crc kubenswrapper[4612]: E0227 07:51:07.204182 4612 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 07:51:07.70416868 +0000 UTC m=+125.558098678 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pdjlb" (UID: "db4d6d91-61f5-4beb-809c-63ae9028a911") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 07:51:07 crc kubenswrapper[4612]: I0227 07:51:07.242445 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-65bdd5fb46-f8cwg" event={"ID":"8da140b1-a24b-4c2a-bc20-dbf244d0283e","Type":"ContainerStarted","Data":"123ad4605465bb9521f8ad344e05484fa6205b50d75bef0a9b3e23dd0cfc4bd0"} Feb 27 07:51:07 crc kubenswrapper[4612]: I0227 07:51:07.242496 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-65bdd5fb46-f8cwg" event={"ID":"8da140b1-a24b-4c2a-bc20-dbf244d0283e","Type":"ContainerStarted","Data":"a9feef3c425f375dd5a74438f987b5f6944eb30246a99c1477e7b1b83c3cbf82"} Feb 27 07:51:07 crc kubenswrapper[4612]: I0227 07:51:07.243812 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-65bdd5fb46-f8cwg" Feb 27 07:51:07 crc kubenswrapper[4612]: I0227 07:51:07.245297 4612 patch_prober.go:28] interesting pod/route-controller-manager-65bdd5fb46-f8cwg container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.44:8443/healthz\": dial tcp 10.217.0.44:8443: connect: connection refused" start-of-body= Feb 27 07:51:07 crc kubenswrapper[4612]: I0227 07:51:07.245326 4612 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-65bdd5fb46-f8cwg" podUID="8da140b1-a24b-4c2a-bc20-dbf244d0283e" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.44:8443/healthz\": dial tcp 10.217.0.44:8443: connect: connection refused" Feb 27 07:51:07 crc kubenswrapper[4612]: I0227 07:51:07.252073 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-cw57r" event={"ID":"25a0e1aa-86e2-4acf-885b-16c26efc337a","Type":"ContainerStarted","Data":"e1fbf15965a8f19e11e1df413b16054aa35bd11c6e95d079505049935dbf01bb"} Feb 27 07:51:07 crc kubenswrapper[4612]: I0227 07:51:07.252112 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-cw57r" event={"ID":"25a0e1aa-86e2-4acf-885b-16c26efc337a","Type":"ContainerStarted","Data":"6733cf88c65c1bde183c36648efdb1aabfa33f8f70fc864e1241faeba9a3a86e"} Feb 27 07:51:07 crc kubenswrapper[4612]: I0227 07:51:07.265953 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-bqh6n" event={"ID":"9421f529-9b1e-4bd0-8032-7baa608981b4","Type":"ContainerStarted","Data":"ebf5da7157fe353fda71a91e7545388f631bb23dcca241fa2cdf5f8768539bb8"} Feb 27 07:51:07 crc kubenswrapper[4612]: I0227 07:51:07.265994 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-bqh6n" event={"ID":"9421f529-9b1e-4bd0-8032-7baa608981b4","Type":"ContainerStarted","Data":"302ed63fa073dd97f305c4d8c2fa33d8b1524fde867957148239baebd36d7e25"} Feb 27 07:51:07 crc kubenswrapper[4612]: I0227 07:51:07.272521 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-z9zch" event={"ID":"9bb5eece-3596-485e-a859-b61d201b05f5","Type":"ContainerStarted","Data":"f3b8c3c155296560cd0f5ef2c5b03d731ba6767ef0ca4c1b9480c2dc8b944852"} Feb 27 07:51:07 crc kubenswrapper[4612]: I0227 07:51:07.273087 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-z9zch" Feb 27 07:51:07 crc kubenswrapper[4612]: I0227 07:51:07.278295 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-gcfmk" event={"ID":"c28eca4e-4d90-4bf0-95bf-880f3a9037e3","Type":"ContainerStarted","Data":"b6ea64488ce4e1825babde317a68abf977f64c392d4812f08aac5f542c21e194"} Feb 27 07:51:07 crc kubenswrapper[4612]: I0227 07:51:07.293428 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-96vl6" event={"ID":"620a7a2f-92fc-4a76-900b-021d0dd78d79","Type":"ContainerStarted","Data":"4bbf197d1bb8f1e5ee5f6859579cc4a1486b27c7df1db57f5f8136cd6f100398"} Feb 27 07:51:07 crc kubenswrapper[4612]: I0227 07:51:07.303016 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-ghtdk" event={"ID":"f0d8e4b4-aec4-433c-adff-091c8944e8a3","Type":"ContainerStarted","Data":"5d1a41aa54d0465b7df358d9a40f179d987cf306528710f9e6ef37ef5d70401e"} Feb 27 07:51:07 crc kubenswrapper[4612]: I0227 07:51:07.304123 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 07:51:07 crc kubenswrapper[4612]: E0227 07:51:07.304402 4612 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 07:51:07.804388752 +0000 UTC m=+125.658318750 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 07:51:07 crc kubenswrapper[4612]: I0227 07:51:07.309548 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-65bdd5fb46-f8cwg" podStartSLOduration=5.309534954 podStartE2EDuration="5.309534954s" podCreationTimestamp="2026-02-27 07:51:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 07:51:07.303762045 +0000 UTC m=+125.157692033" watchObservedRunningTime="2026-02-27 07:51:07.309534954 +0000 UTC m=+125.163464952" Feb 27 07:51:07 crc kubenswrapper[4612]: I0227 07:51:07.315818 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-5wvhc" event={"ID":"ff5fcfe1-eb01-4f01-857d-a67712db3b87","Type":"ContainerStarted","Data":"ef9bc1c2f314bd200839c0ae167e11f2499be795c2213508909507e0f44405f2"} Feb 27 07:51:07 crc kubenswrapper[4612]: I0227 07:51:07.322905 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-tcqv8" event={"ID":"efc3c589-c379-45ca-a974-75483d90c5a4","Type":"ContainerStarted","Data":"36c2388fc76756f9d7adf5bcd1ed2719981ed25baa8510c2a888a5c7faa2378e"} Feb 27 07:51:07 crc kubenswrapper[4612]: I0227 07:51:07.323490 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-tcqv8" Feb 27 07:51:07 crc kubenswrapper[4612]: I0227 07:51:07.325783 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-n5jbm" event={"ID":"ad4c7d8c-98d7-40c1-ae6d-21ee5655387f","Type":"ContainerStarted","Data":"d492b001024ee566814b31141a9e2874af6c5eaa112c34de92db6a4ecc157566"} Feb 27 07:51:07 crc kubenswrapper[4612]: I0227 07:51:07.329112 4612 ???:1] "http: TLS handshake error from 192.168.126.11:38036: no serving certificate available for the kubelet" Feb 27 07:51:07 crc kubenswrapper[4612]: I0227 07:51:07.331972 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-gvv4x" event={"ID":"c02d88e5-7055-4864-8b83-d69fdccad379","Type":"ContainerStarted","Data":"ce0059d0dfaf7fabf0b180f446848d755f56d3577e37c31dccc48a22203c068a"} Feb 27 07:51:07 crc kubenswrapper[4612]: I0227 07:51:07.332010 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-gvv4x" event={"ID":"c02d88e5-7055-4864-8b83-d69fdccad379","Type":"ContainerStarted","Data":"ddbfcfe25573635c4af7305609be02bb3008619e72b6cef5d2fac9667fcd7a35"} Feb 27 07:51:07 crc kubenswrapper[4612]: I0227 07:51:07.332532 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-gvv4x" Feb 27 07:51:07 crc kubenswrapper[4612]: I0227 07:51:07.336443 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zpshm" event={"ID":"d951f066-9e39-4e0a-9fa3-da9c868552d1","Type":"ContainerStarted","Data":"92a14b63068be97ae508237518c1318de0668820063fa29a67a1ba27b7a641a2"} Feb 27 07:51:07 crc kubenswrapper[4612]: I0227 07:51:07.348951 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-ks964" event={"ID":"0b1512ff-906a-4df2-b0ca-a09ba21f4875","Type":"ContainerStarted","Data":"9bbdfa67c83e28c1bf44abde465de57432282232423c716745fd5988dd9af4ee"} Feb 27 07:51:07 crc kubenswrapper[4612]: I0227 07:51:07.363463 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-q6h7h" event={"ID":"094b001c-e80c-4d07-a161-e00514667462","Type":"ContainerStarted","Data":"05f9b45704ef974c3c078d8d5db73da180ff9184c7d313261cc1bec6dba28a6b"} Feb 27 07:51:07 crc kubenswrapper[4612]: I0227 07:51:07.393993 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2rrfz" event={"ID":"d9aa8f66-da03-49e5-b7b4-d17135ccaa03","Type":"ContainerStarted","Data":"49353416b6041382617aa44dbf6c8c85b4b1203260aab1fdb9278e5e0b51f4cf"} Feb 27 07:51:07 crc kubenswrapper[4612]: I0227 07:51:07.405270 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pdjlb\" (UID: \"db4d6d91-61f5-4beb-809c-63ae9028a911\") " pod="openshift-image-registry/image-registry-697d97f7c8-pdjlb" Feb 27 07:51:07 crc kubenswrapper[4612]: E0227 07:51:07.406737 4612 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 07:51:07.906720662 +0000 UTC m=+125.760650740 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pdjlb" (UID: "db4d6d91-61f5-4beb-809c-63ae9028a911") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 07:51:07 crc kubenswrapper[4612]: I0227 07:51:07.408642 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-4lpn2" event={"ID":"4138a6e8-dd4e-44d3-9103-8f328b794227","Type":"ContainerStarted","Data":"462f5ee9128378d9d87afc4a7ab384c8465b489f6b625c6013618644cee59411"} Feb 27 07:51:07 crc kubenswrapper[4612]: I0227 07:51:07.412832 4612 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-6cz6v container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.26:8443/healthz\": dial tcp 10.217.0.26:8443: connect: connection refused" start-of-body= Feb 27 07:51:07 crc kubenswrapper[4612]: I0227 07:51:07.412882 4612 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6cz6v" podUID="90d5c002-a228-4c36-b395-5ba0b01aac8e" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.26:8443/healthz\": dial tcp 10.217.0.26:8443: connect: connection refused" Feb 27 07:51:07 crc kubenswrapper[4612]: I0227 07:51:07.420012 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-bqh6n" podStartSLOduration=64.419997128 podStartE2EDuration="1m4.419997128s" podCreationTimestamp="2026-02-27 07:50:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 07:51:07.413493689 +0000 UTC m=+125.267423697" watchObservedRunningTime="2026-02-27 07:51:07.419997128 +0000 UTC m=+125.273927126" Feb 27 07:51:07 crc kubenswrapper[4612]: I0227 07:51:07.421229 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-ghtdk" podStartSLOduration=64.421224312 podStartE2EDuration="1m4.421224312s" podCreationTimestamp="2026-02-27 07:50:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 07:51:07.347039227 +0000 UTC m=+125.200969225" watchObservedRunningTime="2026-02-27 07:51:07.421224312 +0000 UTC m=+125.275154310" Feb 27 07:51:07 crc kubenswrapper[4612]: I0227 07:51:07.463185 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-gcfmk" podStartSLOduration=64.463167728 podStartE2EDuration="1m4.463167728s" podCreationTimestamp="2026-02-27 07:50:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 07:51:07.461151402 +0000 UTC m=+125.315081410" watchObservedRunningTime="2026-02-27 07:51:07.463167728 +0000 UTC m=+125.317097726" Feb 27 07:51:07 crc kubenswrapper[4612]: I0227 07:51:07.507246 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 07:51:07 crc kubenswrapper[4612]: E0227 07:51:07.508557 4612 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 07:51:08.008540508 +0000 UTC m=+125.862470506 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 07:51:07 crc kubenswrapper[4612]: I0227 07:51:07.566492 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-96vl6" podStartSLOduration=64.566471094 podStartE2EDuration="1m4.566471094s" podCreationTimestamp="2026-02-27 07:50:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 07:51:07.503483389 +0000 UTC m=+125.357413377" watchObservedRunningTime="2026-02-27 07:51:07.566471094 +0000 UTC m=+125.420401092" Feb 27 07:51:07 crc kubenswrapper[4612]: I0227 07:51:07.609401 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pdjlb\" (UID: \"db4d6d91-61f5-4beb-809c-63ae9028a911\") " pod="openshift-image-registry/image-registry-697d97f7c8-pdjlb" Feb 27 07:51:07 crc kubenswrapper[4612]: E0227 07:51:07.613602 4612 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 07:51:08.113590263 +0000 UTC m=+125.967520261 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pdjlb" (UID: "db4d6d91-61f5-4beb-809c-63ae9028a911") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 07:51:07 crc kubenswrapper[4612]: I0227 07:51:07.640377 4612 patch_prober.go:28] interesting pod/router-default-5444994796-2d7kh container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 27 07:51:07 crc kubenswrapper[4612]: [-]has-synced failed: reason withheld Feb 27 07:51:07 crc kubenswrapper[4612]: [+]process-running ok Feb 27 07:51:07 crc kubenswrapper[4612]: healthz check failed Feb 27 07:51:07 crc kubenswrapper[4612]: I0227 07:51:07.640425 4612 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-2d7kh" podUID="faae1773-bf96-4127-8060-a40da07930ce" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 27 07:51:07 crc kubenswrapper[4612]: I0227 07:51:07.711009 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 07:51:07 crc kubenswrapper[4612]: E0227 07:51:07.711193 4612 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 07:51:08.211166302 +0000 UTC m=+126.065096300 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 07:51:07 crc kubenswrapper[4612]: I0227 07:51:07.711297 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pdjlb\" (UID: \"db4d6d91-61f5-4beb-809c-63ae9028a911\") " pod="openshift-image-registry/image-registry-697d97f7c8-pdjlb" Feb 27 07:51:07 crc kubenswrapper[4612]: I0227 07:51:07.711330 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 07:51:07 crc kubenswrapper[4612]: I0227 07:51:07.711354 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 07:51:07 crc kubenswrapper[4612]: I0227 07:51:07.711371 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 07:51:07 crc kubenswrapper[4612]: I0227 07:51:07.711390 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 07:51:07 crc kubenswrapper[4612]: E0227 07:51:07.711944 4612 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 07:51:08.211936783 +0000 UTC m=+126.065866781 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pdjlb" (UID: "db4d6d91-61f5-4beb-809c-63ae9028a911") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 07:51:07 crc kubenswrapper[4612]: I0227 07:51:07.750311 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Feb 27 07:51:07 crc kubenswrapper[4612]: I0227 07:51:07.750621 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-z9zch" podStartSLOduration=64.750600708 podStartE2EDuration="1m4.750600708s" podCreationTimestamp="2026-02-27 07:50:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 07:51:07.568005057 +0000 UTC m=+125.421935055" watchObservedRunningTime="2026-02-27 07:51:07.750600708 +0000 UTC m=+125.604530726" Feb 27 07:51:07 crc kubenswrapper[4612]: I0227 07:51:07.751438 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-cw57r" podStartSLOduration=64.751433641 podStartE2EDuration="1m4.751433641s" podCreationTimestamp="2026-02-27 07:50:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 07:51:07.698151043 +0000 UTC m=+125.552081041" watchObservedRunningTime="2026-02-27 07:51:07.751433641 +0000 UTC m=+125.605363639" Feb 27 07:51:07 crc kubenswrapper[4612]: I0227 07:51:07.753825 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 07:51:07 crc kubenswrapper[4612]: I0227 07:51:07.754572 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Feb 27 07:51:07 crc kubenswrapper[4612]: I0227 07:51:07.754789 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Feb 27 07:51:07 crc kubenswrapper[4612]: I0227 07:51:07.778592 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Feb 27 07:51:07 crc kubenswrapper[4612]: I0227 07:51:07.779763 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 07:51:07 crc kubenswrapper[4612]: I0227 07:51:07.788481 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 07:51:07 crc kubenswrapper[4612]: I0227 07:51:07.795337 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 07:51:07 crc kubenswrapper[4612]: I0227 07:51:07.815179 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-gvv4x" podStartSLOduration=64.815165068 podStartE2EDuration="1m4.815165068s" podCreationTimestamp="2026-02-27 07:50:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 07:51:07.812259107 +0000 UTC m=+125.666189105" watchObservedRunningTime="2026-02-27 07:51:07.815165068 +0000 UTC m=+125.669095066" Feb 27 07:51:07 crc kubenswrapper[4612]: I0227 07:51:07.815639 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 07:51:07 crc kubenswrapper[4612]: E0227 07:51:07.816013 4612 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 07:51:08.316000371 +0000 UTC m=+126.169930369 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 07:51:07 crc kubenswrapper[4612]: I0227 07:51:07.917495 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pdjlb\" (UID: \"db4d6d91-61f5-4beb-809c-63ae9028a911\") " pod="openshift-image-registry/image-registry-697d97f7c8-pdjlb" Feb 27 07:51:07 crc kubenswrapper[4612]: E0227 07:51:07.917944 4612 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 07:51:08.417927869 +0000 UTC m=+126.271857857 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pdjlb" (UID: "db4d6d91-61f5-4beb-809c-63ae9028a911") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 07:51:07 crc kubenswrapper[4612]: I0227 07:51:07.971890 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zpshm" podStartSLOduration=64.971870826 podStartE2EDuration="1m4.971870826s" podCreationTimestamp="2026-02-27 07:50:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 07:51:07.969502031 +0000 UTC m=+125.823432029" watchObservedRunningTime="2026-02-27 07:51:07.971870826 +0000 UTC m=+125.825800824" Feb 27 07:51:07 crc kubenswrapper[4612]: I0227 07:51:07.980995 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 27 07:51:07 crc kubenswrapper[4612]: I0227 07:51:07.996771 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 07:51:08 crc kubenswrapper[4612]: I0227 07:51:08.018332 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 07:51:08 crc kubenswrapper[4612]: E0227 07:51:08.018722 4612 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 07:51:08.518705086 +0000 UTC m=+126.372635084 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 07:51:08 crc kubenswrapper[4612]: I0227 07:51:08.018873 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 27 07:51:08 crc kubenswrapper[4612]: I0227 07:51:08.120082 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pdjlb\" (UID: \"db4d6d91-61f5-4beb-809c-63ae9028a911\") " pod="openshift-image-registry/image-registry-697d97f7c8-pdjlb" Feb 27 07:51:08 crc kubenswrapper[4612]: E0227 07:51:08.120456 4612 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 07:51:08.62043968 +0000 UTC m=+126.474369678 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pdjlb" (UID: "db4d6d91-61f5-4beb-809c-63ae9028a911") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 07:51:08 crc kubenswrapper[4612]: I0227 07:51:08.169891 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-q6h7h" podStartSLOduration=65.169875302 podStartE2EDuration="1m5.169875302s" podCreationTimestamp="2026-02-27 07:50:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 07:51:08.169146482 +0000 UTC m=+126.023076480" watchObservedRunningTime="2026-02-27 07:51:08.169875302 +0000 UTC m=+126.023805300" Feb 27 07:51:08 crc kubenswrapper[4612]: I0227 07:51:08.205216 4612 ???:1] "http: TLS handshake error from 192.168.126.11:38044: no serving certificate available for the kubelet" Feb 27 07:51:08 crc kubenswrapper[4612]: I0227 07:51:08.220957 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 07:51:08 crc kubenswrapper[4612]: E0227 07:51:08.221476 4612 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 07:51:08.721459274 +0000 UTC m=+126.575389272 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 07:51:08 crc kubenswrapper[4612]: I0227 07:51:08.322469 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pdjlb\" (UID: \"db4d6d91-61f5-4beb-809c-63ae9028a911\") " pod="openshift-image-registry/image-registry-697d97f7c8-pdjlb" Feb 27 07:51:08 crc kubenswrapper[4612]: E0227 07:51:08.322768 4612 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 07:51:08.822757425 +0000 UTC m=+126.676687423 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pdjlb" (UID: "db4d6d91-61f5-4beb-809c-63ae9028a911") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 07:51:08 crc kubenswrapper[4612]: I0227 07:51:08.347959 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2rrfz" podStartSLOduration=65.347930219 podStartE2EDuration="1m5.347930219s" podCreationTimestamp="2026-02-27 07:50:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 07:51:08.300634366 +0000 UTC m=+126.154564354" watchObservedRunningTime="2026-02-27 07:51:08.347930219 +0000 UTC m=+126.201860227" Feb 27 07:51:08 crc kubenswrapper[4612]: I0227 07:51:08.394432 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-tcqv8" podStartSLOduration=10.39441524 podStartE2EDuration="10.39441524s" podCreationTimestamp="2026-02-27 07:50:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 07:51:08.352935657 +0000 UTC m=+126.206865655" watchObservedRunningTime="2026-02-27 07:51:08.39441524 +0000 UTC m=+126.248345228" Feb 27 07:51:08 crc kubenswrapper[4612]: I0227 07:51:08.396121 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-4lpn2" podStartSLOduration=65.396113907 podStartE2EDuration="1m5.396113907s" podCreationTimestamp="2026-02-27 07:50:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 07:51:08.3933428 +0000 UTC m=+126.247272798" watchObservedRunningTime="2026-02-27 07:51:08.396113907 +0000 UTC m=+126.250043905" Feb 27 07:51:08 crc kubenswrapper[4612]: I0227 07:51:08.425769 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 07:51:08 crc kubenswrapper[4612]: E0227 07:51:08.426093 4612 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 07:51:08.926078212 +0000 UTC m=+126.780008210 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 07:51:08 crc kubenswrapper[4612]: I0227 07:51:08.445769 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-n5jbm" event={"ID":"ad4c7d8c-98d7-40c1-ae6d-21ee5655387f","Type":"ContainerStarted","Data":"9d84fb46cae311409cc26c7c91bb7cae7ae35e93dd19866d29045dd3174e5dd8"} Feb 27 07:51:08 crc kubenswrapper[4612]: I0227 07:51:08.449677 4612 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-multus/cni-sysctl-allowlist-ds-v589k" podUID="2522fcbd-fe8e-46cf-95ba-620316513df1" containerName="kube-multus-additional-cni-plugins" containerID="cri-o://19989c4604cb1985b76a2c1d2f8276aa3c4c5fb242e7ebf2619f837c570f9715" gracePeriod=30 Feb 27 07:51:08 crc kubenswrapper[4612]: I0227 07:51:08.473541 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6cz6v" Feb 27 07:51:08 crc kubenswrapper[4612]: I0227 07:51:08.478137 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-65bdd5fb46-f8cwg" Feb 27 07:51:08 crc kubenswrapper[4612]: I0227 07:51:08.499781 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-ks964" podStartSLOduration=65.499756683 podStartE2EDuration="1m5.499756683s" podCreationTimestamp="2026-02-27 07:50:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 07:51:08.4582822 +0000 UTC m=+126.312212198" watchObservedRunningTime="2026-02-27 07:51:08.499756683 +0000 UTC m=+126.353686701" Feb 27 07:51:08 crc kubenswrapper[4612]: I0227 07:51:08.534943 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pdjlb\" (UID: \"db4d6d91-61f5-4beb-809c-63ae9028a911\") " pod="openshift-image-registry/image-registry-697d97f7c8-pdjlb" Feb 27 07:51:08 crc kubenswrapper[4612]: E0227 07:51:08.535209 4612 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 07:51:09.035198079 +0000 UTC m=+126.889128077 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pdjlb" (UID: "db4d6d91-61f5-4beb-809c-63ae9028a911") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 07:51:08 crc kubenswrapper[4612]: I0227 07:51:08.611112 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-5wvhc" podStartSLOduration=65.61107867 podStartE2EDuration="1m5.61107867s" podCreationTimestamp="2026-02-27 07:50:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 07:51:08.520594557 +0000 UTC m=+126.374524585" watchObservedRunningTime="2026-02-27 07:51:08.61107867 +0000 UTC m=+126.465008668" Feb 27 07:51:08 crc kubenswrapper[4612]: I0227 07:51:08.637237 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 07:51:08 crc kubenswrapper[4612]: E0227 07:51:08.639997 4612 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 07:51:09.139982827 +0000 UTC m=+126.993912825 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 07:51:08 crc kubenswrapper[4612]: I0227 07:51:08.681312 4612 patch_prober.go:28] interesting pod/router-default-5444994796-2d7kh container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 27 07:51:08 crc kubenswrapper[4612]: [-]has-synced failed: reason withheld Feb 27 07:51:08 crc kubenswrapper[4612]: [+]process-running ok Feb 27 07:51:08 crc kubenswrapper[4612]: healthz check failed Feb 27 07:51:08 crc kubenswrapper[4612]: I0227 07:51:08.681501 4612 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-2d7kh" podUID="faae1773-bf96-4127-8060-a40da07930ce" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 27 07:51:08 crc kubenswrapper[4612]: I0227 07:51:08.743592 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pdjlb\" (UID: \"db4d6d91-61f5-4beb-809c-63ae9028a911\") " pod="openshift-image-registry/image-registry-697d97f7c8-pdjlb" Feb 27 07:51:08 crc kubenswrapper[4612]: E0227 07:51:08.744007 4612 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 07:51:09.243992183 +0000 UTC m=+127.097922181 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pdjlb" (UID: "db4d6d91-61f5-4beb-809c-63ae9028a911") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 07:51:08 crc kubenswrapper[4612]: I0227 07:51:08.844820 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 07:51:08 crc kubenswrapper[4612]: E0227 07:51:08.845200 4612 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 07:51:09.345184422 +0000 UTC m=+127.199114410 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 07:51:08 crc kubenswrapper[4612]: I0227 07:51:08.946766 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pdjlb\" (UID: \"db4d6d91-61f5-4beb-809c-63ae9028a911\") " pod="openshift-image-registry/image-registry-697d97f7c8-pdjlb" Feb 27 07:51:08 crc kubenswrapper[4612]: E0227 07:51:08.947045 4612 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 07:51:09.447034568 +0000 UTC m=+127.300964566 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pdjlb" (UID: "db4d6d91-61f5-4beb-809c-63ae9028a911") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 07:51:09 crc kubenswrapper[4612]: I0227 07:51:09.049774 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 07:51:09 crc kubenswrapper[4612]: E0227 07:51:09.049938 4612 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 07:51:09.549911883 +0000 UTC m=+127.403841881 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 07:51:09 crc kubenswrapper[4612]: I0227 07:51:09.049983 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pdjlb\" (UID: \"db4d6d91-61f5-4beb-809c-63ae9028a911\") " pod="openshift-image-registry/image-registry-697d97f7c8-pdjlb" Feb 27 07:51:09 crc kubenswrapper[4612]: E0227 07:51:09.050269 4612 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 07:51:09.550256933 +0000 UTC m=+127.404186931 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pdjlb" (UID: "db4d6d91-61f5-4beb-809c-63ae9028a911") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 07:51:09 crc kubenswrapper[4612]: I0227 07:51:09.150546 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 07:51:09 crc kubenswrapper[4612]: E0227 07:51:09.150924 4612 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 07:51:09.650905046 +0000 UTC m=+127.504835044 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 07:51:09 crc kubenswrapper[4612]: I0227 07:51:09.253442 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pdjlb\" (UID: \"db4d6d91-61f5-4beb-809c-63ae9028a911\") " pod="openshift-image-registry/image-registry-697d97f7c8-pdjlb" Feb 27 07:51:09 crc kubenswrapper[4612]: E0227 07:51:09.253980 4612 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 07:51:09.753963376 +0000 UTC m=+127.607893374 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pdjlb" (UID: "db4d6d91-61f5-4beb-809c-63ae9028a911") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 07:51:09 crc kubenswrapper[4612]: I0227 07:51:09.290987 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 27 07:51:09 crc kubenswrapper[4612]: I0227 07:51:09.355331 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 07:51:09 crc kubenswrapper[4612]: E0227 07:51:09.355586 4612 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 07:51:09.855571516 +0000 UTC m=+127.709501514 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 07:51:09 crc kubenswrapper[4612]: I0227 07:51:09.456772 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pdjlb\" (UID: \"db4d6d91-61f5-4beb-809c-63ae9028a911\") " pod="openshift-image-registry/image-registry-697d97f7c8-pdjlb" Feb 27 07:51:09 crc kubenswrapper[4612]: E0227 07:51:09.457038 4612 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 07:51:09.957027092 +0000 UTC m=+127.810957090 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pdjlb" (UID: "db4d6d91-61f5-4beb-809c-63ae9028a911") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 07:51:09 crc kubenswrapper[4612]: I0227 07:51:09.483484 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-p5fp8"] Feb 27 07:51:09 crc kubenswrapper[4612]: I0227 07:51:09.491137 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-p5fp8" Feb 27 07:51:09 crc kubenswrapper[4612]: I0227 07:51:09.497678 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Feb 27 07:51:09 crc kubenswrapper[4612]: I0227 07:51:09.506911 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-p5fp8"] Feb 27 07:51:09 crc kubenswrapper[4612]: I0227 07:51:09.514913 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-n5jbm" event={"ID":"ad4c7d8c-98d7-40c1-ae6d-21ee5655387f","Type":"ContainerStarted","Data":"a46eed9bf7df91379f9925bc6ddf91e4133a7ff9bd243efdf6ebe993ac366e14"} Feb 27 07:51:09 crc kubenswrapper[4612]: I0227 07:51:09.547637 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"fd2d688588bee4de55a250acf3aa281c565f817d6beb64a9d466f6b59c8a9b05"} Feb 27 07:51:09 crc kubenswrapper[4612]: I0227 07:51:09.547669 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"9224d6a9b1135ed665f88875104a8e54ce57d9640732b08d0a1df7f3c6e8159f"} Feb 27 07:51:09 crc kubenswrapper[4612]: I0227 07:51:09.548510 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 07:51:09 crc kubenswrapper[4612]: I0227 07:51:09.558055 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 07:51:09 crc kubenswrapper[4612]: E0227 07:51:09.558401 4612 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 07:51:10.058386555 +0000 UTC m=+127.912316553 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 07:51:09 crc kubenswrapper[4612]: W0227 07:51:09.611949 4612 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9d751cbb_f2e2_430d_9754_c882a5e924a5.slice/crio-ca74527dac1e8420752d831051dcddb03b9f15d820c4aa780fca7e6771763a8d WatchSource:0}: Error finding container ca74527dac1e8420752d831051dcddb03b9f15d820c4aa780fca7e6771763a8d: Status 404 returned error can't find the container with id ca74527dac1e8420752d831051dcddb03b9f15d820c4aa780fca7e6771763a8d Feb 27 07:51:09 crc kubenswrapper[4612]: I0227 07:51:09.615958 4612 ???:1] "http: TLS handshake error from 192.168.126.11:38046: no serving certificate available for the kubelet" Feb 27 07:51:09 crc kubenswrapper[4612]: I0227 07:51:09.649902 4612 patch_prober.go:28] interesting pod/router-default-5444994796-2d7kh container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 27 07:51:09 crc kubenswrapper[4612]: [-]has-synced failed: reason withheld Feb 27 07:51:09 crc kubenswrapper[4612]: [+]process-running ok Feb 27 07:51:09 crc kubenswrapper[4612]: healthz check failed Feb 27 07:51:09 crc kubenswrapper[4612]: I0227 07:51:09.649965 4612 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-2d7kh" podUID="faae1773-bf96-4127-8060-a40da07930ce" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 27 07:51:09 crc kubenswrapper[4612]: I0227 07:51:09.655989 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-qlhr9"] Feb 27 07:51:09 crc kubenswrapper[4612]: I0227 07:51:09.659389 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pdjlb\" (UID: \"db4d6d91-61f5-4beb-809c-63ae9028a911\") " pod="openshift-image-registry/image-registry-697d97f7c8-pdjlb" Feb 27 07:51:09 crc kubenswrapper[4612]: I0227 07:51:09.659679 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/201de829-c0b7-4d50-9a3e-ef65d3c06916-utilities\") pod \"community-operators-p5fp8\" (UID: \"201de829-c0b7-4d50-9a3e-ef65d3c06916\") " pod="openshift-marketplace/community-operators-p5fp8" Feb 27 07:51:09 crc kubenswrapper[4612]: I0227 07:51:09.659857 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/201de829-c0b7-4d50-9a3e-ef65d3c06916-catalog-content\") pod \"community-operators-p5fp8\" (UID: \"201de829-c0b7-4d50-9a3e-ef65d3c06916\") " pod="openshift-marketplace/community-operators-p5fp8" Feb 27 07:51:09 crc kubenswrapper[4612]: I0227 07:51:09.660028 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m629f\" (UniqueName: \"kubernetes.io/projected/201de829-c0b7-4d50-9a3e-ef65d3c06916-kube-api-access-m629f\") pod \"community-operators-p5fp8\" (UID: \"201de829-c0b7-4d50-9a3e-ef65d3c06916\") " pod="openshift-marketplace/community-operators-p5fp8" Feb 27 07:51:09 crc kubenswrapper[4612]: E0227 07:51:09.660678 4612 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 07:51:10.160660684 +0000 UTC m=+128.014590772 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pdjlb" (UID: "db4d6d91-61f5-4beb-809c-63ae9028a911") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 07:51:09 crc kubenswrapper[4612]: I0227 07:51:09.671022 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qlhr9" Feb 27 07:51:09 crc kubenswrapper[4612]: I0227 07:51:09.672645 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-qlhr9"] Feb 27 07:51:09 crc kubenswrapper[4612]: I0227 07:51:09.680046 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Feb 27 07:51:09 crc kubenswrapper[4612]: I0227 07:51:09.763223 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 07:51:09 crc kubenswrapper[4612]: I0227 07:51:09.763408 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/201de829-c0b7-4d50-9a3e-ef65d3c06916-utilities\") pod \"community-operators-p5fp8\" (UID: \"201de829-c0b7-4d50-9a3e-ef65d3c06916\") " pod="openshift-marketplace/community-operators-p5fp8" Feb 27 07:51:09 crc kubenswrapper[4612]: I0227 07:51:09.763445 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7faf5006-23b1-4ef4-ad29-e0e676340a7c-catalog-content\") pod \"certified-operators-qlhr9\" (UID: \"7faf5006-23b1-4ef4-ad29-e0e676340a7c\") " pod="openshift-marketplace/certified-operators-qlhr9" Feb 27 07:51:09 crc kubenswrapper[4612]: I0227 07:51:09.763466 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/201de829-c0b7-4d50-9a3e-ef65d3c06916-catalog-content\") pod \"community-operators-p5fp8\" (UID: \"201de829-c0b7-4d50-9a3e-ef65d3c06916\") " pod="openshift-marketplace/community-operators-p5fp8" Feb 27 07:51:09 crc kubenswrapper[4612]: I0227 07:51:09.763483 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m629f\" (UniqueName: \"kubernetes.io/projected/201de829-c0b7-4d50-9a3e-ef65d3c06916-kube-api-access-m629f\") pod \"community-operators-p5fp8\" (UID: \"201de829-c0b7-4d50-9a3e-ef65d3c06916\") " pod="openshift-marketplace/community-operators-p5fp8" Feb 27 07:51:09 crc kubenswrapper[4612]: I0227 07:51:09.763506 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7faf5006-23b1-4ef4-ad29-e0e676340a7c-utilities\") pod \"certified-operators-qlhr9\" (UID: \"7faf5006-23b1-4ef4-ad29-e0e676340a7c\") " pod="openshift-marketplace/certified-operators-qlhr9" Feb 27 07:51:09 crc kubenswrapper[4612]: I0227 07:51:09.763556 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qfjdb\" (UniqueName: \"kubernetes.io/projected/7faf5006-23b1-4ef4-ad29-e0e676340a7c-kube-api-access-qfjdb\") pod \"certified-operators-qlhr9\" (UID: \"7faf5006-23b1-4ef4-ad29-e0e676340a7c\") " pod="openshift-marketplace/certified-operators-qlhr9" Feb 27 07:51:09 crc kubenswrapper[4612]: E0227 07:51:09.763666 4612 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 07:51:10.263650351 +0000 UTC m=+128.117580339 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 07:51:09 crc kubenswrapper[4612]: I0227 07:51:09.764006 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/201de829-c0b7-4d50-9a3e-ef65d3c06916-utilities\") pod \"community-operators-p5fp8\" (UID: \"201de829-c0b7-4d50-9a3e-ef65d3c06916\") " pod="openshift-marketplace/community-operators-p5fp8" Feb 27 07:51:09 crc kubenswrapper[4612]: I0227 07:51:09.764203 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/201de829-c0b7-4d50-9a3e-ef65d3c06916-catalog-content\") pod \"community-operators-p5fp8\" (UID: \"201de829-c0b7-4d50-9a3e-ef65d3c06916\") " pod="openshift-marketplace/community-operators-p5fp8" Feb 27 07:51:09 crc kubenswrapper[4612]: I0227 07:51:09.827643 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m629f\" (UniqueName: \"kubernetes.io/projected/201de829-c0b7-4d50-9a3e-ef65d3c06916-kube-api-access-m629f\") pod \"community-operators-p5fp8\" (UID: \"201de829-c0b7-4d50-9a3e-ef65d3c06916\") " pod="openshift-marketplace/community-operators-p5fp8" Feb 27 07:51:09 crc kubenswrapper[4612]: I0227 07:51:09.834933 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-p5fp8" Feb 27 07:51:09 crc kubenswrapper[4612]: I0227 07:51:09.864805 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pdjlb\" (UID: \"db4d6d91-61f5-4beb-809c-63ae9028a911\") " pod="openshift-image-registry/image-registry-697d97f7c8-pdjlb" Feb 27 07:51:09 crc kubenswrapper[4612]: I0227 07:51:09.864854 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qfjdb\" (UniqueName: \"kubernetes.io/projected/7faf5006-23b1-4ef4-ad29-e0e676340a7c-kube-api-access-qfjdb\") pod \"certified-operators-qlhr9\" (UID: \"7faf5006-23b1-4ef4-ad29-e0e676340a7c\") " pod="openshift-marketplace/certified-operators-qlhr9" Feb 27 07:51:09 crc kubenswrapper[4612]: I0227 07:51:09.864889 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7faf5006-23b1-4ef4-ad29-e0e676340a7c-catalog-content\") pod \"certified-operators-qlhr9\" (UID: \"7faf5006-23b1-4ef4-ad29-e0e676340a7c\") " pod="openshift-marketplace/certified-operators-qlhr9" Feb 27 07:51:09 crc kubenswrapper[4612]: I0227 07:51:09.864912 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7faf5006-23b1-4ef4-ad29-e0e676340a7c-utilities\") pod \"certified-operators-qlhr9\" (UID: \"7faf5006-23b1-4ef4-ad29-e0e676340a7c\") " pod="openshift-marketplace/certified-operators-qlhr9" Feb 27 07:51:09 crc kubenswrapper[4612]: I0227 07:51:09.865251 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7faf5006-23b1-4ef4-ad29-e0e676340a7c-utilities\") pod \"certified-operators-qlhr9\" (UID: \"7faf5006-23b1-4ef4-ad29-e0e676340a7c\") " pod="openshift-marketplace/certified-operators-qlhr9" Feb 27 07:51:09 crc kubenswrapper[4612]: E0227 07:51:09.865498 4612 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 07:51:10.365486967 +0000 UTC m=+128.219416955 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pdjlb" (UID: "db4d6d91-61f5-4beb-809c-63ae9028a911") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 07:51:09 crc kubenswrapper[4612]: I0227 07:51:09.866149 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7faf5006-23b1-4ef4-ad29-e0e676340a7c-catalog-content\") pod \"certified-operators-qlhr9\" (UID: \"7faf5006-23b1-4ef4-ad29-e0e676340a7c\") " pod="openshift-marketplace/certified-operators-qlhr9" Feb 27 07:51:09 crc kubenswrapper[4612]: I0227 07:51:09.889797 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-qcsjx"] Feb 27 07:51:09 crc kubenswrapper[4612]: I0227 07:51:09.890724 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-qcsjx" Feb 27 07:51:09 crc kubenswrapper[4612]: I0227 07:51:09.931400 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qfjdb\" (UniqueName: \"kubernetes.io/projected/7faf5006-23b1-4ef4-ad29-e0e676340a7c-kube-api-access-qfjdb\") pod \"certified-operators-qlhr9\" (UID: \"7faf5006-23b1-4ef4-ad29-e0e676340a7c\") " pod="openshift-marketplace/certified-operators-qlhr9" Feb 27 07:51:09 crc kubenswrapper[4612]: I0227 07:51:09.970047 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 07:51:09 crc kubenswrapper[4612]: E0227 07:51:09.970420 4612 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 07:51:10.470405138 +0000 UTC m=+128.324335136 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 07:51:09 crc kubenswrapper[4612]: I0227 07:51:09.988709 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-qcsjx"] Feb 27 07:51:10 crc kubenswrapper[4612]: I0227 07:51:10.006979 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qlhr9" Feb 27 07:51:10 crc kubenswrapper[4612]: I0227 07:51:10.029520 4612 patch_prober.go:28] interesting pod/downloads-7954f5f757-zbccb container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.9:8080/\": dial tcp 10.217.0.9:8080: connect: connection refused" start-of-body= Feb 27 07:51:10 crc kubenswrapper[4612]: I0227 07:51:10.029560 4612 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-zbccb" podUID="a5bbba10-16da-4fce-ae5f-311fc5d0d6c1" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.9:8080/\": dial tcp 10.217.0.9:8080: connect: connection refused" Feb 27 07:51:10 crc kubenswrapper[4612]: I0227 07:51:10.029858 4612 patch_prober.go:28] interesting pod/downloads-7954f5f757-zbccb container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.9:8080/\": dial tcp 10.217.0.9:8080: connect: connection refused" start-of-body= Feb 27 07:51:10 crc kubenswrapper[4612]: I0227 07:51:10.029880 4612 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-zbccb" podUID="a5bbba10-16da-4fce-ae5f-311fc5d0d6c1" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.9:8080/\": dial tcp 10.217.0.9:8080: connect: connection refused" Feb 27 07:51:10 crc kubenswrapper[4612]: I0227 07:51:10.058127 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-99cd5"] Feb 27 07:51:10 crc kubenswrapper[4612]: I0227 07:51:10.059221 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-99cd5" Feb 27 07:51:10 crc kubenswrapper[4612]: I0227 07:51:10.072073 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pdjlb\" (UID: \"db4d6d91-61f5-4beb-809c-63ae9028a911\") " pod="openshift-image-registry/image-registry-697d97f7c8-pdjlb" Feb 27 07:51:10 crc kubenswrapper[4612]: I0227 07:51:10.072119 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qwhq4\" (UniqueName: \"kubernetes.io/projected/f167bfef-3a04-4363-aa30-7f75a88a35e9-kube-api-access-qwhq4\") pod \"community-operators-qcsjx\" (UID: \"f167bfef-3a04-4363-aa30-7f75a88a35e9\") " pod="openshift-marketplace/community-operators-qcsjx" Feb 27 07:51:10 crc kubenswrapper[4612]: I0227 07:51:10.072152 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f167bfef-3a04-4363-aa30-7f75a88a35e9-utilities\") pod \"community-operators-qcsjx\" (UID: \"f167bfef-3a04-4363-aa30-7f75a88a35e9\") " pod="openshift-marketplace/community-operators-qcsjx" Feb 27 07:51:10 crc kubenswrapper[4612]: I0227 07:51:10.072200 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f167bfef-3a04-4363-aa30-7f75a88a35e9-catalog-content\") pod \"community-operators-qcsjx\" (UID: \"f167bfef-3a04-4363-aa30-7f75a88a35e9\") " pod="openshift-marketplace/community-operators-qcsjx" Feb 27 07:51:10 crc kubenswrapper[4612]: E0227 07:51:10.072439 4612 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 07:51:10.57242812 +0000 UTC m=+128.426358118 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pdjlb" (UID: "db4d6d91-61f5-4beb-809c-63ae9028a911") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 07:51:10 crc kubenswrapper[4612]: I0227 07:51:10.086151 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-99cd5"] Feb 27 07:51:10 crc kubenswrapper[4612]: I0227 07:51:10.173215 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 07:51:10 crc kubenswrapper[4612]: I0227 07:51:10.173816 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f167bfef-3a04-4363-aa30-7f75a88a35e9-catalog-content\") pod \"community-operators-qcsjx\" (UID: \"f167bfef-3a04-4363-aa30-7f75a88a35e9\") " pod="openshift-marketplace/community-operators-qcsjx" Feb 27 07:51:10 crc kubenswrapper[4612]: I0227 07:51:10.174852 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f167bfef-3a04-4363-aa30-7f75a88a35e9-catalog-content\") pod \"community-operators-qcsjx\" (UID: \"f167bfef-3a04-4363-aa30-7f75a88a35e9\") " pod="openshift-marketplace/community-operators-qcsjx" Feb 27 07:51:10 crc kubenswrapper[4612]: E0227 07:51:10.174922 4612 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 07:51:10.674901363 +0000 UTC m=+128.528831361 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 07:51:10 crc kubenswrapper[4612]: I0227 07:51:10.174991 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b4a05b89-0ef2-4f07-b1de-f558a36415a3-utilities\") pod \"certified-operators-99cd5\" (UID: \"b4a05b89-0ef2-4f07-b1de-f558a36415a3\") " pod="openshift-marketplace/certified-operators-99cd5" Feb 27 07:51:10 crc kubenswrapper[4612]: I0227 07:51:10.175043 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pdjlb\" (UID: \"db4d6d91-61f5-4beb-809c-63ae9028a911\") " pod="openshift-image-registry/image-registry-697d97f7c8-pdjlb" Feb 27 07:51:10 crc kubenswrapper[4612]: I0227 07:51:10.175081 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rkh9l\" (UniqueName: \"kubernetes.io/projected/b4a05b89-0ef2-4f07-b1de-f558a36415a3-kube-api-access-rkh9l\") pod \"certified-operators-99cd5\" (UID: \"b4a05b89-0ef2-4f07-b1de-f558a36415a3\") " pod="openshift-marketplace/certified-operators-99cd5" Feb 27 07:51:10 crc kubenswrapper[4612]: I0227 07:51:10.175131 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qwhq4\" (UniqueName: \"kubernetes.io/projected/f167bfef-3a04-4363-aa30-7f75a88a35e9-kube-api-access-qwhq4\") pod \"community-operators-qcsjx\" (UID: \"f167bfef-3a04-4363-aa30-7f75a88a35e9\") " pod="openshift-marketplace/community-operators-qcsjx" Feb 27 07:51:10 crc kubenswrapper[4612]: I0227 07:51:10.175187 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f167bfef-3a04-4363-aa30-7f75a88a35e9-utilities\") pod \"community-operators-qcsjx\" (UID: \"f167bfef-3a04-4363-aa30-7f75a88a35e9\") " pod="openshift-marketplace/community-operators-qcsjx" Feb 27 07:51:10 crc kubenswrapper[4612]: I0227 07:51:10.175227 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b4a05b89-0ef2-4f07-b1de-f558a36415a3-catalog-content\") pod \"certified-operators-99cd5\" (UID: \"b4a05b89-0ef2-4f07-b1de-f558a36415a3\") " pod="openshift-marketplace/certified-operators-99cd5" Feb 27 07:51:10 crc kubenswrapper[4612]: E0227 07:51:10.175587 4612 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 07:51:10.675578252 +0000 UTC m=+128.529508270 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pdjlb" (UID: "db4d6d91-61f5-4beb-809c-63ae9028a911") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 07:51:10 crc kubenswrapper[4612]: I0227 07:51:10.176286 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f167bfef-3a04-4363-aa30-7f75a88a35e9-utilities\") pod \"community-operators-qcsjx\" (UID: \"f167bfef-3a04-4363-aa30-7f75a88a35e9\") " pod="openshift-marketplace/community-operators-qcsjx" Feb 27 07:51:10 crc kubenswrapper[4612]: I0227 07:51:10.232762 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qwhq4\" (UniqueName: \"kubernetes.io/projected/f167bfef-3a04-4363-aa30-7f75a88a35e9-kube-api-access-qwhq4\") pod \"community-operators-qcsjx\" (UID: \"f167bfef-3a04-4363-aa30-7f75a88a35e9\") " pod="openshift-marketplace/community-operators-qcsjx" Feb 27 07:51:10 crc kubenswrapper[4612]: I0227 07:51:10.254605 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-qcsjx" Feb 27 07:51:10 crc kubenswrapper[4612]: I0227 07:51:10.276618 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 07:51:10 crc kubenswrapper[4612]: I0227 07:51:10.276830 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b4a05b89-0ef2-4f07-b1de-f558a36415a3-catalog-content\") pod \"certified-operators-99cd5\" (UID: \"b4a05b89-0ef2-4f07-b1de-f558a36415a3\") " pod="openshift-marketplace/certified-operators-99cd5" Feb 27 07:51:10 crc kubenswrapper[4612]: I0227 07:51:10.276880 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b4a05b89-0ef2-4f07-b1de-f558a36415a3-utilities\") pod \"certified-operators-99cd5\" (UID: \"b4a05b89-0ef2-4f07-b1de-f558a36415a3\") " pod="openshift-marketplace/certified-operators-99cd5" Feb 27 07:51:10 crc kubenswrapper[4612]: I0227 07:51:10.276908 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rkh9l\" (UniqueName: \"kubernetes.io/projected/b4a05b89-0ef2-4f07-b1de-f558a36415a3-kube-api-access-rkh9l\") pod \"certified-operators-99cd5\" (UID: \"b4a05b89-0ef2-4f07-b1de-f558a36415a3\") " pod="openshift-marketplace/certified-operators-99cd5" Feb 27 07:51:10 crc kubenswrapper[4612]: E0227 07:51:10.277223 4612 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 07:51:10.777202933 +0000 UTC m=+128.631132931 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 07:51:10 crc kubenswrapper[4612]: I0227 07:51:10.277595 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b4a05b89-0ef2-4f07-b1de-f558a36415a3-catalog-content\") pod \"certified-operators-99cd5\" (UID: \"b4a05b89-0ef2-4f07-b1de-f558a36415a3\") " pod="openshift-marketplace/certified-operators-99cd5" Feb 27 07:51:10 crc kubenswrapper[4612]: I0227 07:51:10.277877 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b4a05b89-0ef2-4f07-b1de-f558a36415a3-utilities\") pod \"certified-operators-99cd5\" (UID: \"b4a05b89-0ef2-4f07-b1de-f558a36415a3\") " pod="openshift-marketplace/certified-operators-99cd5" Feb 27 07:51:10 crc kubenswrapper[4612]: I0227 07:51:10.325071 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rkh9l\" (UniqueName: \"kubernetes.io/projected/b4a05b89-0ef2-4f07-b1de-f558a36415a3-kube-api-access-rkh9l\") pod \"certified-operators-99cd5\" (UID: \"b4a05b89-0ef2-4f07-b1de-f558a36415a3\") " pod="openshift-marketplace/certified-operators-99cd5" Feb 27 07:51:10 crc kubenswrapper[4612]: I0227 07:51:10.374970 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-99cd5" Feb 27 07:51:10 crc kubenswrapper[4612]: I0227 07:51:10.381452 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pdjlb\" (UID: \"db4d6d91-61f5-4beb-809c-63ae9028a911\") " pod="openshift-image-registry/image-registry-697d97f7c8-pdjlb" Feb 27 07:51:10 crc kubenswrapper[4612]: E0227 07:51:10.381858 4612 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 07:51:10.881844546 +0000 UTC m=+128.735774554 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pdjlb" (UID: "db4d6d91-61f5-4beb-809c-63ae9028a911") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 07:51:10 crc kubenswrapper[4612]: I0227 07:51:10.397148 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-z9zch" Feb 27 07:51:10 crc kubenswrapper[4612]: I0227 07:51:10.482708 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 07:51:10 crc kubenswrapper[4612]: E0227 07:51:10.483022 4612 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 07:51:10.983007174 +0000 UTC m=+128.836937172 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 07:51:10 crc kubenswrapper[4612]: I0227 07:51:10.585331 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pdjlb\" (UID: \"db4d6d91-61f5-4beb-809c-63ae9028a911\") " pod="openshift-image-registry/image-registry-697d97f7c8-pdjlb" Feb 27 07:51:10 crc kubenswrapper[4612]: E0227 07:51:10.585894 4612 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 07:51:11.085880429 +0000 UTC m=+128.939810427 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pdjlb" (UID: "db4d6d91-61f5-4beb-809c-63ae9028a911") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 07:51:10 crc kubenswrapper[4612]: I0227 07:51:10.590941 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"3b227958221c3f14749f4b2a7af6e3f631606565992183cf7ce834564807930c"} Feb 27 07:51:10 crc kubenswrapper[4612]: I0227 07:51:10.590979 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"4396ccf89a38ab383a4ea63d5f4ffafab5f3bc23730db0534a20f08d48ae5d2d"} Feb 27 07:51:10 crc kubenswrapper[4612]: I0227 07:51:10.637876 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-n5jbm" event={"ID":"ad4c7d8c-98d7-40c1-ae6d-21ee5655387f","Type":"ContainerStarted","Data":"23f82c2f1dc3e06a706061d1a811858d4cf282d43aff1d9a4f15f89b7c0abb7d"} Feb 27 07:51:10 crc kubenswrapper[4612]: I0227 07:51:10.643372 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-2d7kh" Feb 27 07:51:10 crc kubenswrapper[4612]: I0227 07:51:10.652384 4612 patch_prober.go:28] interesting pod/router-default-5444994796-2d7kh container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 27 07:51:10 crc kubenswrapper[4612]: [-]has-synced failed: reason withheld Feb 27 07:51:10 crc kubenswrapper[4612]: [+]process-running ok Feb 27 07:51:10 crc kubenswrapper[4612]: healthz check failed Feb 27 07:51:10 crc kubenswrapper[4612]: I0227 07:51:10.652427 4612 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-2d7kh" podUID="faae1773-bf96-4127-8060-a40da07930ce" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 27 07:51:10 crc kubenswrapper[4612]: I0227 07:51:10.655567 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"145e36d67f979530cfe3353a7db5ab7dd0128c2cce6192adaa9b20de533de273"} Feb 27 07:51:10 crc kubenswrapper[4612]: I0227 07:51:10.655617 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"ca74527dac1e8420752d831051dcddb03b9f15d820c4aa780fca7e6771763a8d"} Feb 27 07:51:10 crc kubenswrapper[4612]: I0227 07:51:10.665960 4612 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-8wk54" Feb 27 07:51:10 crc kubenswrapper[4612]: I0227 07:51:10.666367 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-8wk54" Feb 27 07:51:10 crc kubenswrapper[4612]: I0227 07:51:10.674233 4612 patch_prober.go:28] interesting pod/console-f9d7485db-8wk54 container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.13:8443/health\": dial tcp 10.217.0.13:8443: connect: connection refused" start-of-body= Feb 27 07:51:10 crc kubenswrapper[4612]: I0227 07:51:10.674271 4612 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-8wk54" podUID="da284312-b98a-48d6-8cff-f8071c334e0b" containerName="console" probeResult="failure" output="Get \"https://10.217.0.13:8443/health\": dial tcp 10.217.0.13:8443: connect: connection refused" Feb 27 07:51:10 crc kubenswrapper[4612]: I0227 07:51:10.679404 4612 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2rrfz" Feb 27 07:51:10 crc kubenswrapper[4612]: I0227 07:51:10.679687 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2rrfz" Feb 27 07:51:10 crc kubenswrapper[4612]: I0227 07:51:10.680068 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-hp72h" Feb 27 07:51:10 crc kubenswrapper[4612]: I0227 07:51:10.691201 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 07:51:10 crc kubenswrapper[4612]: E0227 07:51:10.691844 4612 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 07:51:11.191822588 +0000 UTC m=+129.045752586 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 07:51:10 crc kubenswrapper[4612]: I0227 07:51:10.703179 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-n5jbm" podStartSLOduration=12.703165471 podStartE2EDuration="12.703165471s" podCreationTimestamp="2026-02-27 07:50:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 07:51:10.70095747 +0000 UTC m=+128.554887478" watchObservedRunningTime="2026-02-27 07:51:10.703165471 +0000 UTC m=+128.557095469" Feb 27 07:51:10 crc kubenswrapper[4612]: I0227 07:51:10.716981 4612 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2rrfz" Feb 27 07:51:10 crc kubenswrapper[4612]: I0227 07:51:10.792564 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pdjlb\" (UID: \"db4d6d91-61f5-4beb-809c-63ae9028a911\") " pod="openshift-image-registry/image-registry-697d97f7c8-pdjlb" Feb 27 07:51:10 crc kubenswrapper[4612]: E0227 07:51:10.796208 4612 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 07:51:11.296195044 +0000 UTC m=+129.150125042 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pdjlb" (UID: "db4d6d91-61f5-4beb-809c-63ae9028a911") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 07:51:10 crc kubenswrapper[4612]: I0227 07:51:10.896116 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 07:51:10 crc kubenswrapper[4612]: E0227 07:51:10.896739 4612 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 07:51:11.396722495 +0000 UTC m=+129.250652483 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 07:51:11 crc kubenswrapper[4612]: I0227 07:51:11.001437 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pdjlb\" (UID: \"db4d6d91-61f5-4beb-809c-63ae9028a911\") " pod="openshift-image-registry/image-registry-697d97f7c8-pdjlb" Feb 27 07:51:11 crc kubenswrapper[4612]: E0227 07:51:11.001755 4612 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 07:51:11.501743069 +0000 UTC m=+129.355673067 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pdjlb" (UID: "db4d6d91-61f5-4beb-809c-63ae9028a911") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 07:51:11 crc kubenswrapper[4612]: I0227 07:51:11.060762 4612 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Feb 27 07:51:11 crc kubenswrapper[4612]: I0227 07:51:11.103983 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 07:51:11 crc kubenswrapper[4612]: E0227 07:51:11.104171 4612 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 07:51:11.6041455 +0000 UTC m=+129.458075498 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 07:51:11 crc kubenswrapper[4612]: I0227 07:51:11.104336 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pdjlb\" (UID: \"db4d6d91-61f5-4beb-809c-63ae9028a911\") " pod="openshift-image-registry/image-registry-697d97f7c8-pdjlb" Feb 27 07:51:11 crc kubenswrapper[4612]: E0227 07:51:11.104665 4612 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 07:51:11.604656135 +0000 UTC m=+129.458586133 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pdjlb" (UID: "db4d6d91-61f5-4beb-809c-63ae9028a911") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 07:51:11 crc kubenswrapper[4612]: I0227 07:51:11.146487 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Feb 27 07:51:11 crc kubenswrapper[4612]: I0227 07:51:11.147348 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 27 07:51:11 crc kubenswrapper[4612]: I0227 07:51:11.153722 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Feb 27 07:51:11 crc kubenswrapper[4612]: I0227 07:51:11.154278 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Feb 27 07:51:11 crc kubenswrapper[4612]: I0227 07:51:11.188209 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Feb 27 07:51:11 crc kubenswrapper[4612]: I0227 07:51:11.209816 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 07:51:11 crc kubenswrapper[4612]: E0227 07:51:11.210160 4612 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 07:51:11.710141791 +0000 UTC m=+129.564071789 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 07:51:11 crc kubenswrapper[4612]: I0227 07:51:11.287746 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-p5fp8"] Feb 27 07:51:11 crc kubenswrapper[4612]: I0227 07:51:11.294098 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-qlhr9"] Feb 27 07:51:11 crc kubenswrapper[4612]: I0227 07:51:11.314117 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/4819654e-3f95-4da0-b2ed-91d67547381c-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"4819654e-3f95-4da0-b2ed-91d67547381c\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 27 07:51:11 crc kubenswrapper[4612]: I0227 07:51:11.314292 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4819654e-3f95-4da0-b2ed-91d67547381c-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"4819654e-3f95-4da0-b2ed-91d67547381c\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 27 07:51:11 crc kubenswrapper[4612]: I0227 07:51:11.314446 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pdjlb\" (UID: \"db4d6d91-61f5-4beb-809c-63ae9028a911\") " pod="openshift-image-registry/image-registry-697d97f7c8-pdjlb" Feb 27 07:51:11 crc kubenswrapper[4612]: E0227 07:51:11.314791 4612 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 07:51:11.814776125 +0000 UTC m=+129.668706123 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pdjlb" (UID: "db4d6d91-61f5-4beb-809c-63ae9028a911") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 07:51:11 crc kubenswrapper[4612]: I0227 07:51:11.427229 4612 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-cw57r" Feb 27 07:51:11 crc kubenswrapper[4612]: I0227 07:51:11.427486 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-cw57r" Feb 27 07:51:11 crc kubenswrapper[4612]: I0227 07:51:11.431406 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 07:51:11 crc kubenswrapper[4612]: I0227 07:51:11.431682 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/4819654e-3f95-4da0-b2ed-91d67547381c-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"4819654e-3f95-4da0-b2ed-91d67547381c\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 27 07:51:11 crc kubenswrapper[4612]: I0227 07:51:11.432133 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4819654e-3f95-4da0-b2ed-91d67547381c-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"4819654e-3f95-4da0-b2ed-91d67547381c\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 27 07:51:11 crc kubenswrapper[4612]: E0227 07:51:11.432498 4612 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 07:51:11.932481968 +0000 UTC m=+129.786411976 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 07:51:11 crc kubenswrapper[4612]: I0227 07:51:11.432533 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/4819654e-3f95-4da0-b2ed-91d67547381c-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"4819654e-3f95-4da0-b2ed-91d67547381c\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 27 07:51:11 crc kubenswrapper[4612]: I0227 07:51:11.477484 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4819654e-3f95-4da0-b2ed-91d67547381c-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"4819654e-3f95-4da0-b2ed-91d67547381c\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 27 07:51:11 crc kubenswrapper[4612]: I0227 07:51:11.534464 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pdjlb\" (UID: \"db4d6d91-61f5-4beb-809c-63ae9028a911\") " pod="openshift-image-registry/image-registry-697d97f7c8-pdjlb" Feb 27 07:51:11 crc kubenswrapper[4612]: E0227 07:51:11.535802 4612 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 07:51:12.035791315 +0000 UTC m=+129.889721313 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pdjlb" (UID: "db4d6d91-61f5-4beb-809c-63ae9028a911") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 07:51:11 crc kubenswrapper[4612]: I0227 07:51:11.608415 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Feb 27 07:51:11 crc kubenswrapper[4612]: I0227 07:51:11.612635 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 27 07:51:11 crc kubenswrapper[4612]: I0227 07:51:11.617663 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Feb 27 07:51:11 crc kubenswrapper[4612]: I0227 07:51:11.617893 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Feb 27 07:51:11 crc kubenswrapper[4612]: I0227 07:51:11.618041 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Feb 27 07:51:11 crc kubenswrapper[4612]: I0227 07:51:11.636720 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 07:51:11 crc kubenswrapper[4612]: I0227 07:51:11.636900 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/14304c19-40b6-4ba0-aaa9-7024441a03e3-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"14304c19-40b6-4ba0-aaa9-7024441a03e3\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 27 07:51:11 crc kubenswrapper[4612]: I0227 07:51:11.636977 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/14304c19-40b6-4ba0-aaa9-7024441a03e3-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"14304c19-40b6-4ba0-aaa9-7024441a03e3\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 27 07:51:11 crc kubenswrapper[4612]: E0227 07:51:11.637101 4612 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 07:51:12.137082787 +0000 UTC m=+129.991012785 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 07:51:11 crc kubenswrapper[4612]: I0227 07:51:11.640898 4612 patch_prober.go:28] interesting pod/router-default-5444994796-2d7kh container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 27 07:51:11 crc kubenswrapper[4612]: [-]has-synced failed: reason withheld Feb 27 07:51:11 crc kubenswrapper[4612]: [+]process-running ok Feb 27 07:51:11 crc kubenswrapper[4612]: healthz check failed Feb 27 07:51:11 crc kubenswrapper[4612]: I0227 07:51:11.640938 4612 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-2d7kh" podUID="faae1773-bf96-4127-8060-a40da07930ce" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 27 07:51:11 crc kubenswrapper[4612]: I0227 07:51:11.674997 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-w4brj"] Feb 27 07:51:11 crc kubenswrapper[4612]: I0227 07:51:11.675981 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-w4brj" Feb 27 07:51:11 crc kubenswrapper[4612]: I0227 07:51:11.685037 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Feb 27 07:51:11 crc kubenswrapper[4612]: I0227 07:51:11.692561 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p5fp8" event={"ID":"201de829-c0b7-4d50-9a3e-ef65d3c06916","Type":"ContainerStarted","Data":"0521755a1fd13d9dfd37e8c466d6d6f38844dd6dff978b91c3083d738fc1e1a4"} Feb 27 07:51:11 crc kubenswrapper[4612]: I0227 07:51:11.694435 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qlhr9" event={"ID":"7faf5006-23b1-4ef4-ad29-e0e676340a7c","Type":"ContainerStarted","Data":"3909413230e119c86dc9abe3e069de2ff1342a9e4a7331713645d53104392a1a"} Feb 27 07:51:11 crc kubenswrapper[4612]: I0227 07:51:11.702305 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-sxt8c" Feb 27 07:51:11 crc kubenswrapper[4612]: I0227 07:51:11.716852 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2rrfz" Feb 27 07:51:11 crc kubenswrapper[4612]: I0227 07:51:11.739259 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0bed0a85-ad8c-4670-8193-0e9a90e88d78-utilities\") pod \"redhat-marketplace-w4brj\" (UID: \"0bed0a85-ad8c-4670-8193-0e9a90e88d78\") " pod="openshift-marketplace/redhat-marketplace-w4brj" Feb 27 07:51:11 crc kubenswrapper[4612]: I0227 07:51:11.739300 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/14304c19-40b6-4ba0-aaa9-7024441a03e3-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"14304c19-40b6-4ba0-aaa9-7024441a03e3\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 27 07:51:11 crc kubenswrapper[4612]: I0227 07:51:11.739357 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0bed0a85-ad8c-4670-8193-0e9a90e88d78-catalog-content\") pod \"redhat-marketplace-w4brj\" (UID: \"0bed0a85-ad8c-4670-8193-0e9a90e88d78\") " pod="openshift-marketplace/redhat-marketplace-w4brj" Feb 27 07:51:11 crc kubenswrapper[4612]: I0227 07:51:11.739426 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cc6gh\" (UniqueName: \"kubernetes.io/projected/0bed0a85-ad8c-4670-8193-0e9a90e88d78-kube-api-access-cc6gh\") pod \"redhat-marketplace-w4brj\" (UID: \"0bed0a85-ad8c-4670-8193-0e9a90e88d78\") " pod="openshift-marketplace/redhat-marketplace-w4brj" Feb 27 07:51:11 crc kubenswrapper[4612]: I0227 07:51:11.739486 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/14304c19-40b6-4ba0-aaa9-7024441a03e3-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"14304c19-40b6-4ba0-aaa9-7024441a03e3\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 27 07:51:11 crc kubenswrapper[4612]: I0227 07:51:11.739530 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pdjlb\" (UID: \"db4d6d91-61f5-4beb-809c-63ae9028a911\") " pod="openshift-image-registry/image-registry-697d97f7c8-pdjlb" Feb 27 07:51:11 crc kubenswrapper[4612]: I0227 07:51:11.742166 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/14304c19-40b6-4ba0-aaa9-7024441a03e3-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"14304c19-40b6-4ba0-aaa9-7024441a03e3\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 27 07:51:11 crc kubenswrapper[4612]: E0227 07:51:11.742911 4612 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 07:51:12.242901463 +0000 UTC m=+130.096831461 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pdjlb" (UID: "db4d6d91-61f5-4beb-809c-63ae9028a911") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 07:51:11 crc kubenswrapper[4612]: I0227 07:51:11.758338 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-99cd5"] Feb 27 07:51:11 crc kubenswrapper[4612]: I0227 07:51:11.771946 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 27 07:51:11 crc kubenswrapper[4612]: I0227 07:51:11.783998 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-w4brj"] Feb 27 07:51:11 crc kubenswrapper[4612]: I0227 07:51:11.810585 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/14304c19-40b6-4ba0-aaa9-7024441a03e3-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"14304c19-40b6-4ba0-aaa9-7024441a03e3\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 27 07:51:11 crc kubenswrapper[4612]: I0227 07:51:11.850757 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 07:51:11 crc kubenswrapper[4612]: I0227 07:51:11.851123 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0bed0a85-ad8c-4670-8193-0e9a90e88d78-catalog-content\") pod \"redhat-marketplace-w4brj\" (UID: \"0bed0a85-ad8c-4670-8193-0e9a90e88d78\") " pod="openshift-marketplace/redhat-marketplace-w4brj" Feb 27 07:51:11 crc kubenswrapper[4612]: I0227 07:51:11.851159 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cc6gh\" (UniqueName: \"kubernetes.io/projected/0bed0a85-ad8c-4670-8193-0e9a90e88d78-kube-api-access-cc6gh\") pod \"redhat-marketplace-w4brj\" (UID: \"0bed0a85-ad8c-4670-8193-0e9a90e88d78\") " pod="openshift-marketplace/redhat-marketplace-w4brj" Feb 27 07:51:11 crc kubenswrapper[4612]: I0227 07:51:11.851224 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0bed0a85-ad8c-4670-8193-0e9a90e88d78-utilities\") pod \"redhat-marketplace-w4brj\" (UID: \"0bed0a85-ad8c-4670-8193-0e9a90e88d78\") " pod="openshift-marketplace/redhat-marketplace-w4brj" Feb 27 07:51:11 crc kubenswrapper[4612]: I0227 07:51:11.851649 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0bed0a85-ad8c-4670-8193-0e9a90e88d78-utilities\") pod \"redhat-marketplace-w4brj\" (UID: \"0bed0a85-ad8c-4670-8193-0e9a90e88d78\") " pod="openshift-marketplace/redhat-marketplace-w4brj" Feb 27 07:51:11 crc kubenswrapper[4612]: E0227 07:51:11.851740 4612 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-27 07:51:12.351724931 +0000 UTC m=+130.205654929 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 07:51:11 crc kubenswrapper[4612]: I0227 07:51:11.852100 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0bed0a85-ad8c-4670-8193-0e9a90e88d78-catalog-content\") pod \"redhat-marketplace-w4brj\" (UID: \"0bed0a85-ad8c-4670-8193-0e9a90e88d78\") " pod="openshift-marketplace/redhat-marketplace-w4brj" Feb 27 07:51:11 crc kubenswrapper[4612]: I0227 07:51:11.903125 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cc6gh\" (UniqueName: \"kubernetes.io/projected/0bed0a85-ad8c-4670-8193-0e9a90e88d78-kube-api-access-cc6gh\") pod \"redhat-marketplace-w4brj\" (UID: \"0bed0a85-ad8c-4670-8193-0e9a90e88d78\") " pod="openshift-marketplace/redhat-marketplace-w4brj" Feb 27 07:51:11 crc kubenswrapper[4612]: I0227 07:51:11.931026 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-qcsjx"] Feb 27 07:51:11 crc kubenswrapper[4612]: I0227 07:51:11.954507 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pdjlb\" (UID: \"db4d6d91-61f5-4beb-809c-63ae9028a911\") " pod="openshift-image-registry/image-registry-697d97f7c8-pdjlb" Feb 27 07:51:11 crc kubenswrapper[4612]: E0227 07:51:11.954826 4612 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-27 07:51:12.454814942 +0000 UTC m=+130.308744940 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pdjlb" (UID: "db4d6d91-61f5-4beb-809c-63ae9028a911") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 27 07:51:11 crc kubenswrapper[4612]: I0227 07:51:11.955454 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 27 07:51:11 crc kubenswrapper[4612]: W0227 07:51:11.976534 4612 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf167bfef_3a04_4363_aa30_7f75a88a35e9.slice/crio-95e30b402c583c3375346400e6dd10960d5f7f6b2bab0ceac8b32fa3e35e5ad8 WatchSource:0}: Error finding container 95e30b402c583c3375346400e6dd10960d5f7f6b2bab0ceac8b32fa3e35e5ad8: Status 404 returned error can't find the container with id 95e30b402c583c3375346400e6dd10960d5f7f6b2bab0ceac8b32fa3e35e5ad8 Feb 27 07:51:12 crc kubenswrapper[4612]: I0227 07:51:12.002924 4612 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2026-02-27T07:51:11.060788956Z","Handler":null,"Name":""} Feb 27 07:51:12 crc kubenswrapper[4612]: I0227 07:51:12.009163 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-w4brj" Feb 27 07:51:12 crc kubenswrapper[4612]: I0227 07:51:12.029413 4612 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Feb 27 07:51:12 crc kubenswrapper[4612]: I0227 07:51:12.029447 4612 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Feb 27 07:51:12 crc kubenswrapper[4612]: E0227 07:51:12.044353 4612 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="19989c4604cb1985b76a2c1d2f8276aa3c4c5fb242e7ebf2619f837c570f9715" cmd=["/bin/bash","-c","test -f /ready/ready"] Feb 27 07:51:12 crc kubenswrapper[4612]: I0227 07:51:12.050133 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-trzgg"] Feb 27 07:51:12 crc kubenswrapper[4612]: I0227 07:51:12.051339 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-trzgg" Feb 27 07:51:12 crc kubenswrapper[4612]: I0227 07:51:12.056919 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 27 07:51:12 crc kubenswrapper[4612]: I0227 07:51:12.057145 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3d35510b-1105-415a-b866-3c3fd63a646b-utilities\") pod \"redhat-marketplace-trzgg\" (UID: \"3d35510b-1105-415a-b866-3c3fd63a646b\") " pod="openshift-marketplace/redhat-marketplace-trzgg" Feb 27 07:51:12 crc kubenswrapper[4612]: I0227 07:51:12.057189 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4kwtg\" (UniqueName: \"kubernetes.io/projected/3d35510b-1105-415a-b866-3c3fd63a646b-kube-api-access-4kwtg\") pod \"redhat-marketplace-trzgg\" (UID: \"3d35510b-1105-415a-b866-3c3fd63a646b\") " pod="openshift-marketplace/redhat-marketplace-trzgg" Feb 27 07:51:12 crc kubenswrapper[4612]: I0227 07:51:12.057208 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3d35510b-1105-415a-b866-3c3fd63a646b-catalog-content\") pod \"redhat-marketplace-trzgg\" (UID: \"3d35510b-1105-415a-b866-3c3fd63a646b\") " pod="openshift-marketplace/redhat-marketplace-trzgg" Feb 27 07:51:12 crc kubenswrapper[4612]: I0227 07:51:12.072032 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-trzgg"] Feb 27 07:51:12 crc kubenswrapper[4612]: E0227 07:51:12.072251 4612 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="19989c4604cb1985b76a2c1d2f8276aa3c4c5fb242e7ebf2619f837c570f9715" cmd=["/bin/bash","-c","test -f /ready/ready"] Feb 27 07:51:12 crc kubenswrapper[4612]: E0227 07:51:12.110517 4612 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="19989c4604cb1985b76a2c1d2f8276aa3c4c5fb242e7ebf2619f837c570f9715" cmd=["/bin/bash","-c","test -f /ready/ready"] Feb 27 07:51:12 crc kubenswrapper[4612]: E0227 07:51:12.110579 4612 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openshift-multus/cni-sysctl-allowlist-ds-v589k" podUID="2522fcbd-fe8e-46cf-95ba-620316513df1" containerName="kube-multus-additional-cni-plugins" Feb 27 07:51:12 crc kubenswrapper[4612]: I0227 07:51:12.129196 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Feb 27 07:51:12 crc kubenswrapper[4612]: I0227 07:51:12.158960 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pdjlb\" (UID: \"db4d6d91-61f5-4beb-809c-63ae9028a911\") " pod="openshift-image-registry/image-registry-697d97f7c8-pdjlb" Feb 27 07:51:12 crc kubenswrapper[4612]: I0227 07:51:12.159068 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3d35510b-1105-415a-b866-3c3fd63a646b-utilities\") pod \"redhat-marketplace-trzgg\" (UID: \"3d35510b-1105-415a-b866-3c3fd63a646b\") " pod="openshift-marketplace/redhat-marketplace-trzgg" Feb 27 07:51:12 crc kubenswrapper[4612]: I0227 07:51:12.159122 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4kwtg\" (UniqueName: \"kubernetes.io/projected/3d35510b-1105-415a-b866-3c3fd63a646b-kube-api-access-4kwtg\") pod \"redhat-marketplace-trzgg\" (UID: \"3d35510b-1105-415a-b866-3c3fd63a646b\") " pod="openshift-marketplace/redhat-marketplace-trzgg" Feb 27 07:51:12 crc kubenswrapper[4612]: I0227 07:51:12.159148 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3d35510b-1105-415a-b866-3c3fd63a646b-catalog-content\") pod \"redhat-marketplace-trzgg\" (UID: \"3d35510b-1105-415a-b866-3c3fd63a646b\") " pod="openshift-marketplace/redhat-marketplace-trzgg" Feb 27 07:51:12 crc kubenswrapper[4612]: I0227 07:51:12.159820 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3d35510b-1105-415a-b866-3c3fd63a646b-catalog-content\") pod \"redhat-marketplace-trzgg\" (UID: \"3d35510b-1105-415a-b866-3c3fd63a646b\") " pod="openshift-marketplace/redhat-marketplace-trzgg" Feb 27 07:51:12 crc kubenswrapper[4612]: I0227 07:51:12.197086 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4kwtg\" (UniqueName: \"kubernetes.io/projected/3d35510b-1105-415a-b866-3c3fd63a646b-kube-api-access-4kwtg\") pod \"redhat-marketplace-trzgg\" (UID: \"3d35510b-1105-415a-b866-3c3fd63a646b\") " pod="openshift-marketplace/redhat-marketplace-trzgg" Feb 27 07:51:12 crc kubenswrapper[4612]: I0227 07:51:12.163616 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3d35510b-1105-415a-b866-3c3fd63a646b-utilities\") pod \"redhat-marketplace-trzgg\" (UID: \"3d35510b-1105-415a-b866-3c3fd63a646b\") " pod="openshift-marketplace/redhat-marketplace-trzgg" Feb 27 07:51:12 crc kubenswrapper[4612]: I0227 07:51:12.286495 4612 patch_prober.go:28] interesting pod/apiserver-76f77b778f-cw57r container/openshift-apiserver namespace/openshift-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Feb 27 07:51:12 crc kubenswrapper[4612]: [+]log ok Feb 27 07:51:12 crc kubenswrapper[4612]: [+]etcd ok Feb 27 07:51:12 crc kubenswrapper[4612]: [+]poststarthook/start-apiserver-admission-initializer ok Feb 27 07:51:12 crc kubenswrapper[4612]: [+]poststarthook/generic-apiserver-start-informers ok Feb 27 07:51:12 crc kubenswrapper[4612]: [+]poststarthook/max-in-flight-filter ok Feb 27 07:51:12 crc kubenswrapper[4612]: [+]poststarthook/storage-object-count-tracker-hook ok Feb 27 07:51:12 crc kubenswrapper[4612]: [+]poststarthook/image.openshift.io-apiserver-caches ok Feb 27 07:51:12 crc kubenswrapper[4612]: [-]poststarthook/authorization.openshift.io-bootstrapclusterroles failed: reason withheld Feb 27 07:51:12 crc kubenswrapper[4612]: [-]poststarthook/authorization.openshift.io-ensurenodebootstrap-sa failed: reason withheld Feb 27 07:51:12 crc kubenswrapper[4612]: [+]poststarthook/project.openshift.io-projectcache ok Feb 27 07:51:12 crc kubenswrapper[4612]: [+]poststarthook/project.openshift.io-projectauthorizationcache ok Feb 27 07:51:12 crc kubenswrapper[4612]: [+]poststarthook/openshift.io-startinformers ok Feb 27 07:51:12 crc kubenswrapper[4612]: [+]poststarthook/openshift.io-restmapperupdater ok Feb 27 07:51:12 crc kubenswrapper[4612]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Feb 27 07:51:12 crc kubenswrapper[4612]: livez check failed Feb 27 07:51:12 crc kubenswrapper[4612]: I0227 07:51:12.286546 4612 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-apiserver/apiserver-76f77b778f-cw57r" podUID="25a0e1aa-86e2-4acf-885b-16c26efc337a" containerName="openshift-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 27 07:51:12 crc kubenswrapper[4612]: I0227 07:51:12.315838 4612 ???:1] "http: TLS handshake error from 192.168.126.11:48308: no serving certificate available for the kubelet" Feb 27 07:51:12 crc kubenswrapper[4612]: I0227 07:51:12.385766 4612 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Feb 27 07:51:12 crc kubenswrapper[4612]: I0227 07:51:12.385847 4612 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pdjlb\" (UID: \"db4d6d91-61f5-4beb-809c-63ae9028a911\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-pdjlb" Feb 27 07:51:12 crc kubenswrapper[4612]: I0227 07:51:12.407682 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-trzgg" Feb 27 07:51:12 crc kubenswrapper[4612]: I0227 07:51:12.603142 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Feb 27 07:51:12 crc kubenswrapper[4612]: I0227 07:51:12.622316 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pdjlb\" (UID: \"db4d6d91-61f5-4beb-809c-63ae9028a911\") " pod="openshift-image-registry/image-registry-697d97f7c8-pdjlb" Feb 27 07:51:12 crc kubenswrapper[4612]: I0227 07:51:12.643642 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-fwkxc"] Feb 27 07:51:12 crc kubenswrapper[4612]: I0227 07:51:12.644717 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-fwkxc" Feb 27 07:51:12 crc kubenswrapper[4612]: I0227 07:51:12.649212 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Feb 27 07:51:12 crc kubenswrapper[4612]: I0227 07:51:12.663587 4612 patch_prober.go:28] interesting pod/router-default-5444994796-2d7kh container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 27 07:51:12 crc kubenswrapper[4612]: [-]has-synced failed: reason withheld Feb 27 07:51:12 crc kubenswrapper[4612]: [+]process-running ok Feb 27 07:51:12 crc kubenswrapper[4612]: healthz check failed Feb 27 07:51:12 crc kubenswrapper[4612]: I0227 07:51:12.663630 4612 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-2d7kh" podUID="faae1773-bf96-4127-8060-a40da07930ce" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 27 07:51:12 crc kubenswrapper[4612]: I0227 07:51:12.736876 4612 generic.go:334] "Generic (PLEG): container finished" podID="f167bfef-3a04-4363-aa30-7f75a88a35e9" containerID="28c5a083b3d3ea88b579e2dec6c6a702efbf9dbaa452442b29f513993fe5cd55" exitCode=0 Feb 27 07:51:12 crc kubenswrapper[4612]: I0227 07:51:12.736949 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qcsjx" event={"ID":"f167bfef-3a04-4363-aa30-7f75a88a35e9","Type":"ContainerDied","Data":"28c5a083b3d3ea88b579e2dec6c6a702efbf9dbaa452442b29f513993fe5cd55"} Feb 27 07:51:12 crc kubenswrapper[4612]: I0227 07:51:12.736971 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qcsjx" event={"ID":"f167bfef-3a04-4363-aa30-7f75a88a35e9","Type":"ContainerStarted","Data":"95e30b402c583c3375346400e6dd10960d5f7f6b2bab0ceac8b32fa3e35e5ad8"} Feb 27 07:51:12 crc kubenswrapper[4612]: I0227 07:51:12.744476 4612 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 27 07:51:12 crc kubenswrapper[4612]: I0227 07:51:12.758902 4612 generic.go:334] "Generic (PLEG): container finished" podID="7faf5006-23b1-4ef4-ad29-e0e676340a7c" containerID="2fb6c4cf5724b631a0b34f2088cfbfa5e0a7c694db0d32f4f330dca0cc39c7a2" exitCode=0 Feb 27 07:51:12 crc kubenswrapper[4612]: I0227 07:51:12.758960 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qlhr9" event={"ID":"7faf5006-23b1-4ef4-ad29-e0e676340a7c","Type":"ContainerDied","Data":"2fb6c4cf5724b631a0b34f2088cfbfa5e0a7c694db0d32f4f330dca0cc39c7a2"} Feb 27 07:51:12 crc kubenswrapper[4612]: I0227 07:51:12.771143 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"4819654e-3f95-4da0-b2ed-91d67547381c","Type":"ContainerStarted","Data":"da67ff02997072e0f3b621f584d388debdc218afc0b24599c17986d5e5bac4f5"} Feb 27 07:51:12 crc kubenswrapper[4612]: I0227 07:51:12.796881 4612 generic.go:334] "Generic (PLEG): container finished" podID="b4a05b89-0ef2-4f07-b1de-f558a36415a3" containerID="8ac8333061750bbd7f5d78a4ef26f5753f05fb46272b5a603b24767136bc54ac" exitCode=0 Feb 27 07:51:12 crc kubenswrapper[4612]: I0227 07:51:12.796955 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-99cd5" event={"ID":"b4a05b89-0ef2-4f07-b1de-f558a36415a3","Type":"ContainerDied","Data":"8ac8333061750bbd7f5d78a4ef26f5753f05fb46272b5a603b24767136bc54ac"} Feb 27 07:51:12 crc kubenswrapper[4612]: I0227 07:51:12.796979 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-99cd5" event={"ID":"b4a05b89-0ef2-4f07-b1de-f558a36415a3","Type":"ContainerStarted","Data":"00c5fe7ac50ab88468386ef0cc47f1da9a51457ee9170cc7305347b0d2b3a8cb"} Feb 27 07:51:12 crc kubenswrapper[4612]: I0227 07:51:12.798203 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a321b7e1-dbb6-4839-b112-14d7674639c0-utilities\") pod \"redhat-operators-fwkxc\" (UID: \"a321b7e1-dbb6-4839-b112-14d7674639c0\") " pod="openshift-marketplace/redhat-operators-fwkxc" Feb 27 07:51:12 crc kubenswrapper[4612]: I0227 07:51:12.798230 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a321b7e1-dbb6-4839-b112-14d7674639c0-catalog-content\") pod \"redhat-operators-fwkxc\" (UID: \"a321b7e1-dbb6-4839-b112-14d7674639c0\") " pod="openshift-marketplace/redhat-operators-fwkxc" Feb 27 07:51:12 crc kubenswrapper[4612]: I0227 07:51:12.798277 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5zpf5\" (UniqueName: \"kubernetes.io/projected/a321b7e1-dbb6-4839-b112-14d7674639c0-kube-api-access-5zpf5\") pod \"redhat-operators-fwkxc\" (UID: \"a321b7e1-dbb6-4839-b112-14d7674639c0\") " pod="openshift-marketplace/redhat-operators-fwkxc" Feb 27 07:51:12 crc kubenswrapper[4612]: I0227 07:51:12.804289 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-fwkxc"] Feb 27 07:51:12 crc kubenswrapper[4612]: I0227 07:51:12.840995 4612 generic.go:334] "Generic (PLEG): container finished" podID="201de829-c0b7-4d50-9a3e-ef65d3c06916" containerID="c3a1bd90b4e84308b9a8a12e1bcd3c87b094f875095f113890440ccd5ae909ad" exitCode=0 Feb 27 07:51:12 crc kubenswrapper[4612]: I0227 07:51:12.842207 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p5fp8" event={"ID":"201de829-c0b7-4d50-9a3e-ef65d3c06916","Type":"ContainerDied","Data":"c3a1bd90b4e84308b9a8a12e1bcd3c87b094f875095f113890440ccd5ae909ad"} Feb 27 07:51:12 crc kubenswrapper[4612]: I0227 07:51:12.873145 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Feb 27 07:51:12 crc kubenswrapper[4612]: I0227 07:51:12.879163 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-pdjlb" Feb 27 07:51:12 crc kubenswrapper[4612]: I0227 07:51:12.903638 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5zpf5\" (UniqueName: \"kubernetes.io/projected/a321b7e1-dbb6-4839-b112-14d7674639c0-kube-api-access-5zpf5\") pod \"redhat-operators-fwkxc\" (UID: \"a321b7e1-dbb6-4839-b112-14d7674639c0\") " pod="openshift-marketplace/redhat-operators-fwkxc" Feb 27 07:51:12 crc kubenswrapper[4612]: I0227 07:51:12.903747 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a321b7e1-dbb6-4839-b112-14d7674639c0-utilities\") pod \"redhat-operators-fwkxc\" (UID: \"a321b7e1-dbb6-4839-b112-14d7674639c0\") " pod="openshift-marketplace/redhat-operators-fwkxc" Feb 27 07:51:12 crc kubenswrapper[4612]: I0227 07:51:12.903771 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a321b7e1-dbb6-4839-b112-14d7674639c0-catalog-content\") pod \"redhat-operators-fwkxc\" (UID: \"a321b7e1-dbb6-4839-b112-14d7674639c0\") " pod="openshift-marketplace/redhat-operators-fwkxc" Feb 27 07:51:12 crc kubenswrapper[4612]: I0227 07:51:12.906841 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a321b7e1-dbb6-4839-b112-14d7674639c0-utilities\") pod \"redhat-operators-fwkxc\" (UID: \"a321b7e1-dbb6-4839-b112-14d7674639c0\") " pod="openshift-marketplace/redhat-operators-fwkxc" Feb 27 07:51:12 crc kubenswrapper[4612]: I0227 07:51:12.907078 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a321b7e1-dbb6-4839-b112-14d7674639c0-catalog-content\") pod \"redhat-operators-fwkxc\" (UID: \"a321b7e1-dbb6-4839-b112-14d7674639c0\") " pod="openshift-marketplace/redhat-operators-fwkxc" Feb 27 07:51:12 crc kubenswrapper[4612]: I0227 07:51:12.952487 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Feb 27 07:51:12 crc kubenswrapper[4612]: I0227 07:51:12.953092 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Feb 27 07:51:12 crc kubenswrapper[4612]: I0227 07:51:12.953400 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5zpf5\" (UniqueName: \"kubernetes.io/projected/a321b7e1-dbb6-4839-b112-14d7674639c0-kube-api-access-5zpf5\") pod \"redhat-operators-fwkxc\" (UID: \"a321b7e1-dbb6-4839-b112-14d7674639c0\") " pod="openshift-marketplace/redhat-operators-fwkxc" Feb 27 07:51:12 crc kubenswrapper[4612]: I0227 07:51:12.980021 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-w4brj"] Feb 27 07:51:13 crc kubenswrapper[4612]: I0227 07:51:13.045018 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-fwkxc" Feb 27 07:51:13 crc kubenswrapper[4612]: I0227 07:51:13.073039 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-fjg45"] Feb 27 07:51:13 crc kubenswrapper[4612]: I0227 07:51:13.074220 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-fjg45" Feb 27 07:51:13 crc kubenswrapper[4612]: I0227 07:51:13.107606 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-fjg45"] Feb 27 07:51:13 crc kubenswrapper[4612]: I0227 07:51:13.220270 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-97g6s\" (UniqueName: \"kubernetes.io/projected/7676c2d7-9df2-4dc8-a2bd-cde398845d26-kube-api-access-97g6s\") pod \"redhat-operators-fjg45\" (UID: \"7676c2d7-9df2-4dc8-a2bd-cde398845d26\") " pod="openshift-marketplace/redhat-operators-fjg45" Feb 27 07:51:13 crc kubenswrapper[4612]: I0227 07:51:13.220338 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7676c2d7-9df2-4dc8-a2bd-cde398845d26-catalog-content\") pod \"redhat-operators-fjg45\" (UID: \"7676c2d7-9df2-4dc8-a2bd-cde398845d26\") " pod="openshift-marketplace/redhat-operators-fjg45" Feb 27 07:51:13 crc kubenswrapper[4612]: I0227 07:51:13.220376 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7676c2d7-9df2-4dc8-a2bd-cde398845d26-utilities\") pod \"redhat-operators-fjg45\" (UID: \"7676c2d7-9df2-4dc8-a2bd-cde398845d26\") " pod="openshift-marketplace/redhat-operators-fjg45" Feb 27 07:51:13 crc kubenswrapper[4612]: I0227 07:51:13.257825 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-trzgg"] Feb 27 07:51:13 crc kubenswrapper[4612]: I0227 07:51:13.258222 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=1.258200029 podStartE2EDuration="1.258200029s" podCreationTimestamp="2026-02-27 07:51:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 07:51:13.232657286 +0000 UTC m=+131.086587284" watchObservedRunningTime="2026-02-27 07:51:13.258200029 +0000 UTC m=+131.112130027" Feb 27 07:51:13 crc kubenswrapper[4612]: I0227 07:51:13.322495 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-97g6s\" (UniqueName: \"kubernetes.io/projected/7676c2d7-9df2-4dc8-a2bd-cde398845d26-kube-api-access-97g6s\") pod \"redhat-operators-fjg45\" (UID: \"7676c2d7-9df2-4dc8-a2bd-cde398845d26\") " pod="openshift-marketplace/redhat-operators-fjg45" Feb 27 07:51:13 crc kubenswrapper[4612]: I0227 07:51:13.325623 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7676c2d7-9df2-4dc8-a2bd-cde398845d26-catalog-content\") pod \"redhat-operators-fjg45\" (UID: \"7676c2d7-9df2-4dc8-a2bd-cde398845d26\") " pod="openshift-marketplace/redhat-operators-fjg45" Feb 27 07:51:13 crc kubenswrapper[4612]: I0227 07:51:13.325669 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7676c2d7-9df2-4dc8-a2bd-cde398845d26-utilities\") pod \"redhat-operators-fjg45\" (UID: \"7676c2d7-9df2-4dc8-a2bd-cde398845d26\") " pod="openshift-marketplace/redhat-operators-fjg45" Feb 27 07:51:13 crc kubenswrapper[4612]: I0227 07:51:13.326227 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7676c2d7-9df2-4dc8-a2bd-cde398845d26-utilities\") pod \"redhat-operators-fjg45\" (UID: \"7676c2d7-9df2-4dc8-a2bd-cde398845d26\") " pod="openshift-marketplace/redhat-operators-fjg45" Feb 27 07:51:13 crc kubenswrapper[4612]: I0227 07:51:13.326474 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7676c2d7-9df2-4dc8-a2bd-cde398845d26-catalog-content\") pod \"redhat-operators-fjg45\" (UID: \"7676c2d7-9df2-4dc8-a2bd-cde398845d26\") " pod="openshift-marketplace/redhat-operators-fjg45" Feb 27 07:51:13 crc kubenswrapper[4612]: I0227 07:51:13.331846 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Feb 27 07:51:13 crc kubenswrapper[4612]: I0227 07:51:13.388916 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-97g6s\" (UniqueName: \"kubernetes.io/projected/7676c2d7-9df2-4dc8-a2bd-cde398845d26-kube-api-access-97g6s\") pod \"redhat-operators-fjg45\" (UID: \"7676c2d7-9df2-4dc8-a2bd-cde398845d26\") " pod="openshift-marketplace/redhat-operators-fjg45" Feb 27 07:51:13 crc kubenswrapper[4612]: I0227 07:51:13.411638 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-fjg45" Feb 27 07:51:13 crc kubenswrapper[4612]: I0227 07:51:13.641571 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-pdjlb"] Feb 27 07:51:13 crc kubenswrapper[4612]: I0227 07:51:13.646738 4612 patch_prober.go:28] interesting pod/router-default-5444994796-2d7kh container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 27 07:51:13 crc kubenswrapper[4612]: [-]has-synced failed: reason withheld Feb 27 07:51:13 crc kubenswrapper[4612]: [+]process-running ok Feb 27 07:51:13 crc kubenswrapper[4612]: healthz check failed Feb 27 07:51:13 crc kubenswrapper[4612]: I0227 07:51:13.646796 4612 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-2d7kh" podUID="faae1773-bf96-4127-8060-a40da07930ce" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 27 07:51:13 crc kubenswrapper[4612]: W0227 07:51:13.667801 4612 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddb4d6d91_61f5_4beb_809c_63ae9028a911.slice/crio-064fd80a4871b89d6878288363b599fb3e86994c3833cddac443d4d4dc4c77cb WatchSource:0}: Error finding container 064fd80a4871b89d6878288363b599fb3e86994c3833cddac443d4d4dc4c77cb: Status 404 returned error can't find the container with id 064fd80a4871b89d6878288363b599fb3e86994c3833cddac443d4d4dc4c77cb Feb 27 07:51:13 crc kubenswrapper[4612]: I0227 07:51:13.696558 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-fwkxc"] Feb 27 07:51:13 crc kubenswrapper[4612]: W0227 07:51:13.735160 4612 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda321b7e1_dbb6_4839_b112_14d7674639c0.slice/crio-6c30c404089e69affb635ab954734cef84c8d90b79d3e9221b069d1534b0cf54 WatchSource:0}: Error finding container 6c30c404089e69affb635ab954734cef84c8d90b79d3e9221b069d1534b0cf54: Status 404 returned error can't find the container with id 6c30c404089e69affb635ab954734cef84c8d90b79d3e9221b069d1534b0cf54 Feb 27 07:51:13 crc kubenswrapper[4612]: I0227 07:51:13.910951 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"14304c19-40b6-4ba0-aaa9-7024441a03e3","Type":"ContainerStarted","Data":"895cfc27aff1a465d8d31fb6764b8ae66d4ebde5a75a89752924329f7b87035b"} Feb 27 07:51:13 crc kubenswrapper[4612]: I0227 07:51:13.913025 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fwkxc" event={"ID":"a321b7e1-dbb6-4839-b112-14d7674639c0","Type":"ContainerStarted","Data":"6c30c404089e69affb635ab954734cef84c8d90b79d3e9221b069d1534b0cf54"} Feb 27 07:51:13 crc kubenswrapper[4612]: I0227 07:51:13.923067 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-pdjlb" event={"ID":"db4d6d91-61f5-4beb-809c-63ae9028a911","Type":"ContainerStarted","Data":"064fd80a4871b89d6878288363b599fb3e86994c3833cddac443d4d4dc4c77cb"} Feb 27 07:51:13 crc kubenswrapper[4612]: I0227 07:51:13.937060 4612 generic.go:334] "Generic (PLEG): container finished" podID="0bed0a85-ad8c-4670-8193-0e9a90e88d78" containerID="a4f3f08c42a2b8f5989c0e5f57f73f47697425566b9354deab7746a5d4af3958" exitCode=0 Feb 27 07:51:13 crc kubenswrapper[4612]: I0227 07:51:13.937132 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-w4brj" event={"ID":"0bed0a85-ad8c-4670-8193-0e9a90e88d78","Type":"ContainerDied","Data":"a4f3f08c42a2b8f5989c0e5f57f73f47697425566b9354deab7746a5d4af3958"} Feb 27 07:51:13 crc kubenswrapper[4612]: I0227 07:51:13.937157 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-w4brj" event={"ID":"0bed0a85-ad8c-4670-8193-0e9a90e88d78","Type":"ContainerStarted","Data":"1e00925de47b036507b4e069657a360fb4493d49c80db12572efba3e2366c650"} Feb 27 07:51:13 crc kubenswrapper[4612]: I0227 07:51:13.958432 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"4819654e-3f95-4da0-b2ed-91d67547381c","Type":"ContainerStarted","Data":"3d0690fe6663ade8622dda478e05216a80286418f8d8563410e7d5065d46ad2d"} Feb 27 07:51:13 crc kubenswrapper[4612]: I0227 07:51:13.982989 4612 generic.go:334] "Generic (PLEG): container finished" podID="3d35510b-1105-415a-b866-3c3fd63a646b" containerID="ec3552c14e5b1f1fcc3cb85c4168c6bcdcf7d462e6c062a92794d15c9bbdcd7f" exitCode=0 Feb 27 07:51:13 crc kubenswrapper[4612]: I0227 07:51:13.984066 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-trzgg" event={"ID":"3d35510b-1105-415a-b866-3c3fd63a646b","Type":"ContainerDied","Data":"ec3552c14e5b1f1fcc3cb85c4168c6bcdcf7d462e6c062a92794d15c9bbdcd7f"} Feb 27 07:51:13 crc kubenswrapper[4612]: I0227 07:51:13.984105 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-trzgg" event={"ID":"3d35510b-1105-415a-b866-3c3fd63a646b","Type":"ContainerStarted","Data":"8096dd2568edca1253a3d02cb9bad54dd90684eed31e4a52a6fc8808e0c88293"} Feb 27 07:51:13 crc kubenswrapper[4612]: I0227 07:51:13.987125 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-8-crc" podStartSLOduration=2.987115545 podStartE2EDuration="2.987115545s" podCreationTimestamp="2026-02-27 07:51:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 07:51:13.986621472 +0000 UTC m=+131.840551470" watchObservedRunningTime="2026-02-27 07:51:13.987115545 +0000 UTC m=+131.841045543" Feb 27 07:51:14 crc kubenswrapper[4612]: I0227 07:51:14.014944 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-fjg45"] Feb 27 07:51:14 crc kubenswrapper[4612]: I0227 07:51:14.638279 4612 patch_prober.go:28] interesting pod/router-default-5444994796-2d7kh container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 27 07:51:14 crc kubenswrapper[4612]: [-]has-synced failed: reason withheld Feb 27 07:51:14 crc kubenswrapper[4612]: [+]process-running ok Feb 27 07:51:14 crc kubenswrapper[4612]: healthz check failed Feb 27 07:51:14 crc kubenswrapper[4612]: I0227 07:51:14.638536 4612 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-2d7kh" podUID="faae1773-bf96-4127-8060-a40da07930ce" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 27 07:51:15 crc kubenswrapper[4612]: I0227 07:51:15.003519 4612 generic.go:334] "Generic (PLEG): container finished" podID="4819654e-3f95-4da0-b2ed-91d67547381c" containerID="3d0690fe6663ade8622dda478e05216a80286418f8d8563410e7d5065d46ad2d" exitCode=0 Feb 27 07:51:15 crc kubenswrapper[4612]: I0227 07:51:15.003606 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"4819654e-3f95-4da0-b2ed-91d67547381c","Type":"ContainerDied","Data":"3d0690fe6663ade8622dda478e05216a80286418f8d8563410e7d5065d46ad2d"} Feb 27 07:51:15 crc kubenswrapper[4612]: I0227 07:51:15.010654 4612 generic.go:334] "Generic (PLEG): container finished" podID="7676c2d7-9df2-4dc8-a2bd-cde398845d26" containerID="f78d4b506918d6b031a0ab9625bd103a819e4528a72a33626159b8fc4213bfd8" exitCode=0 Feb 27 07:51:15 crc kubenswrapper[4612]: I0227 07:51:15.010725 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fjg45" event={"ID":"7676c2d7-9df2-4dc8-a2bd-cde398845d26","Type":"ContainerDied","Data":"f78d4b506918d6b031a0ab9625bd103a819e4528a72a33626159b8fc4213bfd8"} Feb 27 07:51:15 crc kubenswrapper[4612]: I0227 07:51:15.010768 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fjg45" event={"ID":"7676c2d7-9df2-4dc8-a2bd-cde398845d26","Type":"ContainerStarted","Data":"5dd86e2dd7860913a54b819f2b93c1ae9b8d82b3e622ad3d886a97a478d55778"} Feb 27 07:51:15 crc kubenswrapper[4612]: I0227 07:51:15.017053 4612 generic.go:334] "Generic (PLEG): container finished" podID="14304c19-40b6-4ba0-aaa9-7024441a03e3" containerID="0fcc89836554944ebac074b2d9f4a2677a3579bbeac21e1c51be291bcbd4276b" exitCode=0 Feb 27 07:51:15 crc kubenswrapper[4612]: I0227 07:51:15.017193 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"14304c19-40b6-4ba0-aaa9-7024441a03e3","Type":"ContainerDied","Data":"0fcc89836554944ebac074b2d9f4a2677a3579bbeac21e1c51be291bcbd4276b"} Feb 27 07:51:15 crc kubenswrapper[4612]: I0227 07:51:15.030904 4612 generic.go:334] "Generic (PLEG): container finished" podID="a321b7e1-dbb6-4839-b112-14d7674639c0" containerID="041be02357fc54b1f0d14a282c3a15a136c5b290322b49ee5403ee58e1b3e221" exitCode=0 Feb 27 07:51:15 crc kubenswrapper[4612]: I0227 07:51:15.031169 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fwkxc" event={"ID":"a321b7e1-dbb6-4839-b112-14d7674639c0","Type":"ContainerDied","Data":"041be02357fc54b1f0d14a282c3a15a136c5b290322b49ee5403ee58e1b3e221"} Feb 27 07:51:15 crc kubenswrapper[4612]: I0227 07:51:15.042565 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-pdjlb" event={"ID":"db4d6d91-61f5-4beb-809c-63ae9028a911","Type":"ContainerStarted","Data":"4cc002862ee09b83f37e5daf8763574294e4fa40534d98b44c07ee66a930c157"} Feb 27 07:51:15 crc kubenswrapper[4612]: I0227 07:51:15.044345 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-pdjlb" Feb 27 07:51:15 crc kubenswrapper[4612]: I0227 07:51:15.113034 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-pdjlb" podStartSLOduration=72.113014631 podStartE2EDuration="1m12.113014631s" podCreationTimestamp="2026-02-27 07:50:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 07:51:15.08464316 +0000 UTC m=+132.938573148" watchObservedRunningTime="2026-02-27 07:51:15.113014631 +0000 UTC m=+132.966944629" Feb 27 07:51:15 crc kubenswrapper[4612]: I0227 07:51:15.637108 4612 patch_prober.go:28] interesting pod/router-default-5444994796-2d7kh container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 27 07:51:15 crc kubenswrapper[4612]: [-]has-synced failed: reason withheld Feb 27 07:51:15 crc kubenswrapper[4612]: [+]process-running ok Feb 27 07:51:15 crc kubenswrapper[4612]: healthz check failed Feb 27 07:51:15 crc kubenswrapper[4612]: I0227 07:51:15.637698 4612 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-2d7kh" podUID="faae1773-bf96-4127-8060-a40da07930ce" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 27 07:51:16 crc kubenswrapper[4612]: I0227 07:51:16.067754 4612 generic.go:334] "Generic (PLEG): container finished" podID="ea3d0df0-517f-4865-964d-3e12a313696b" containerID="9f4a63ddb5dc79f146a7585e5535f6c947de9edebadc177ca8b4362f7aa13188" exitCode=0 Feb 27 07:51:16 crc kubenswrapper[4612]: I0227 07:51:16.067832 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29536305-vmtnx" event={"ID":"ea3d0df0-517f-4865-964d-3e12a313696b","Type":"ContainerDied","Data":"9f4a63ddb5dc79f146a7585e5535f6c947de9edebadc177ca8b4362f7aa13188"} Feb 27 07:51:16 crc kubenswrapper[4612]: I0227 07:51:16.431645 4612 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-cw57r" Feb 27 07:51:16 crc kubenswrapper[4612]: I0227 07:51:16.435473 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-cw57r" Feb 27 07:51:16 crc kubenswrapper[4612]: I0227 07:51:16.551731 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 27 07:51:16 crc kubenswrapper[4612]: I0227 07:51:16.592474 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/4819654e-3f95-4da0-b2ed-91d67547381c-kubelet-dir\") pod \"4819654e-3f95-4da0-b2ed-91d67547381c\" (UID: \"4819654e-3f95-4da0-b2ed-91d67547381c\") " Feb 27 07:51:16 crc kubenswrapper[4612]: I0227 07:51:16.592536 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4819654e-3f95-4da0-b2ed-91d67547381c-kube-api-access\") pod \"4819654e-3f95-4da0-b2ed-91d67547381c\" (UID: \"4819654e-3f95-4da0-b2ed-91d67547381c\") " Feb 27 07:51:16 crc kubenswrapper[4612]: I0227 07:51:16.595237 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4819654e-3f95-4da0-b2ed-91d67547381c-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "4819654e-3f95-4da0-b2ed-91d67547381c" (UID: "4819654e-3f95-4da0-b2ed-91d67547381c"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 27 07:51:16 crc kubenswrapper[4612]: I0227 07:51:16.625070 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4819654e-3f95-4da0-b2ed-91d67547381c-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "4819654e-3f95-4da0-b2ed-91d67547381c" (UID: "4819654e-3f95-4da0-b2ed-91d67547381c"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 07:51:16 crc kubenswrapper[4612]: I0227 07:51:16.641394 4612 patch_prober.go:28] interesting pod/router-default-5444994796-2d7kh container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 27 07:51:16 crc kubenswrapper[4612]: [-]has-synced failed: reason withheld Feb 27 07:51:16 crc kubenswrapper[4612]: [+]process-running ok Feb 27 07:51:16 crc kubenswrapper[4612]: healthz check failed Feb 27 07:51:16 crc kubenswrapper[4612]: I0227 07:51:16.641442 4612 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-2d7kh" podUID="faae1773-bf96-4127-8060-a40da07930ce" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 27 07:51:16 crc kubenswrapper[4612]: I0227 07:51:16.693870 4612 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/4819654e-3f95-4da0-b2ed-91d67547381c-kubelet-dir\") on node \"crc\" DevicePath \"\"" Feb 27 07:51:16 crc kubenswrapper[4612]: I0227 07:51:16.693900 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4819654e-3f95-4da0-b2ed-91d67547381c-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 27 07:51:16 crc kubenswrapper[4612]: I0227 07:51:16.697233 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-tcqv8" Feb 27 07:51:16 crc kubenswrapper[4612]: I0227 07:51:16.762932 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 27 07:51:16 crc kubenswrapper[4612]: I0227 07:51:16.795417 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/14304c19-40b6-4ba0-aaa9-7024441a03e3-kube-api-access\") pod \"14304c19-40b6-4ba0-aaa9-7024441a03e3\" (UID: \"14304c19-40b6-4ba0-aaa9-7024441a03e3\") " Feb 27 07:51:16 crc kubenswrapper[4612]: I0227 07:51:16.795490 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/14304c19-40b6-4ba0-aaa9-7024441a03e3-kubelet-dir\") pod \"14304c19-40b6-4ba0-aaa9-7024441a03e3\" (UID: \"14304c19-40b6-4ba0-aaa9-7024441a03e3\") " Feb 27 07:51:16 crc kubenswrapper[4612]: I0227 07:51:16.796853 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/14304c19-40b6-4ba0-aaa9-7024441a03e3-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "14304c19-40b6-4ba0-aaa9-7024441a03e3" (UID: "14304c19-40b6-4ba0-aaa9-7024441a03e3"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 27 07:51:16 crc kubenswrapper[4612]: I0227 07:51:16.803260 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/14304c19-40b6-4ba0-aaa9-7024441a03e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "14304c19-40b6-4ba0-aaa9-7024441a03e3" (UID: "14304c19-40b6-4ba0-aaa9-7024441a03e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 07:51:16 crc kubenswrapper[4612]: I0227 07:51:16.924580 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/14304c19-40b6-4ba0-aaa9-7024441a03e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 27 07:51:16 crc kubenswrapper[4612]: I0227 07:51:16.924621 4612 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/14304c19-40b6-4ba0-aaa9-7024441a03e3-kubelet-dir\") on node \"crc\" DevicePath \"\"" Feb 27 07:51:17 crc kubenswrapper[4612]: I0227 07:51:17.202160 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"14304c19-40b6-4ba0-aaa9-7024441a03e3","Type":"ContainerDied","Data":"895cfc27aff1a465d8d31fb6764b8ae66d4ebde5a75a89752924329f7b87035b"} Feb 27 07:51:17 crc kubenswrapper[4612]: I0227 07:51:17.202197 4612 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="895cfc27aff1a465d8d31fb6764b8ae66d4ebde5a75a89752924329f7b87035b" Feb 27 07:51:17 crc kubenswrapper[4612]: I0227 07:51:17.202326 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 27 07:51:17 crc kubenswrapper[4612]: I0227 07:51:17.214272 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 27 07:51:17 crc kubenswrapper[4612]: I0227 07:51:17.214744 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"4819654e-3f95-4da0-b2ed-91d67547381c","Type":"ContainerDied","Data":"da67ff02997072e0f3b621f584d388debdc218afc0b24599c17986d5e5bac4f5"} Feb 27 07:51:17 crc kubenswrapper[4612]: I0227 07:51:17.214780 4612 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="da67ff02997072e0f3b621f584d388debdc218afc0b24599c17986d5e5bac4f5" Feb 27 07:51:17 crc kubenswrapper[4612]: I0227 07:51:17.469999 4612 ???:1] "http: TLS handshake error from 192.168.126.11:48324: no serving certificate available for the kubelet" Feb 27 07:51:17 crc kubenswrapper[4612]: I0227 07:51:17.624581 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29536305-vmtnx" Feb 27 07:51:17 crc kubenswrapper[4612]: I0227 07:51:17.635408 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ea3d0df0-517f-4865-964d-3e12a313696b-secret-volume\") pod \"ea3d0df0-517f-4865-964d-3e12a313696b\" (UID: \"ea3d0df0-517f-4865-964d-3e12a313696b\") " Feb 27 07:51:17 crc kubenswrapper[4612]: I0227 07:51:17.636832 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ea3d0df0-517f-4865-964d-3e12a313696b-config-volume\") pod \"ea3d0df0-517f-4865-964d-3e12a313696b\" (UID: \"ea3d0df0-517f-4865-964d-3e12a313696b\") " Feb 27 07:51:17 crc kubenswrapper[4612]: I0227 07:51:17.636982 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4rxck\" (UniqueName: \"kubernetes.io/projected/ea3d0df0-517f-4865-964d-3e12a313696b-kube-api-access-4rxck\") pod \"ea3d0df0-517f-4865-964d-3e12a313696b\" (UID: \"ea3d0df0-517f-4865-964d-3e12a313696b\") " Feb 27 07:51:17 crc kubenswrapper[4612]: I0227 07:51:17.637560 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ea3d0df0-517f-4865-964d-3e12a313696b-config-volume" (OuterVolumeSpecName: "config-volume") pod "ea3d0df0-517f-4865-964d-3e12a313696b" (UID: "ea3d0df0-517f-4865-964d-3e12a313696b"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 07:51:17 crc kubenswrapper[4612]: I0227 07:51:17.638463 4612 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ea3d0df0-517f-4865-964d-3e12a313696b-config-volume\") on node \"crc\" DevicePath \"\"" Feb 27 07:51:17 crc kubenswrapper[4612]: I0227 07:51:17.638534 4612 patch_prober.go:28] interesting pod/router-default-5444994796-2d7kh container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 27 07:51:17 crc kubenswrapper[4612]: [-]has-synced failed: reason withheld Feb 27 07:51:17 crc kubenswrapper[4612]: [+]process-running ok Feb 27 07:51:17 crc kubenswrapper[4612]: healthz check failed Feb 27 07:51:17 crc kubenswrapper[4612]: I0227 07:51:17.638583 4612 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-2d7kh" podUID="faae1773-bf96-4127-8060-a40da07930ce" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 27 07:51:17 crc kubenswrapper[4612]: I0227 07:51:17.661292 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ea3d0df0-517f-4865-964d-3e12a313696b-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "ea3d0df0-517f-4865-964d-3e12a313696b" (UID: "ea3d0df0-517f-4865-964d-3e12a313696b"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 07:51:17 crc kubenswrapper[4612]: I0227 07:51:17.672931 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ea3d0df0-517f-4865-964d-3e12a313696b-kube-api-access-4rxck" (OuterVolumeSpecName: "kube-api-access-4rxck") pod "ea3d0df0-517f-4865-964d-3e12a313696b" (UID: "ea3d0df0-517f-4865-964d-3e12a313696b"). InnerVolumeSpecName "kube-api-access-4rxck". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 07:51:17 crc kubenswrapper[4612]: I0227 07:51:17.698931 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-hb4dm" Feb 27 07:51:17 crc kubenswrapper[4612]: I0227 07:51:17.741327 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4rxck\" (UniqueName: \"kubernetes.io/projected/ea3d0df0-517f-4865-964d-3e12a313696b-kube-api-access-4rxck\") on node \"crc\" DevicePath \"\"" Feb 27 07:51:17 crc kubenswrapper[4612]: I0227 07:51:17.741361 4612 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ea3d0df0-517f-4865-964d-3e12a313696b-secret-volume\") on node \"crc\" DevicePath \"\"" Feb 27 07:51:18 crc kubenswrapper[4612]: I0227 07:51:18.221597 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29536305-vmtnx" event={"ID":"ea3d0df0-517f-4865-964d-3e12a313696b","Type":"ContainerDied","Data":"bf05a7f96e03026c6cbda7c4f30a3f4f88ae90748101d570a2eaf5bc3d545924"} Feb 27 07:51:18 crc kubenswrapper[4612]: I0227 07:51:18.221635 4612 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bf05a7f96e03026c6cbda7c4f30a3f4f88ae90748101d570a2eaf5bc3d545924" Feb 27 07:51:18 crc kubenswrapper[4612]: I0227 07:51:18.221698 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29536305-vmtnx" Feb 27 07:51:18 crc kubenswrapper[4612]: I0227 07:51:18.627682 4612 ???:1] "http: TLS handshake error from 192.168.126.11:48330: no serving certificate available for the kubelet" Feb 27 07:51:18 crc kubenswrapper[4612]: I0227 07:51:18.641197 4612 patch_prober.go:28] interesting pod/router-default-5444994796-2d7kh container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 27 07:51:18 crc kubenswrapper[4612]: [-]has-synced failed: reason withheld Feb 27 07:51:18 crc kubenswrapper[4612]: [+]process-running ok Feb 27 07:51:18 crc kubenswrapper[4612]: healthz check failed Feb 27 07:51:18 crc kubenswrapper[4612]: I0227 07:51:18.641562 4612 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-2d7kh" podUID="faae1773-bf96-4127-8060-a40da07930ce" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 27 07:51:19 crc kubenswrapper[4612]: I0227 07:51:19.636758 4612 patch_prober.go:28] interesting pod/router-default-5444994796-2d7kh container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 27 07:51:19 crc kubenswrapper[4612]: [-]has-synced failed: reason withheld Feb 27 07:51:19 crc kubenswrapper[4612]: [+]process-running ok Feb 27 07:51:19 crc kubenswrapper[4612]: healthz check failed Feb 27 07:51:19 crc kubenswrapper[4612]: I0227 07:51:19.636806 4612 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-2d7kh" podUID="faae1773-bf96-4127-8060-a40da07930ce" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 27 07:51:20 crc kubenswrapper[4612]: I0227 07:51:20.028271 4612 patch_prober.go:28] interesting pod/downloads-7954f5f757-zbccb container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.9:8080/\": dial tcp 10.217.0.9:8080: connect: connection refused" start-of-body= Feb 27 07:51:20 crc kubenswrapper[4612]: I0227 07:51:20.028243 4612 patch_prober.go:28] interesting pod/downloads-7954f5f757-zbccb container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.9:8080/\": dial tcp 10.217.0.9:8080: connect: connection refused" start-of-body= Feb 27 07:51:20 crc kubenswrapper[4612]: I0227 07:51:20.028325 4612 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-zbccb" podUID="a5bbba10-16da-4fce-ae5f-311fc5d0d6c1" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.9:8080/\": dial tcp 10.217.0.9:8080: connect: connection refused" Feb 27 07:51:20 crc kubenswrapper[4612]: I0227 07:51:20.028325 4612 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-zbccb" podUID="a5bbba10-16da-4fce-ae5f-311fc5d0d6c1" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.9:8080/\": dial tcp 10.217.0.9:8080: connect: connection refused" Feb 27 07:51:20 crc kubenswrapper[4612]: I0227 07:51:20.636733 4612 patch_prober.go:28] interesting pod/router-default-5444994796-2d7kh container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 27 07:51:20 crc kubenswrapper[4612]: [-]has-synced failed: reason withheld Feb 27 07:51:20 crc kubenswrapper[4612]: [+]process-running ok Feb 27 07:51:20 crc kubenswrapper[4612]: healthz check failed Feb 27 07:51:20 crc kubenswrapper[4612]: I0227 07:51:20.636790 4612 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-2d7kh" podUID="faae1773-bf96-4127-8060-a40da07930ce" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 27 07:51:20 crc kubenswrapper[4612]: I0227 07:51:20.660319 4612 patch_prober.go:28] interesting pod/console-f9d7485db-8wk54 container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.13:8443/health\": dial tcp 10.217.0.13:8443: connect: connection refused" start-of-body= Feb 27 07:51:20 crc kubenswrapper[4612]: I0227 07:51:20.660366 4612 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-8wk54" podUID="da284312-b98a-48d6-8cff-f8071c334e0b" containerName="console" probeResult="failure" output="Get \"https://10.217.0.13:8443/health\": dial tcp 10.217.0.13:8443: connect: connection refused" Feb 27 07:51:21 crc kubenswrapper[4612]: I0227 07:51:21.637093 4612 patch_prober.go:28] interesting pod/router-default-5444994796-2d7kh container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 27 07:51:21 crc kubenswrapper[4612]: [-]has-synced failed: reason withheld Feb 27 07:51:21 crc kubenswrapper[4612]: [+]process-running ok Feb 27 07:51:21 crc kubenswrapper[4612]: healthz check failed Feb 27 07:51:21 crc kubenswrapper[4612]: I0227 07:51:21.637495 4612 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-2d7kh" podUID="faae1773-bf96-4127-8060-a40da07930ce" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 27 07:51:22 crc kubenswrapper[4612]: E0227 07:51:22.019700 4612 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="19989c4604cb1985b76a2c1d2f8276aa3c4c5fb242e7ebf2619f837c570f9715" cmd=["/bin/bash","-c","test -f /ready/ready"] Feb 27 07:51:22 crc kubenswrapper[4612]: E0227 07:51:22.024136 4612 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="19989c4604cb1985b76a2c1d2f8276aa3c4c5fb242e7ebf2619f837c570f9715" cmd=["/bin/bash","-c","test -f /ready/ready"] Feb 27 07:51:22 crc kubenswrapper[4612]: E0227 07:51:22.025973 4612 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="19989c4604cb1985b76a2c1d2f8276aa3c4c5fb242e7ebf2619f837c570f9715" cmd=["/bin/bash","-c","test -f /ready/ready"] Feb 27 07:51:22 crc kubenswrapper[4612]: E0227 07:51:22.026046 4612 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openshift-multus/cni-sysctl-allowlist-ds-v589k" podUID="2522fcbd-fe8e-46cf-95ba-620316513df1" containerName="kube-multus-additional-cni-plugins" Feb 27 07:51:22 crc kubenswrapper[4612]: I0227 07:51:22.141216 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-65bdd5fb46-f8cwg"] Feb 27 07:51:22 crc kubenswrapper[4612]: I0227 07:51:22.141632 4612 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-65bdd5fb46-f8cwg" podUID="8da140b1-a24b-4c2a-bc20-dbf244d0283e" containerName="route-controller-manager" containerID="cri-o://123ad4605465bb9521f8ad344e05484fa6205b50d75bef0a9b3e23dd0cfc4bd0" gracePeriod=30 Feb 27 07:51:22 crc kubenswrapper[4612]: I0227 07:51:22.643681 4612 patch_prober.go:28] interesting pod/router-default-5444994796-2d7kh container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 27 07:51:22 crc kubenswrapper[4612]: [-]has-synced failed: reason withheld Feb 27 07:51:22 crc kubenswrapper[4612]: [+]process-running ok Feb 27 07:51:22 crc kubenswrapper[4612]: healthz check failed Feb 27 07:51:22 crc kubenswrapper[4612]: I0227 07:51:22.643761 4612 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-2d7kh" podUID="faae1773-bf96-4127-8060-a40da07930ce" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 27 07:51:23 crc kubenswrapper[4612]: I0227 07:51:23.340240 4612 generic.go:334] "Generic (PLEG): container finished" podID="8da140b1-a24b-4c2a-bc20-dbf244d0283e" containerID="123ad4605465bb9521f8ad344e05484fa6205b50d75bef0a9b3e23dd0cfc4bd0" exitCode=0 Feb 27 07:51:23 crc kubenswrapper[4612]: I0227 07:51:23.340288 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-65bdd5fb46-f8cwg" event={"ID":"8da140b1-a24b-4c2a-bc20-dbf244d0283e","Type":"ContainerDied","Data":"123ad4605465bb9521f8ad344e05484fa6205b50d75bef0a9b3e23dd0cfc4bd0"} Feb 27 07:51:23 crc kubenswrapper[4612]: I0227 07:51:23.637249 4612 patch_prober.go:28] interesting pod/router-default-5444994796-2d7kh container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 27 07:51:23 crc kubenswrapper[4612]: [-]has-synced failed: reason withheld Feb 27 07:51:23 crc kubenswrapper[4612]: [+]process-running ok Feb 27 07:51:23 crc kubenswrapper[4612]: healthz check failed Feb 27 07:51:23 crc kubenswrapper[4612]: I0227 07:51:23.637409 4612 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-2d7kh" podUID="faae1773-bf96-4127-8060-a40da07930ce" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 27 07:51:24 crc kubenswrapper[4612]: I0227 07:51:24.636932 4612 patch_prober.go:28] interesting pod/router-default-5444994796-2d7kh container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 27 07:51:24 crc kubenswrapper[4612]: [-]has-synced failed: reason withheld Feb 27 07:51:24 crc kubenswrapper[4612]: [+]process-running ok Feb 27 07:51:24 crc kubenswrapper[4612]: healthz check failed Feb 27 07:51:24 crc kubenswrapper[4612]: I0227 07:51:24.637227 4612 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-2d7kh" podUID="faae1773-bf96-4127-8060-a40da07930ce" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 27 07:51:25 crc kubenswrapper[4612]: I0227 07:51:25.645025 4612 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-2d7kh" Feb 27 07:51:25 crc kubenswrapper[4612]: I0227 07:51:25.651161 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-2d7kh" Feb 27 07:51:25 crc kubenswrapper[4612]: I0227 07:51:25.928599 4612 patch_prober.go:28] interesting pod/route-controller-manager-65bdd5fb46-f8cwg container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.44:8443/healthz\": dial tcp 10.217.0.44:8443: connect: connection refused" start-of-body= Feb 27 07:51:25 crc kubenswrapper[4612]: I0227 07:51:25.928660 4612 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-65bdd5fb46-f8cwg" podUID="8da140b1-a24b-4c2a-bc20-dbf244d0283e" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.44:8443/healthz\": dial tcp 10.217.0.44:8443: connect: connection refused" Feb 27 07:51:27 crc kubenswrapper[4612]: I0227 07:51:27.731883 4612 ???:1] "http: TLS handshake error from 192.168.126.11:47466: no serving certificate available for the kubelet" Feb 27 07:51:30 crc kubenswrapper[4612]: I0227 07:51:30.032387 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-zbccb" Feb 27 07:51:30 crc kubenswrapper[4612]: I0227 07:51:30.663894 4612 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-8wk54" Feb 27 07:51:30 crc kubenswrapper[4612]: I0227 07:51:30.672408 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-8wk54" Feb 27 07:51:32 crc kubenswrapper[4612]: E0227 07:51:32.012378 4612 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="19989c4604cb1985b76a2c1d2f8276aa3c4c5fb242e7ebf2619f837c570f9715" cmd=["/bin/bash","-c","test -f /ready/ready"] Feb 27 07:51:32 crc kubenswrapper[4612]: E0227 07:51:32.014715 4612 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="19989c4604cb1985b76a2c1d2f8276aa3c4c5fb242e7ebf2619f837c570f9715" cmd=["/bin/bash","-c","test -f /ready/ready"] Feb 27 07:51:32 crc kubenswrapper[4612]: E0227 07:51:32.016177 4612 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="19989c4604cb1985b76a2c1d2f8276aa3c4c5fb242e7ebf2619f837c570f9715" cmd=["/bin/bash","-c","test -f /ready/ready"] Feb 27 07:51:32 crc kubenswrapper[4612]: E0227 07:51:32.016211 4612 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openshift-multus/cni-sysctl-allowlist-ds-v589k" podUID="2522fcbd-fe8e-46cf-95ba-620316513df1" containerName="kube-multus-additional-cni-plugins" Feb 27 07:51:32 crc kubenswrapper[4612]: I0227 07:51:32.886431 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-pdjlb" Feb 27 07:51:33 crc kubenswrapper[4612]: I0227 07:51:33.738609 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-65bdd5fb46-f8cwg" Feb 27 07:51:33 crc kubenswrapper[4612]: I0227 07:51:33.770484 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-8455cd459b-d6dhv"] Feb 27 07:51:33 crc kubenswrapper[4612]: E0227 07:51:33.770679 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ea3d0df0-517f-4865-964d-3e12a313696b" containerName="collect-profiles" Feb 27 07:51:33 crc kubenswrapper[4612]: I0227 07:51:33.770705 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="ea3d0df0-517f-4865-964d-3e12a313696b" containerName="collect-profiles" Feb 27 07:51:33 crc kubenswrapper[4612]: E0227 07:51:33.770717 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="14304c19-40b6-4ba0-aaa9-7024441a03e3" containerName="pruner" Feb 27 07:51:33 crc kubenswrapper[4612]: I0227 07:51:33.770723 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="14304c19-40b6-4ba0-aaa9-7024441a03e3" containerName="pruner" Feb 27 07:51:33 crc kubenswrapper[4612]: E0227 07:51:33.770732 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4819654e-3f95-4da0-b2ed-91d67547381c" containerName="pruner" Feb 27 07:51:33 crc kubenswrapper[4612]: I0227 07:51:33.770738 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="4819654e-3f95-4da0-b2ed-91d67547381c" containerName="pruner" Feb 27 07:51:33 crc kubenswrapper[4612]: E0227 07:51:33.770746 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8da140b1-a24b-4c2a-bc20-dbf244d0283e" containerName="route-controller-manager" Feb 27 07:51:33 crc kubenswrapper[4612]: I0227 07:51:33.770753 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="8da140b1-a24b-4c2a-bc20-dbf244d0283e" containerName="route-controller-manager" Feb 27 07:51:33 crc kubenswrapper[4612]: I0227 07:51:33.774014 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="4819654e-3f95-4da0-b2ed-91d67547381c" containerName="pruner" Feb 27 07:51:33 crc kubenswrapper[4612]: I0227 07:51:33.774059 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="14304c19-40b6-4ba0-aaa9-7024441a03e3" containerName="pruner" Feb 27 07:51:33 crc kubenswrapper[4612]: I0227 07:51:33.774072 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="ea3d0df0-517f-4865-964d-3e12a313696b" containerName="collect-profiles" Feb 27 07:51:33 crc kubenswrapper[4612]: I0227 07:51:33.774084 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="8da140b1-a24b-4c2a-bc20-dbf244d0283e" containerName="route-controller-manager" Feb 27 07:51:33 crc kubenswrapper[4612]: I0227 07:51:33.774858 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-8455cd459b-d6dhv" Feb 27 07:51:33 crc kubenswrapper[4612]: I0227 07:51:33.776481 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-8455cd459b-d6dhv"] Feb 27 07:51:33 crc kubenswrapper[4612]: I0227 07:51:33.894043 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8da140b1-a24b-4c2a-bc20-dbf244d0283e-serving-cert\") pod \"8da140b1-a24b-4c2a-bc20-dbf244d0283e\" (UID: \"8da140b1-a24b-4c2a-bc20-dbf244d0283e\") " Feb 27 07:51:33 crc kubenswrapper[4612]: I0227 07:51:33.894134 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8da140b1-a24b-4c2a-bc20-dbf244d0283e-client-ca\") pod \"8da140b1-a24b-4c2a-bc20-dbf244d0283e\" (UID: \"8da140b1-a24b-4c2a-bc20-dbf244d0283e\") " Feb 27 07:51:33 crc kubenswrapper[4612]: I0227 07:51:33.894200 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8da140b1-a24b-4c2a-bc20-dbf244d0283e-config\") pod \"8da140b1-a24b-4c2a-bc20-dbf244d0283e\" (UID: \"8da140b1-a24b-4c2a-bc20-dbf244d0283e\") " Feb 27 07:51:33 crc kubenswrapper[4612]: I0227 07:51:33.894303 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gmqw9\" (UniqueName: \"kubernetes.io/projected/8da140b1-a24b-4c2a-bc20-dbf244d0283e-kube-api-access-gmqw9\") pod \"8da140b1-a24b-4c2a-bc20-dbf244d0283e\" (UID: \"8da140b1-a24b-4c2a-bc20-dbf244d0283e\") " Feb 27 07:51:33 crc kubenswrapper[4612]: I0227 07:51:33.894669 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e4a7c189-8cdc-4970-8bef-e70742f2fa7e-serving-cert\") pod \"route-controller-manager-8455cd459b-d6dhv\" (UID: \"e4a7c189-8cdc-4970-8bef-e70742f2fa7e\") " pod="openshift-route-controller-manager/route-controller-manager-8455cd459b-d6dhv" Feb 27 07:51:33 crc kubenswrapper[4612]: I0227 07:51:33.894772 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7thm5\" (UniqueName: \"kubernetes.io/projected/e4a7c189-8cdc-4970-8bef-e70742f2fa7e-kube-api-access-7thm5\") pod \"route-controller-manager-8455cd459b-d6dhv\" (UID: \"e4a7c189-8cdc-4970-8bef-e70742f2fa7e\") " pod="openshift-route-controller-manager/route-controller-manager-8455cd459b-d6dhv" Feb 27 07:51:33 crc kubenswrapper[4612]: I0227 07:51:33.894838 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e4a7c189-8cdc-4970-8bef-e70742f2fa7e-config\") pod \"route-controller-manager-8455cd459b-d6dhv\" (UID: \"e4a7c189-8cdc-4970-8bef-e70742f2fa7e\") " pod="openshift-route-controller-manager/route-controller-manager-8455cd459b-d6dhv" Feb 27 07:51:33 crc kubenswrapper[4612]: I0227 07:51:33.894919 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e4a7c189-8cdc-4970-8bef-e70742f2fa7e-client-ca\") pod \"route-controller-manager-8455cd459b-d6dhv\" (UID: \"e4a7c189-8cdc-4970-8bef-e70742f2fa7e\") " pod="openshift-route-controller-manager/route-controller-manager-8455cd459b-d6dhv" Feb 27 07:51:33 crc kubenswrapper[4612]: I0227 07:51:33.895963 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8da140b1-a24b-4c2a-bc20-dbf244d0283e-client-ca" (OuterVolumeSpecName: "client-ca") pod "8da140b1-a24b-4c2a-bc20-dbf244d0283e" (UID: "8da140b1-a24b-4c2a-bc20-dbf244d0283e"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 07:51:33 crc kubenswrapper[4612]: I0227 07:51:33.896601 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8da140b1-a24b-4c2a-bc20-dbf244d0283e-config" (OuterVolumeSpecName: "config") pod "8da140b1-a24b-4c2a-bc20-dbf244d0283e" (UID: "8da140b1-a24b-4c2a-bc20-dbf244d0283e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 07:51:33 crc kubenswrapper[4612]: I0227 07:51:33.899899 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8da140b1-a24b-4c2a-bc20-dbf244d0283e-kube-api-access-gmqw9" (OuterVolumeSpecName: "kube-api-access-gmqw9") pod "8da140b1-a24b-4c2a-bc20-dbf244d0283e" (UID: "8da140b1-a24b-4c2a-bc20-dbf244d0283e"). InnerVolumeSpecName "kube-api-access-gmqw9". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 07:51:33 crc kubenswrapper[4612]: I0227 07:51:33.906323 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8da140b1-a24b-4c2a-bc20-dbf244d0283e-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8da140b1-a24b-4c2a-bc20-dbf244d0283e" (UID: "8da140b1-a24b-4c2a-bc20-dbf244d0283e"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 07:51:33 crc kubenswrapper[4612]: I0227 07:51:33.996422 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7thm5\" (UniqueName: \"kubernetes.io/projected/e4a7c189-8cdc-4970-8bef-e70742f2fa7e-kube-api-access-7thm5\") pod \"route-controller-manager-8455cd459b-d6dhv\" (UID: \"e4a7c189-8cdc-4970-8bef-e70742f2fa7e\") " pod="openshift-route-controller-manager/route-controller-manager-8455cd459b-d6dhv" Feb 27 07:51:33 crc kubenswrapper[4612]: I0227 07:51:33.996780 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e4a7c189-8cdc-4970-8bef-e70742f2fa7e-config\") pod \"route-controller-manager-8455cd459b-d6dhv\" (UID: \"e4a7c189-8cdc-4970-8bef-e70742f2fa7e\") " pod="openshift-route-controller-manager/route-controller-manager-8455cd459b-d6dhv" Feb 27 07:51:33 crc kubenswrapper[4612]: I0227 07:51:33.996830 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e4a7c189-8cdc-4970-8bef-e70742f2fa7e-client-ca\") pod \"route-controller-manager-8455cd459b-d6dhv\" (UID: \"e4a7c189-8cdc-4970-8bef-e70742f2fa7e\") " pod="openshift-route-controller-manager/route-controller-manager-8455cd459b-d6dhv" Feb 27 07:51:33 crc kubenswrapper[4612]: I0227 07:51:33.996941 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e4a7c189-8cdc-4970-8bef-e70742f2fa7e-serving-cert\") pod \"route-controller-manager-8455cd459b-d6dhv\" (UID: \"e4a7c189-8cdc-4970-8bef-e70742f2fa7e\") " pod="openshift-route-controller-manager/route-controller-manager-8455cd459b-d6dhv" Feb 27 07:51:33 crc kubenswrapper[4612]: I0227 07:51:33.996985 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gmqw9\" (UniqueName: \"kubernetes.io/projected/8da140b1-a24b-4c2a-bc20-dbf244d0283e-kube-api-access-gmqw9\") on node \"crc\" DevicePath \"\"" Feb 27 07:51:33 crc kubenswrapper[4612]: I0227 07:51:33.997000 4612 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8da140b1-a24b-4c2a-bc20-dbf244d0283e-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 27 07:51:33 crc kubenswrapper[4612]: I0227 07:51:33.997012 4612 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8da140b1-a24b-4c2a-bc20-dbf244d0283e-client-ca\") on node \"crc\" DevicePath \"\"" Feb 27 07:51:33 crc kubenswrapper[4612]: I0227 07:51:33.997023 4612 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8da140b1-a24b-4c2a-bc20-dbf244d0283e-config\") on node \"crc\" DevicePath \"\"" Feb 27 07:51:33 crc kubenswrapper[4612]: I0227 07:51:33.997883 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e4a7c189-8cdc-4970-8bef-e70742f2fa7e-client-ca\") pod \"route-controller-manager-8455cd459b-d6dhv\" (UID: \"e4a7c189-8cdc-4970-8bef-e70742f2fa7e\") " pod="openshift-route-controller-manager/route-controller-manager-8455cd459b-d6dhv" Feb 27 07:51:33 crc kubenswrapper[4612]: I0227 07:51:33.997998 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e4a7c189-8cdc-4970-8bef-e70742f2fa7e-config\") pod \"route-controller-manager-8455cd459b-d6dhv\" (UID: \"e4a7c189-8cdc-4970-8bef-e70742f2fa7e\") " pod="openshift-route-controller-manager/route-controller-manager-8455cd459b-d6dhv" Feb 27 07:51:34 crc kubenswrapper[4612]: I0227 07:51:34.001603 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e4a7c189-8cdc-4970-8bef-e70742f2fa7e-serving-cert\") pod \"route-controller-manager-8455cd459b-d6dhv\" (UID: \"e4a7c189-8cdc-4970-8bef-e70742f2fa7e\") " pod="openshift-route-controller-manager/route-controller-manager-8455cd459b-d6dhv" Feb 27 07:51:34 crc kubenswrapper[4612]: I0227 07:51:34.014080 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7thm5\" (UniqueName: \"kubernetes.io/projected/e4a7c189-8cdc-4970-8bef-e70742f2fa7e-kube-api-access-7thm5\") pod \"route-controller-manager-8455cd459b-d6dhv\" (UID: \"e4a7c189-8cdc-4970-8bef-e70742f2fa7e\") " pod="openshift-route-controller-manager/route-controller-manager-8455cd459b-d6dhv" Feb 27 07:51:34 crc kubenswrapper[4612]: I0227 07:51:34.093502 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-8455cd459b-d6dhv" Feb 27 07:51:34 crc kubenswrapper[4612]: I0227 07:51:34.437349 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-65bdd5fb46-f8cwg" event={"ID":"8da140b1-a24b-4c2a-bc20-dbf244d0283e","Type":"ContainerDied","Data":"a9feef3c425f375dd5a74438f987b5f6944eb30246a99c1477e7b1b83c3cbf82"} Feb 27 07:51:34 crc kubenswrapper[4612]: I0227 07:51:34.437403 4612 scope.go:117] "RemoveContainer" containerID="123ad4605465bb9521f8ad344e05484fa6205b50d75bef0a9b3e23dd0cfc4bd0" Feb 27 07:51:34 crc kubenswrapper[4612]: I0227 07:51:34.437533 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-65bdd5fb46-f8cwg" Feb 27 07:51:34 crc kubenswrapper[4612]: I0227 07:51:34.473168 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-65bdd5fb46-f8cwg"] Feb 27 07:51:34 crc kubenswrapper[4612]: I0227 07:51:34.476014 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-65bdd5fb46-f8cwg"] Feb 27 07:51:34 crc kubenswrapper[4612]: I0227 07:51:34.858975 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8da140b1-a24b-4c2a-bc20-dbf244d0283e" path="/var/lib/kubelet/pods/8da140b1-a24b-4c2a-bc20-dbf244d0283e/volumes" Feb 27 07:51:35 crc kubenswrapper[4612]: I0227 07:51:35.451988 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-controller-manager_controller-manager-879f6c89f-sxt8c_738d2b6e-2b7c-461d-b51e-f35eb5d41363/controller-manager/0.log" Feb 27 07:51:35 crc kubenswrapper[4612]: I0227 07:51:35.452031 4612 generic.go:334] "Generic (PLEG): container finished" podID="738d2b6e-2b7c-461d-b51e-f35eb5d41363" containerID="80058f3d8b6825a6aa02bba10ea42f4f51630b27d2fc39c813f4cc5bb5412319" exitCode=137 Feb 27 07:51:35 crc kubenswrapper[4612]: I0227 07:51:35.452057 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-sxt8c" event={"ID":"738d2b6e-2b7c-461d-b51e-f35eb5d41363","Type":"ContainerDied","Data":"80058f3d8b6825a6aa02bba10ea42f4f51630b27d2fc39c813f4cc5bb5412319"} Feb 27 07:51:39 crc kubenswrapper[4612]: I0227 07:51:39.474319 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_cni-sysctl-allowlist-ds-v589k_2522fcbd-fe8e-46cf-95ba-620316513df1/kube-multus-additional-cni-plugins/0.log" Feb 27 07:51:39 crc kubenswrapper[4612]: I0227 07:51:39.474547 4612 generic.go:334] "Generic (PLEG): container finished" podID="2522fcbd-fe8e-46cf-95ba-620316513df1" containerID="19989c4604cb1985b76a2c1d2f8276aa3c4c5fb242e7ebf2619f837c570f9715" exitCode=137 Feb 27 07:51:39 crc kubenswrapper[4612]: I0227 07:51:39.474590 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/cni-sysctl-allowlist-ds-v589k" event={"ID":"2522fcbd-fe8e-46cf-95ba-620316513df1","Type":"ContainerDied","Data":"19989c4604cb1985b76a2c1d2f8276aa3c4c5fb242e7ebf2619f837c570f9715"} Feb 27 07:51:41 crc kubenswrapper[4612]: I0227 07:51:41.649966 4612 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-sxt8c container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.16:8443/healthz\": dial tcp 10.217.0.16:8443: connect: connection refused" start-of-body= Feb 27 07:51:41 crc kubenswrapper[4612]: I0227 07:51:41.650022 4612 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-sxt8c" podUID="738d2b6e-2b7c-461d-b51e-f35eb5d41363" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.16:8443/healthz\": dial tcp 10.217.0.16:8443: connect: connection refused" Feb 27 07:51:41 crc kubenswrapper[4612]: I0227 07:51:41.967438 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-gvv4x" Feb 27 07:51:42 crc kubenswrapper[4612]: E0227 07:51:42.012820 4612 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 19989c4604cb1985b76a2c1d2f8276aa3c4c5fb242e7ebf2619f837c570f9715 is running failed: container process not found" containerID="19989c4604cb1985b76a2c1d2f8276aa3c4c5fb242e7ebf2619f837c570f9715" cmd=["/bin/bash","-c","test -f /ready/ready"] Feb 27 07:51:42 crc kubenswrapper[4612]: E0227 07:51:42.014131 4612 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 19989c4604cb1985b76a2c1d2f8276aa3c4c5fb242e7ebf2619f837c570f9715 is running failed: container process not found" containerID="19989c4604cb1985b76a2c1d2f8276aa3c4c5fb242e7ebf2619f837c570f9715" cmd=["/bin/bash","-c","test -f /ready/ready"] Feb 27 07:51:42 crc kubenswrapper[4612]: E0227 07:51:42.014403 4612 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 19989c4604cb1985b76a2c1d2f8276aa3c4c5fb242e7ebf2619f837c570f9715 is running failed: container process not found" containerID="19989c4604cb1985b76a2c1d2f8276aa3c4c5fb242e7ebf2619f837c570f9715" cmd=["/bin/bash","-c","test -f /ready/ready"] Feb 27 07:51:42 crc kubenswrapper[4612]: E0227 07:51:42.014458 4612 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 19989c4604cb1985b76a2c1d2f8276aa3c4c5fb242e7ebf2619f837c570f9715 is running failed: container process not found" probeType="Readiness" pod="openshift-multus/cni-sysctl-allowlist-ds-v589k" podUID="2522fcbd-fe8e-46cf-95ba-620316513df1" containerName="kube-multus-additional-cni-plugins" Feb 27 07:51:42 crc kubenswrapper[4612]: I0227 07:51:42.212142 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-8455cd459b-d6dhv"] Feb 27 07:51:42 crc kubenswrapper[4612]: E0227 07:51:42.384449 4612 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Feb 27 07:51:42 crc kubenswrapper[4612]: E0227 07:51:42.384895 4612 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-5zpf5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-fwkxc_openshift-marketplace(a321b7e1-dbb6-4839-b112-14d7674639c0): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Feb 27 07:51:42 crc kubenswrapper[4612]: E0227 07:51:42.386074 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-fwkxc" podUID="a321b7e1-dbb6-4839-b112-14d7674639c0" Feb 27 07:51:42 crc kubenswrapper[4612]: I0227 07:51:42.868212 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Feb 27 07:51:44 crc kubenswrapper[4612]: E0227 07:51:44.295804 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-fwkxc" podUID="a321b7e1-dbb6-4839-b112-14d7674639c0" Feb 27 07:51:44 crc kubenswrapper[4612]: E0227 07:51:44.381856 4612 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Feb 27 07:51:44 crc kubenswrapper[4612]: E0227 07:51:44.382009 4612 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rkh9l,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-99cd5_openshift-marketplace(b4a05b89-0ef2-4f07-b1de-f558a36415a3): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Feb 27 07:51:44 crc kubenswrapper[4612]: E0227 07:51:44.383174 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-99cd5" podUID="b4a05b89-0ef2-4f07-b1de-f558a36415a3" Feb 27 07:51:44 crc kubenswrapper[4612]: I0227 07:51:44.546858 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=2.546839098 podStartE2EDuration="2.546839098s" podCreationTimestamp="2026-02-27 07:51:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 07:51:44.528635996 +0000 UTC m=+162.382565994" watchObservedRunningTime="2026-02-27 07:51:44.546839098 +0000 UTC m=+162.400769096" Feb 27 07:51:45 crc kubenswrapper[4612]: E0227 07:51:45.968451 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-99cd5" podUID="b4a05b89-0ef2-4f07-b1de-f558a36415a3" Feb 27 07:51:46 crc kubenswrapper[4612]: I0227 07:51:46.066617 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_cni-sysctl-allowlist-ds-v589k_2522fcbd-fe8e-46cf-95ba-620316513df1/kube-multus-additional-cni-plugins/0.log" Feb 27 07:51:46 crc kubenswrapper[4612]: I0227 07:51:46.066812 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-multus/cni-sysctl-allowlist-ds-v589k" Feb 27 07:51:46 crc kubenswrapper[4612]: I0227 07:51:46.082994 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcdhz\" (UniqueName: \"kubernetes.io/projected/2522fcbd-fe8e-46cf-95ba-620316513df1-kube-api-access-pcdhz\") pod \"2522fcbd-fe8e-46cf-95ba-620316513df1\" (UID: \"2522fcbd-fe8e-46cf-95ba-620316513df1\") " Feb 27 07:51:46 crc kubenswrapper[4612]: I0227 07:51:46.083041 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ready\" (UniqueName: \"kubernetes.io/empty-dir/2522fcbd-fe8e-46cf-95ba-620316513df1-ready\") pod \"2522fcbd-fe8e-46cf-95ba-620316513df1\" (UID: \"2522fcbd-fe8e-46cf-95ba-620316513df1\") " Feb 27 07:51:46 crc kubenswrapper[4612]: I0227 07:51:46.083066 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/2522fcbd-fe8e-46cf-95ba-620316513df1-cni-sysctl-allowlist\") pod \"2522fcbd-fe8e-46cf-95ba-620316513df1\" (UID: \"2522fcbd-fe8e-46cf-95ba-620316513df1\") " Feb 27 07:51:46 crc kubenswrapper[4612]: I0227 07:51:46.083162 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/2522fcbd-fe8e-46cf-95ba-620316513df1-tuning-conf-dir\") pod \"2522fcbd-fe8e-46cf-95ba-620316513df1\" (UID: \"2522fcbd-fe8e-46cf-95ba-620316513df1\") " Feb 27 07:51:46 crc kubenswrapper[4612]: I0227 07:51:46.083397 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2522fcbd-fe8e-46cf-95ba-620316513df1-tuning-conf-dir" (OuterVolumeSpecName: "tuning-conf-dir") pod "2522fcbd-fe8e-46cf-95ba-620316513df1" (UID: "2522fcbd-fe8e-46cf-95ba-620316513df1"). InnerVolumeSpecName "tuning-conf-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 27 07:51:46 crc kubenswrapper[4612]: I0227 07:51:46.084633 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2522fcbd-fe8e-46cf-95ba-620316513df1-ready" (OuterVolumeSpecName: "ready") pod "2522fcbd-fe8e-46cf-95ba-620316513df1" (UID: "2522fcbd-fe8e-46cf-95ba-620316513df1"). InnerVolumeSpecName "ready". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 07:51:46 crc kubenswrapper[4612]: I0227 07:51:46.085732 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2522fcbd-fe8e-46cf-95ba-620316513df1-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "2522fcbd-fe8e-46cf-95ba-620316513df1" (UID: "2522fcbd-fe8e-46cf-95ba-620316513df1"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 07:51:46 crc kubenswrapper[4612]: I0227 07:51:46.093742 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2522fcbd-fe8e-46cf-95ba-620316513df1-kube-api-access-pcdhz" (OuterVolumeSpecName: "kube-api-access-pcdhz") pod "2522fcbd-fe8e-46cf-95ba-620316513df1" (UID: "2522fcbd-fe8e-46cf-95ba-620316513df1"). InnerVolumeSpecName "kube-api-access-pcdhz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 07:51:46 crc kubenswrapper[4612]: E0227 07:51:46.110876 4612 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Feb 27 07:51:46 crc kubenswrapper[4612]: E0227 07:51:46.111023 4612 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-m629f,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-p5fp8_openshift-marketplace(201de829-c0b7-4d50-9a3e-ef65d3c06916): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Feb 27 07:51:46 crc kubenswrapper[4612]: E0227 07:51:46.112381 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-p5fp8" podUID="201de829-c0b7-4d50-9a3e-ef65d3c06916" Feb 27 07:51:46 crc kubenswrapper[4612]: E0227 07:51:46.117216 4612 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Feb 27 07:51:46 crc kubenswrapper[4612]: E0227 07:51:46.117318 4612 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-qwhq4,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-qcsjx_openshift-marketplace(f167bfef-3a04-4363-aa30-7f75a88a35e9): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Feb 27 07:51:46 crc kubenswrapper[4612]: E0227 07:51:46.118388 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-qcsjx" podUID="f167bfef-3a04-4363-aa30-7f75a88a35e9" Feb 27 07:51:46 crc kubenswrapper[4612]: E0227 07:51:46.125259 4612 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Feb 27 07:51:46 crc kubenswrapper[4612]: E0227 07:51:46.125382 4612 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-cc6gh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-w4brj_openshift-marketplace(0bed0a85-ad8c-4670-8193-0e9a90e88d78): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Feb 27 07:51:46 crc kubenswrapper[4612]: E0227 07:51:46.126561 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-w4brj" podUID="0bed0a85-ad8c-4670-8193-0e9a90e88d78" Feb 27 07:51:46 crc kubenswrapper[4612]: E0227 07:51:46.170725 4612 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Feb 27 07:51:46 crc kubenswrapper[4612]: E0227 07:51:46.171253 4612 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-97g6s,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-fjg45_openshift-marketplace(7676c2d7-9df2-4dc8-a2bd-cde398845d26): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Feb 27 07:51:46 crc kubenswrapper[4612]: E0227 07:51:46.174302 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-fjg45" podUID="7676c2d7-9df2-4dc8-a2bd-cde398845d26" Feb 27 07:51:46 crc kubenswrapper[4612]: E0227 07:51:46.180535 4612 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Feb 27 07:51:46 crc kubenswrapper[4612]: E0227 07:51:46.180676 4612 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-qfjdb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-qlhr9_openshift-marketplace(7faf5006-23b1-4ef4-ad29-e0e676340a7c): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Feb 27 07:51:46 crc kubenswrapper[4612]: E0227 07:51:46.182132 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-qlhr9" podUID="7faf5006-23b1-4ef4-ad29-e0e676340a7c" Feb 27 07:51:46 crc kubenswrapper[4612]: I0227 07:51:46.184927 4612 reconciler_common.go:293] "Volume detached for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/2522fcbd-fe8e-46cf-95ba-620316513df1-tuning-conf-dir\") on node \"crc\" DevicePath \"\"" Feb 27 07:51:46 crc kubenswrapper[4612]: I0227 07:51:46.184992 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcdhz\" (UniqueName: \"kubernetes.io/projected/2522fcbd-fe8e-46cf-95ba-620316513df1-kube-api-access-pcdhz\") on node \"crc\" DevicePath \"\"" Feb 27 07:51:46 crc kubenswrapper[4612]: I0227 07:51:46.185006 4612 reconciler_common.go:293] "Volume detached for volume \"ready\" (UniqueName: \"kubernetes.io/empty-dir/2522fcbd-fe8e-46cf-95ba-620316513df1-ready\") on node \"crc\" DevicePath \"\"" Feb 27 07:51:46 crc kubenswrapper[4612]: I0227 07:51:46.185042 4612 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/2522fcbd-fe8e-46cf-95ba-620316513df1-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Feb 27 07:51:46 crc kubenswrapper[4612]: I0227 07:51:46.415579 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-controller-manager_controller-manager-879f6c89f-sxt8c_738d2b6e-2b7c-461d-b51e-f35eb5d41363/controller-manager/0.log" Feb 27 07:51:46 crc kubenswrapper[4612]: I0227 07:51:46.415909 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-sxt8c" Feb 27 07:51:46 crc kubenswrapper[4612]: I0227 07:51:46.489570 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/738d2b6e-2b7c-461d-b51e-f35eb5d41363-serving-cert\") pod \"738d2b6e-2b7c-461d-b51e-f35eb5d41363\" (UID: \"738d2b6e-2b7c-461d-b51e-f35eb5d41363\") " Feb 27 07:51:46 crc kubenswrapper[4612]: I0227 07:51:46.489614 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hp9f5\" (UniqueName: \"kubernetes.io/projected/738d2b6e-2b7c-461d-b51e-f35eb5d41363-kube-api-access-hp9f5\") pod \"738d2b6e-2b7c-461d-b51e-f35eb5d41363\" (UID: \"738d2b6e-2b7c-461d-b51e-f35eb5d41363\") " Feb 27 07:51:46 crc kubenswrapper[4612]: I0227 07:51:46.489762 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/738d2b6e-2b7c-461d-b51e-f35eb5d41363-proxy-ca-bundles\") pod \"738d2b6e-2b7c-461d-b51e-f35eb5d41363\" (UID: \"738d2b6e-2b7c-461d-b51e-f35eb5d41363\") " Feb 27 07:51:46 crc kubenswrapper[4612]: I0227 07:51:46.489788 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/738d2b6e-2b7c-461d-b51e-f35eb5d41363-client-ca\") pod \"738d2b6e-2b7c-461d-b51e-f35eb5d41363\" (UID: \"738d2b6e-2b7c-461d-b51e-f35eb5d41363\") " Feb 27 07:51:46 crc kubenswrapper[4612]: I0227 07:51:46.489805 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/738d2b6e-2b7c-461d-b51e-f35eb5d41363-config\") pod \"738d2b6e-2b7c-461d-b51e-f35eb5d41363\" (UID: \"738d2b6e-2b7c-461d-b51e-f35eb5d41363\") " Feb 27 07:51:46 crc kubenswrapper[4612]: I0227 07:51:46.490524 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/738d2b6e-2b7c-461d-b51e-f35eb5d41363-config" (OuterVolumeSpecName: "config") pod "738d2b6e-2b7c-461d-b51e-f35eb5d41363" (UID: "738d2b6e-2b7c-461d-b51e-f35eb5d41363"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 07:51:46 crc kubenswrapper[4612]: I0227 07:51:46.490745 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/738d2b6e-2b7c-461d-b51e-f35eb5d41363-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "738d2b6e-2b7c-461d-b51e-f35eb5d41363" (UID: "738d2b6e-2b7c-461d-b51e-f35eb5d41363"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 07:51:46 crc kubenswrapper[4612]: I0227 07:51:46.490795 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/738d2b6e-2b7c-461d-b51e-f35eb5d41363-client-ca" (OuterVolumeSpecName: "client-ca") pod "738d2b6e-2b7c-461d-b51e-f35eb5d41363" (UID: "738d2b6e-2b7c-461d-b51e-f35eb5d41363"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 07:51:46 crc kubenswrapper[4612]: I0227 07:51:46.494103 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/738d2b6e-2b7c-461d-b51e-f35eb5d41363-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "738d2b6e-2b7c-461d-b51e-f35eb5d41363" (UID: "738d2b6e-2b7c-461d-b51e-f35eb5d41363"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 07:51:46 crc kubenswrapper[4612]: I0227 07:51:46.494193 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/738d2b6e-2b7c-461d-b51e-f35eb5d41363-kube-api-access-hp9f5" (OuterVolumeSpecName: "kube-api-access-hp9f5") pod "738d2b6e-2b7c-461d-b51e-f35eb5d41363" (UID: "738d2b6e-2b7c-461d-b51e-f35eb5d41363"). InnerVolumeSpecName "kube-api-access-hp9f5". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 07:51:46 crc kubenswrapper[4612]: I0227 07:51:46.505150 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-8455cd459b-d6dhv"] Feb 27 07:51:46 crc kubenswrapper[4612]: W0227 07:51:46.513937 4612 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode4a7c189_8cdc_4970_8bef_e70742f2fa7e.slice/crio-2a85865f086d97793279adef728242392fd3d46bda4db2dcc3bad55aeb116a1e WatchSource:0}: Error finding container 2a85865f086d97793279adef728242392fd3d46bda4db2dcc3bad55aeb116a1e: Status 404 returned error can't find the container with id 2a85865f086d97793279adef728242392fd3d46bda4db2dcc3bad55aeb116a1e Feb 27 07:51:46 crc kubenswrapper[4612]: I0227 07:51:46.525646 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-8455cd459b-d6dhv" event={"ID":"e4a7c189-8cdc-4970-8bef-e70742f2fa7e","Type":"ContainerStarted","Data":"2a85865f086d97793279adef728242392fd3d46bda4db2dcc3bad55aeb116a1e"} Feb 27 07:51:46 crc kubenswrapper[4612]: I0227 07:51:46.528904 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_cni-sysctl-allowlist-ds-v589k_2522fcbd-fe8e-46cf-95ba-620316513df1/kube-multus-additional-cni-plugins/0.log" Feb 27 07:51:46 crc kubenswrapper[4612]: I0227 07:51:46.529078 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-multus/cni-sysctl-allowlist-ds-v589k" Feb 27 07:51:46 crc kubenswrapper[4612]: I0227 07:51:46.529793 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/cni-sysctl-allowlist-ds-v589k" event={"ID":"2522fcbd-fe8e-46cf-95ba-620316513df1","Type":"ContainerDied","Data":"baf8c9a7cabecfe6817cfdfd39ba317bbbf806b6e12b1d396640418d85825448"} Feb 27 07:51:46 crc kubenswrapper[4612]: I0227 07:51:46.529857 4612 scope.go:117] "RemoveContainer" containerID="19989c4604cb1985b76a2c1d2f8276aa3c4c5fb242e7ebf2619f837c570f9715" Feb 27 07:51:46 crc kubenswrapper[4612]: I0227 07:51:46.533088 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-controller-manager_controller-manager-879f6c89f-sxt8c_738d2b6e-2b7c-461d-b51e-f35eb5d41363/controller-manager/0.log" Feb 27 07:51:46 crc kubenswrapper[4612]: I0227 07:51:46.533191 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-sxt8c" event={"ID":"738d2b6e-2b7c-461d-b51e-f35eb5d41363","Type":"ContainerDied","Data":"ffadd5b41c89bb82f7b5c02114da43182b61205b4abd163256d5cbee5663e237"} Feb 27 07:51:46 crc kubenswrapper[4612]: I0227 07:51:46.533278 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-sxt8c" Feb 27 07:51:46 crc kubenswrapper[4612]: I0227 07:51:46.534727 4612 generic.go:334] "Generic (PLEG): container finished" podID="3d35510b-1105-415a-b866-3c3fd63a646b" containerID="b246e89ac50d94a23e23872ec9fe20d3f113cf379791aeaa2921f51441508fb3" exitCode=0 Feb 27 07:51:46 crc kubenswrapper[4612]: I0227 07:51:46.535794 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-trzgg" event={"ID":"3d35510b-1105-415a-b866-3c3fd63a646b","Type":"ContainerDied","Data":"b246e89ac50d94a23e23872ec9fe20d3f113cf379791aeaa2921f51441508fb3"} Feb 27 07:51:46 crc kubenswrapper[4612]: E0227 07:51:46.538486 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-w4brj" podUID="0bed0a85-ad8c-4670-8193-0e9a90e88d78" Feb 27 07:51:46 crc kubenswrapper[4612]: E0227 07:51:46.538568 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-p5fp8" podUID="201de829-c0b7-4d50-9a3e-ef65d3c06916" Feb 27 07:51:46 crc kubenswrapper[4612]: E0227 07:51:46.538611 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-qlhr9" podUID="7faf5006-23b1-4ef4-ad29-e0e676340a7c" Feb 27 07:51:46 crc kubenswrapper[4612]: E0227 07:51:46.538647 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-fjg45" podUID="7676c2d7-9df2-4dc8-a2bd-cde398845d26" Feb 27 07:51:46 crc kubenswrapper[4612]: E0227 07:51:46.538680 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-qcsjx" podUID="f167bfef-3a04-4363-aa30-7f75a88a35e9" Feb 27 07:51:46 crc kubenswrapper[4612]: I0227 07:51:46.547318 4612 scope.go:117] "RemoveContainer" containerID="80058f3d8b6825a6aa02bba10ea42f4f51630b27d2fc39c813f4cc5bb5412319" Feb 27 07:51:46 crc kubenswrapper[4612]: I0227 07:51:46.591340 4612 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/738d2b6e-2b7c-461d-b51e-f35eb5d41363-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Feb 27 07:51:46 crc kubenswrapper[4612]: I0227 07:51:46.591372 4612 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/738d2b6e-2b7c-461d-b51e-f35eb5d41363-client-ca\") on node \"crc\" DevicePath \"\"" Feb 27 07:51:46 crc kubenswrapper[4612]: I0227 07:51:46.591383 4612 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/738d2b6e-2b7c-461d-b51e-f35eb5d41363-config\") on node \"crc\" DevicePath \"\"" Feb 27 07:51:46 crc kubenswrapper[4612]: I0227 07:51:46.591391 4612 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/738d2b6e-2b7c-461d-b51e-f35eb5d41363-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 27 07:51:46 crc kubenswrapper[4612]: I0227 07:51:46.591399 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hp9f5\" (UniqueName: \"kubernetes.io/projected/738d2b6e-2b7c-461d-b51e-f35eb5d41363-kube-api-access-hp9f5\") on node \"crc\" DevicePath \"\"" Feb 27 07:51:46 crc kubenswrapper[4612]: I0227 07:51:46.596823 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-sxt8c"] Feb 27 07:51:46 crc kubenswrapper[4612]: I0227 07:51:46.599852 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-sxt8c"] Feb 27 07:51:46 crc kubenswrapper[4612]: I0227 07:51:46.641088 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-multus/cni-sysctl-allowlist-ds-v589k"] Feb 27 07:51:46 crc kubenswrapper[4612]: I0227 07:51:46.645364 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-multus/cni-sysctl-allowlist-ds-v589k"] Feb 27 07:51:46 crc kubenswrapper[4612]: I0227 07:51:46.693771 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Feb 27 07:51:46 crc kubenswrapper[4612]: E0227 07:51:46.694013 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2522fcbd-fe8e-46cf-95ba-620316513df1" containerName="kube-multus-additional-cni-plugins" Feb 27 07:51:46 crc kubenswrapper[4612]: I0227 07:51:46.694028 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="2522fcbd-fe8e-46cf-95ba-620316513df1" containerName="kube-multus-additional-cni-plugins" Feb 27 07:51:46 crc kubenswrapper[4612]: E0227 07:51:46.694049 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="738d2b6e-2b7c-461d-b51e-f35eb5d41363" containerName="controller-manager" Feb 27 07:51:46 crc kubenswrapper[4612]: I0227 07:51:46.694056 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="738d2b6e-2b7c-461d-b51e-f35eb5d41363" containerName="controller-manager" Feb 27 07:51:46 crc kubenswrapper[4612]: I0227 07:51:46.694191 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="738d2b6e-2b7c-461d-b51e-f35eb5d41363" containerName="controller-manager" Feb 27 07:51:46 crc kubenswrapper[4612]: I0227 07:51:46.694215 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="2522fcbd-fe8e-46cf-95ba-620316513df1" containerName="kube-multus-additional-cni-plugins" Feb 27 07:51:46 crc kubenswrapper[4612]: I0227 07:51:46.694622 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 27 07:51:46 crc kubenswrapper[4612]: I0227 07:51:46.696868 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Feb 27 07:51:46 crc kubenswrapper[4612]: I0227 07:51:46.697025 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Feb 27 07:51:46 crc kubenswrapper[4612]: I0227 07:51:46.706653 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Feb 27 07:51:46 crc kubenswrapper[4612]: I0227 07:51:46.794058 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a00e512e-69e7-441f-93ca-028ef6c14bec-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"a00e512e-69e7-441f-93ca-028ef6c14bec\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 27 07:51:46 crc kubenswrapper[4612]: I0227 07:51:46.794151 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a00e512e-69e7-441f-93ca-028ef6c14bec-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"a00e512e-69e7-441f-93ca-028ef6c14bec\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 27 07:51:46 crc kubenswrapper[4612]: I0227 07:51:46.871106 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2522fcbd-fe8e-46cf-95ba-620316513df1" path="/var/lib/kubelet/pods/2522fcbd-fe8e-46cf-95ba-620316513df1/volumes" Feb 27 07:51:46 crc kubenswrapper[4612]: I0227 07:51:46.872039 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="738d2b6e-2b7c-461d-b51e-f35eb5d41363" path="/var/lib/kubelet/pods/738d2b6e-2b7c-461d-b51e-f35eb5d41363/volumes" Feb 27 07:51:46 crc kubenswrapper[4612]: I0227 07:51:46.895284 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a00e512e-69e7-441f-93ca-028ef6c14bec-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"a00e512e-69e7-441f-93ca-028ef6c14bec\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 27 07:51:46 crc kubenswrapper[4612]: I0227 07:51:46.895781 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a00e512e-69e7-441f-93ca-028ef6c14bec-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"a00e512e-69e7-441f-93ca-028ef6c14bec\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 27 07:51:46 crc kubenswrapper[4612]: I0227 07:51:46.895899 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a00e512e-69e7-441f-93ca-028ef6c14bec-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"a00e512e-69e7-441f-93ca-028ef6c14bec\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 27 07:51:46 crc kubenswrapper[4612]: I0227 07:51:46.916627 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a00e512e-69e7-441f-93ca-028ef6c14bec-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"a00e512e-69e7-441f-93ca-028ef6c14bec\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 27 07:51:47 crc kubenswrapper[4612]: I0227 07:51:47.014102 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 27 07:51:47 crc kubenswrapper[4612]: I0227 07:51:47.268589 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Feb 27 07:51:47 crc kubenswrapper[4612]: I0227 07:51:47.542707 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-trzgg" event={"ID":"3d35510b-1105-415a-b866-3c3fd63a646b","Type":"ContainerStarted","Data":"31441dacb6e0cd39876cc66c5fabf128dd6cabe206b87c0af0b9e46aed2c2d53"} Feb 27 07:51:47 crc kubenswrapper[4612]: I0227 07:51:47.544646 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-8455cd459b-d6dhv" event={"ID":"e4a7c189-8cdc-4970-8bef-e70742f2fa7e","Type":"ContainerStarted","Data":"5161dcb3ef00c65c12ec430cddf23f59b5651948eb0149291f4f1c851be51a88"} Feb 27 07:51:47 crc kubenswrapper[4612]: I0227 07:51:47.544760 4612 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-8455cd459b-d6dhv" podUID="e4a7c189-8cdc-4970-8bef-e70742f2fa7e" containerName="route-controller-manager" containerID="cri-o://5161dcb3ef00c65c12ec430cddf23f59b5651948eb0149291f4f1c851be51a88" gracePeriod=30 Feb 27 07:51:47 crc kubenswrapper[4612]: I0227 07:51:47.545074 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-8455cd459b-d6dhv" Feb 27 07:51:47 crc kubenswrapper[4612]: I0227 07:51:47.551885 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-8455cd459b-d6dhv" Feb 27 07:51:47 crc kubenswrapper[4612]: I0227 07:51:47.555307 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"a00e512e-69e7-441f-93ca-028ef6c14bec","Type":"ContainerStarted","Data":"68c0c968b46355f90d00c276194e20f69e31e6d327d19eb761aa2cb762a21ebf"} Feb 27 07:51:47 crc kubenswrapper[4612]: I0227 07:51:47.555348 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"a00e512e-69e7-441f-93ca-028ef6c14bec","Type":"ContainerStarted","Data":"d6a3331e707fab87e1c3ba00cb1ccc8d926e52553b30072b9c2d8122c2687127"} Feb 27 07:51:47 crc kubenswrapper[4612]: I0227 07:51:47.563143 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-trzgg" podStartSLOduration=2.5182328309999997 podStartE2EDuration="35.563125956s" podCreationTimestamp="2026-02-27 07:51:12 +0000 UTC" firstStartedPulling="2026-02-27 07:51:13.986015515 +0000 UTC m=+131.839945513" lastFinishedPulling="2026-02-27 07:51:47.03090864 +0000 UTC m=+164.884838638" observedRunningTime="2026-02-27 07:51:47.561426969 +0000 UTC m=+165.415356967" watchObservedRunningTime="2026-02-27 07:51:47.563125956 +0000 UTC m=+165.417055964" Feb 27 07:51:47 crc kubenswrapper[4612]: I0227 07:51:47.581526 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-8455cd459b-d6dhv" podStartSLOduration=25.581504983 podStartE2EDuration="25.581504983s" podCreationTimestamp="2026-02-27 07:51:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 07:51:47.580744342 +0000 UTC m=+165.434674360" watchObservedRunningTime="2026-02-27 07:51:47.581504983 +0000 UTC m=+165.435434981" Feb 27 07:51:47 crc kubenswrapper[4612]: I0227 07:51:47.601531 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-9-crc" podStartSLOduration=1.601510564 podStartE2EDuration="1.601510564s" podCreationTimestamp="2026-02-27 07:51:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 07:51:47.599857298 +0000 UTC m=+165.453787306" watchObservedRunningTime="2026-02-27 07:51:47.601510564 +0000 UTC m=+165.455440562" Feb 27 07:51:47 crc kubenswrapper[4612]: I0227 07:51:47.897046 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-8455cd459b-d6dhv" Feb 27 07:51:47 crc kubenswrapper[4612]: I0227 07:51:47.913121 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7thm5\" (UniqueName: \"kubernetes.io/projected/e4a7c189-8cdc-4970-8bef-e70742f2fa7e-kube-api-access-7thm5\") pod \"e4a7c189-8cdc-4970-8bef-e70742f2fa7e\" (UID: \"e4a7c189-8cdc-4970-8bef-e70742f2fa7e\") " Feb 27 07:51:47 crc kubenswrapper[4612]: I0227 07:51:47.913176 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e4a7c189-8cdc-4970-8bef-e70742f2fa7e-config\") pod \"e4a7c189-8cdc-4970-8bef-e70742f2fa7e\" (UID: \"e4a7c189-8cdc-4970-8bef-e70742f2fa7e\") " Feb 27 07:51:47 crc kubenswrapper[4612]: I0227 07:51:47.913201 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e4a7c189-8cdc-4970-8bef-e70742f2fa7e-client-ca\") pod \"e4a7c189-8cdc-4970-8bef-e70742f2fa7e\" (UID: \"e4a7c189-8cdc-4970-8bef-e70742f2fa7e\") " Feb 27 07:51:47 crc kubenswrapper[4612]: I0227 07:51:47.913229 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e4a7c189-8cdc-4970-8bef-e70742f2fa7e-serving-cert\") pod \"e4a7c189-8cdc-4970-8bef-e70742f2fa7e\" (UID: \"e4a7c189-8cdc-4970-8bef-e70742f2fa7e\") " Feb 27 07:51:47 crc kubenswrapper[4612]: I0227 07:51:47.914025 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e4a7c189-8cdc-4970-8bef-e70742f2fa7e-config" (OuterVolumeSpecName: "config") pod "e4a7c189-8cdc-4970-8bef-e70742f2fa7e" (UID: "e4a7c189-8cdc-4970-8bef-e70742f2fa7e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 07:51:47 crc kubenswrapper[4612]: I0227 07:51:47.914361 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e4a7c189-8cdc-4970-8bef-e70742f2fa7e-client-ca" (OuterVolumeSpecName: "client-ca") pod "e4a7c189-8cdc-4970-8bef-e70742f2fa7e" (UID: "e4a7c189-8cdc-4970-8bef-e70742f2fa7e"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 07:51:47 crc kubenswrapper[4612]: I0227 07:51:47.921112 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e4a7c189-8cdc-4970-8bef-e70742f2fa7e-kube-api-access-7thm5" (OuterVolumeSpecName: "kube-api-access-7thm5") pod "e4a7c189-8cdc-4970-8bef-e70742f2fa7e" (UID: "e4a7c189-8cdc-4970-8bef-e70742f2fa7e"). InnerVolumeSpecName "kube-api-access-7thm5". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 07:51:47 crc kubenswrapper[4612]: I0227 07:51:47.922897 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e4a7c189-8cdc-4970-8bef-e70742f2fa7e-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e4a7c189-8cdc-4970-8bef-e70742f2fa7e" (UID: "e4a7c189-8cdc-4970-8bef-e70742f2fa7e"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 07:51:48 crc kubenswrapper[4612]: I0227 07:51:48.003609 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 27 07:51:48 crc kubenswrapper[4612]: I0227 07:51:48.014956 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7thm5\" (UniqueName: \"kubernetes.io/projected/e4a7c189-8cdc-4970-8bef-e70742f2fa7e-kube-api-access-7thm5\") on node \"crc\" DevicePath \"\"" Feb 27 07:51:48 crc kubenswrapper[4612]: I0227 07:51:48.014990 4612 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e4a7c189-8cdc-4970-8bef-e70742f2fa7e-config\") on node \"crc\" DevicePath \"\"" Feb 27 07:51:48 crc kubenswrapper[4612]: I0227 07:51:48.015023 4612 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e4a7c189-8cdc-4970-8bef-e70742f2fa7e-client-ca\") on node \"crc\" DevicePath \"\"" Feb 27 07:51:48 crc kubenswrapper[4612]: I0227 07:51:48.015038 4612 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e4a7c189-8cdc-4970-8bef-e70742f2fa7e-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 27 07:51:48 crc kubenswrapper[4612]: I0227 07:51:48.564035 4612 generic.go:334] "Generic (PLEG): container finished" podID="a00e512e-69e7-441f-93ca-028ef6c14bec" containerID="68c0c968b46355f90d00c276194e20f69e31e6d327d19eb761aa2cb762a21ebf" exitCode=0 Feb 27 07:51:48 crc kubenswrapper[4612]: I0227 07:51:48.564081 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"a00e512e-69e7-441f-93ca-028ef6c14bec","Type":"ContainerDied","Data":"68c0c968b46355f90d00c276194e20f69e31e6d327d19eb761aa2cb762a21ebf"} Feb 27 07:51:48 crc kubenswrapper[4612]: I0227 07:51:48.565804 4612 generic.go:334] "Generic (PLEG): container finished" podID="e4a7c189-8cdc-4970-8bef-e70742f2fa7e" containerID="5161dcb3ef00c65c12ec430cddf23f59b5651948eb0149291f4f1c851be51a88" exitCode=0 Feb 27 07:51:48 crc kubenswrapper[4612]: I0227 07:51:48.565866 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-8455cd459b-d6dhv" Feb 27 07:51:48 crc kubenswrapper[4612]: I0227 07:51:48.565881 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-8455cd459b-d6dhv" event={"ID":"e4a7c189-8cdc-4970-8bef-e70742f2fa7e","Type":"ContainerDied","Data":"5161dcb3ef00c65c12ec430cddf23f59b5651948eb0149291f4f1c851be51a88"} Feb 27 07:51:48 crc kubenswrapper[4612]: I0227 07:51:48.566051 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-8455cd459b-d6dhv" event={"ID":"e4a7c189-8cdc-4970-8bef-e70742f2fa7e","Type":"ContainerDied","Data":"2a85865f086d97793279adef728242392fd3d46bda4db2dcc3bad55aeb116a1e"} Feb 27 07:51:48 crc kubenswrapper[4612]: I0227 07:51:48.566072 4612 scope.go:117] "RemoveContainer" containerID="5161dcb3ef00c65c12ec430cddf23f59b5651948eb0149291f4f1c851be51a88" Feb 27 07:51:48 crc kubenswrapper[4612]: I0227 07:51:48.585527 4612 scope.go:117] "RemoveContainer" containerID="5161dcb3ef00c65c12ec430cddf23f59b5651948eb0149291f4f1c851be51a88" Feb 27 07:51:48 crc kubenswrapper[4612]: E0227 07:51:48.586889 4612 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5161dcb3ef00c65c12ec430cddf23f59b5651948eb0149291f4f1c851be51a88\": container with ID starting with 5161dcb3ef00c65c12ec430cddf23f59b5651948eb0149291f4f1c851be51a88 not found: ID does not exist" containerID="5161dcb3ef00c65c12ec430cddf23f59b5651948eb0149291f4f1c851be51a88" Feb 27 07:51:48 crc kubenswrapper[4612]: I0227 07:51:48.587032 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5161dcb3ef00c65c12ec430cddf23f59b5651948eb0149291f4f1c851be51a88"} err="failed to get container status \"5161dcb3ef00c65c12ec430cddf23f59b5651948eb0149291f4f1c851be51a88\": rpc error: code = NotFound desc = could not find container \"5161dcb3ef00c65c12ec430cddf23f59b5651948eb0149291f4f1c851be51a88\": container with ID starting with 5161dcb3ef00c65c12ec430cddf23f59b5651948eb0149291f4f1c851be51a88 not found: ID does not exist" Feb 27 07:51:48 crc kubenswrapper[4612]: I0227 07:51:48.604006 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-8455cd459b-d6dhv"] Feb 27 07:51:48 crc kubenswrapper[4612]: I0227 07:51:48.609421 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-8455cd459b-d6dhv"] Feb 27 07:51:48 crc kubenswrapper[4612]: I0227 07:51:48.692077 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-66487b7dd-r2cds"] Feb 27 07:51:48 crc kubenswrapper[4612]: E0227 07:51:48.692309 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e4a7c189-8cdc-4970-8bef-e70742f2fa7e" containerName="route-controller-manager" Feb 27 07:51:48 crc kubenswrapper[4612]: I0227 07:51:48.692321 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="e4a7c189-8cdc-4970-8bef-e70742f2fa7e" containerName="route-controller-manager" Feb 27 07:51:48 crc kubenswrapper[4612]: I0227 07:51:48.692411 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="e4a7c189-8cdc-4970-8bef-e70742f2fa7e" containerName="route-controller-manager" Feb 27 07:51:48 crc kubenswrapper[4612]: I0227 07:51:48.693297 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-66487b7dd-r2cds" Feb 27 07:51:48 crc kubenswrapper[4612]: I0227 07:51:48.695105 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7d66dbf864-cd8fc"] Feb 27 07:51:48 crc kubenswrapper[4612]: I0227 07:51:48.696211 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7d66dbf864-cd8fc" Feb 27 07:51:48 crc kubenswrapper[4612]: I0227 07:51:48.700254 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Feb 27 07:51:48 crc kubenswrapper[4612]: I0227 07:51:48.700481 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Feb 27 07:51:48 crc kubenswrapper[4612]: I0227 07:51:48.700828 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Feb 27 07:51:48 crc kubenswrapper[4612]: I0227 07:51:48.701126 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Feb 27 07:51:48 crc kubenswrapper[4612]: I0227 07:51:48.701496 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Feb 27 07:51:48 crc kubenswrapper[4612]: I0227 07:51:48.704587 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Feb 27 07:51:48 crc kubenswrapper[4612]: I0227 07:51:48.704707 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Feb 27 07:51:48 crc kubenswrapper[4612]: I0227 07:51:48.704663 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Feb 27 07:51:48 crc kubenswrapper[4612]: I0227 07:51:48.704867 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Feb 27 07:51:48 crc kubenswrapper[4612]: I0227 07:51:48.705208 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Feb 27 07:51:48 crc kubenswrapper[4612]: I0227 07:51:48.705345 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Feb 27 07:51:48 crc kubenswrapper[4612]: I0227 07:51:48.705643 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Feb 27 07:51:48 crc kubenswrapper[4612]: I0227 07:51:48.709484 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Feb 27 07:51:48 crc kubenswrapper[4612]: I0227 07:51:48.712588 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-66487b7dd-r2cds"] Feb 27 07:51:48 crc kubenswrapper[4612]: I0227 07:51:48.716602 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7d66dbf864-cd8fc"] Feb 27 07:51:48 crc kubenswrapper[4612]: I0227 07:51:48.724778 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/cc6862b7-a2ab-4051-9d83-17774bfc5ac7-client-ca\") pod \"route-controller-manager-7d66dbf864-cd8fc\" (UID: \"cc6862b7-a2ab-4051-9d83-17774bfc5ac7\") " pod="openshift-route-controller-manager/route-controller-manager-7d66dbf864-cd8fc" Feb 27 07:51:48 crc kubenswrapper[4612]: I0227 07:51:48.724816 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lstjj\" (UniqueName: \"kubernetes.io/projected/cc6862b7-a2ab-4051-9d83-17774bfc5ac7-kube-api-access-lstjj\") pod \"route-controller-manager-7d66dbf864-cd8fc\" (UID: \"cc6862b7-a2ab-4051-9d83-17774bfc5ac7\") " pod="openshift-route-controller-manager/route-controller-manager-7d66dbf864-cd8fc" Feb 27 07:51:48 crc kubenswrapper[4612]: I0227 07:51:48.724844 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/458167a3-25b3-4cd1-8628-865ad916d3a6-config\") pod \"controller-manager-66487b7dd-r2cds\" (UID: \"458167a3-25b3-4cd1-8628-865ad916d3a6\") " pod="openshift-controller-manager/controller-manager-66487b7dd-r2cds" Feb 27 07:51:48 crc kubenswrapper[4612]: I0227 07:51:48.724860 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/458167a3-25b3-4cd1-8628-865ad916d3a6-serving-cert\") pod \"controller-manager-66487b7dd-r2cds\" (UID: \"458167a3-25b3-4cd1-8628-865ad916d3a6\") " pod="openshift-controller-manager/controller-manager-66487b7dd-r2cds" Feb 27 07:51:48 crc kubenswrapper[4612]: I0227 07:51:48.724884 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cc6862b7-a2ab-4051-9d83-17774bfc5ac7-serving-cert\") pod \"route-controller-manager-7d66dbf864-cd8fc\" (UID: \"cc6862b7-a2ab-4051-9d83-17774bfc5ac7\") " pod="openshift-route-controller-manager/route-controller-manager-7d66dbf864-cd8fc" Feb 27 07:51:48 crc kubenswrapper[4612]: I0227 07:51:48.724913 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4wjzq\" (UniqueName: \"kubernetes.io/projected/458167a3-25b3-4cd1-8628-865ad916d3a6-kube-api-access-4wjzq\") pod \"controller-manager-66487b7dd-r2cds\" (UID: \"458167a3-25b3-4cd1-8628-865ad916d3a6\") " pod="openshift-controller-manager/controller-manager-66487b7dd-r2cds" Feb 27 07:51:48 crc kubenswrapper[4612]: I0227 07:51:48.724933 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/458167a3-25b3-4cd1-8628-865ad916d3a6-client-ca\") pod \"controller-manager-66487b7dd-r2cds\" (UID: \"458167a3-25b3-4cd1-8628-865ad916d3a6\") " pod="openshift-controller-manager/controller-manager-66487b7dd-r2cds" Feb 27 07:51:48 crc kubenswrapper[4612]: I0227 07:51:48.724959 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cc6862b7-a2ab-4051-9d83-17774bfc5ac7-config\") pod \"route-controller-manager-7d66dbf864-cd8fc\" (UID: \"cc6862b7-a2ab-4051-9d83-17774bfc5ac7\") " pod="openshift-route-controller-manager/route-controller-manager-7d66dbf864-cd8fc" Feb 27 07:51:48 crc kubenswrapper[4612]: I0227 07:51:48.724981 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/458167a3-25b3-4cd1-8628-865ad916d3a6-proxy-ca-bundles\") pod \"controller-manager-66487b7dd-r2cds\" (UID: \"458167a3-25b3-4cd1-8628-865ad916d3a6\") " pod="openshift-controller-manager/controller-manager-66487b7dd-r2cds" Feb 27 07:51:48 crc kubenswrapper[4612]: I0227 07:51:48.826129 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/458167a3-25b3-4cd1-8628-865ad916d3a6-config\") pod \"controller-manager-66487b7dd-r2cds\" (UID: \"458167a3-25b3-4cd1-8628-865ad916d3a6\") " pod="openshift-controller-manager/controller-manager-66487b7dd-r2cds" Feb 27 07:51:48 crc kubenswrapper[4612]: I0227 07:51:48.826513 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/458167a3-25b3-4cd1-8628-865ad916d3a6-serving-cert\") pod \"controller-manager-66487b7dd-r2cds\" (UID: \"458167a3-25b3-4cd1-8628-865ad916d3a6\") " pod="openshift-controller-manager/controller-manager-66487b7dd-r2cds" Feb 27 07:51:48 crc kubenswrapper[4612]: I0227 07:51:48.826550 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cc6862b7-a2ab-4051-9d83-17774bfc5ac7-serving-cert\") pod \"route-controller-manager-7d66dbf864-cd8fc\" (UID: \"cc6862b7-a2ab-4051-9d83-17774bfc5ac7\") " pod="openshift-route-controller-manager/route-controller-manager-7d66dbf864-cd8fc" Feb 27 07:51:48 crc kubenswrapper[4612]: I0227 07:51:48.826578 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4wjzq\" (UniqueName: \"kubernetes.io/projected/458167a3-25b3-4cd1-8628-865ad916d3a6-kube-api-access-4wjzq\") pod \"controller-manager-66487b7dd-r2cds\" (UID: \"458167a3-25b3-4cd1-8628-865ad916d3a6\") " pod="openshift-controller-manager/controller-manager-66487b7dd-r2cds" Feb 27 07:51:48 crc kubenswrapper[4612]: I0227 07:51:48.826601 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/458167a3-25b3-4cd1-8628-865ad916d3a6-client-ca\") pod \"controller-manager-66487b7dd-r2cds\" (UID: \"458167a3-25b3-4cd1-8628-865ad916d3a6\") " pod="openshift-controller-manager/controller-manager-66487b7dd-r2cds" Feb 27 07:51:48 crc kubenswrapper[4612]: I0227 07:51:48.826629 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cc6862b7-a2ab-4051-9d83-17774bfc5ac7-config\") pod \"route-controller-manager-7d66dbf864-cd8fc\" (UID: \"cc6862b7-a2ab-4051-9d83-17774bfc5ac7\") " pod="openshift-route-controller-manager/route-controller-manager-7d66dbf864-cd8fc" Feb 27 07:51:48 crc kubenswrapper[4612]: I0227 07:51:48.826646 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/458167a3-25b3-4cd1-8628-865ad916d3a6-proxy-ca-bundles\") pod \"controller-manager-66487b7dd-r2cds\" (UID: \"458167a3-25b3-4cd1-8628-865ad916d3a6\") " pod="openshift-controller-manager/controller-manager-66487b7dd-r2cds" Feb 27 07:51:48 crc kubenswrapper[4612]: I0227 07:51:48.826705 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/cc6862b7-a2ab-4051-9d83-17774bfc5ac7-client-ca\") pod \"route-controller-manager-7d66dbf864-cd8fc\" (UID: \"cc6862b7-a2ab-4051-9d83-17774bfc5ac7\") " pod="openshift-route-controller-manager/route-controller-manager-7d66dbf864-cd8fc" Feb 27 07:51:48 crc kubenswrapper[4612]: I0227 07:51:48.826734 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lstjj\" (UniqueName: \"kubernetes.io/projected/cc6862b7-a2ab-4051-9d83-17774bfc5ac7-kube-api-access-lstjj\") pod \"route-controller-manager-7d66dbf864-cd8fc\" (UID: \"cc6862b7-a2ab-4051-9d83-17774bfc5ac7\") " pod="openshift-route-controller-manager/route-controller-manager-7d66dbf864-cd8fc" Feb 27 07:51:48 crc kubenswrapper[4612]: I0227 07:51:48.829905 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/458167a3-25b3-4cd1-8628-865ad916d3a6-proxy-ca-bundles\") pod \"controller-manager-66487b7dd-r2cds\" (UID: \"458167a3-25b3-4cd1-8628-865ad916d3a6\") " pod="openshift-controller-manager/controller-manager-66487b7dd-r2cds" Feb 27 07:51:48 crc kubenswrapper[4612]: I0227 07:51:48.830639 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/458167a3-25b3-4cd1-8628-865ad916d3a6-client-ca\") pod \"controller-manager-66487b7dd-r2cds\" (UID: \"458167a3-25b3-4cd1-8628-865ad916d3a6\") " pod="openshift-controller-manager/controller-manager-66487b7dd-r2cds" Feb 27 07:51:48 crc kubenswrapper[4612]: I0227 07:51:48.830782 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cc6862b7-a2ab-4051-9d83-17774bfc5ac7-config\") pod \"route-controller-manager-7d66dbf864-cd8fc\" (UID: \"cc6862b7-a2ab-4051-9d83-17774bfc5ac7\") " pod="openshift-route-controller-manager/route-controller-manager-7d66dbf864-cd8fc" Feb 27 07:51:48 crc kubenswrapper[4612]: I0227 07:51:48.831202 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/cc6862b7-a2ab-4051-9d83-17774bfc5ac7-client-ca\") pod \"route-controller-manager-7d66dbf864-cd8fc\" (UID: \"cc6862b7-a2ab-4051-9d83-17774bfc5ac7\") " pod="openshift-route-controller-manager/route-controller-manager-7d66dbf864-cd8fc" Feb 27 07:51:48 crc kubenswrapper[4612]: I0227 07:51:48.833913 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/458167a3-25b3-4cd1-8628-865ad916d3a6-serving-cert\") pod \"controller-manager-66487b7dd-r2cds\" (UID: \"458167a3-25b3-4cd1-8628-865ad916d3a6\") " pod="openshift-controller-manager/controller-manager-66487b7dd-r2cds" Feb 27 07:51:48 crc kubenswrapper[4612]: I0227 07:51:48.837089 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/458167a3-25b3-4cd1-8628-865ad916d3a6-config\") pod \"controller-manager-66487b7dd-r2cds\" (UID: \"458167a3-25b3-4cd1-8628-865ad916d3a6\") " pod="openshift-controller-manager/controller-manager-66487b7dd-r2cds" Feb 27 07:51:48 crc kubenswrapper[4612]: I0227 07:51:48.844245 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lstjj\" (UniqueName: \"kubernetes.io/projected/cc6862b7-a2ab-4051-9d83-17774bfc5ac7-kube-api-access-lstjj\") pod \"route-controller-manager-7d66dbf864-cd8fc\" (UID: \"cc6862b7-a2ab-4051-9d83-17774bfc5ac7\") " pod="openshift-route-controller-manager/route-controller-manager-7d66dbf864-cd8fc" Feb 27 07:51:48 crc kubenswrapper[4612]: I0227 07:51:48.845021 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cc6862b7-a2ab-4051-9d83-17774bfc5ac7-serving-cert\") pod \"route-controller-manager-7d66dbf864-cd8fc\" (UID: \"cc6862b7-a2ab-4051-9d83-17774bfc5ac7\") " pod="openshift-route-controller-manager/route-controller-manager-7d66dbf864-cd8fc" Feb 27 07:51:48 crc kubenswrapper[4612]: I0227 07:51:48.849153 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4wjzq\" (UniqueName: \"kubernetes.io/projected/458167a3-25b3-4cd1-8628-865ad916d3a6-kube-api-access-4wjzq\") pod \"controller-manager-66487b7dd-r2cds\" (UID: \"458167a3-25b3-4cd1-8628-865ad916d3a6\") " pod="openshift-controller-manager/controller-manager-66487b7dd-r2cds" Feb 27 07:51:48 crc kubenswrapper[4612]: I0227 07:51:48.859273 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e4a7c189-8cdc-4970-8bef-e70742f2fa7e" path="/var/lib/kubelet/pods/e4a7c189-8cdc-4970-8bef-e70742f2fa7e/volumes" Feb 27 07:51:49 crc kubenswrapper[4612]: I0227 07:51:49.015978 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-66487b7dd-r2cds" Feb 27 07:51:49 crc kubenswrapper[4612]: I0227 07:51:49.022287 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7d66dbf864-cd8fc" Feb 27 07:51:49 crc kubenswrapper[4612]: I0227 07:51:49.286023 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7d66dbf864-cd8fc"] Feb 27 07:51:49 crc kubenswrapper[4612]: I0227 07:51:49.453721 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-66487b7dd-r2cds"] Feb 27 07:51:49 crc kubenswrapper[4612]: W0227 07:51:49.458677 4612 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod458167a3_25b3_4cd1_8628_865ad916d3a6.slice/crio-47ac8b300890eb2dbe6d5ee7578f5f522e8820a79526ada246c4cb452b44b1ad WatchSource:0}: Error finding container 47ac8b300890eb2dbe6d5ee7578f5f522e8820a79526ada246c4cb452b44b1ad: Status 404 returned error can't find the container with id 47ac8b300890eb2dbe6d5ee7578f5f522e8820a79526ada246c4cb452b44b1ad Feb 27 07:51:49 crc kubenswrapper[4612]: I0227 07:51:49.579299 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7d66dbf864-cd8fc" event={"ID":"cc6862b7-a2ab-4051-9d83-17774bfc5ac7","Type":"ContainerStarted","Data":"80134093f8f5fbc54255812e2817474d95b0563912dc50683b7a6a006cc7eecf"} Feb 27 07:51:49 crc kubenswrapper[4612]: I0227 07:51:49.579579 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-7d66dbf864-cd8fc" Feb 27 07:51:49 crc kubenswrapper[4612]: I0227 07:51:49.579591 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7d66dbf864-cd8fc" event={"ID":"cc6862b7-a2ab-4051-9d83-17774bfc5ac7","Type":"ContainerStarted","Data":"be2d1a4f1ec68403bc6af09057d10b297350c07592ffd1cd2d6dd0a8a713522a"} Feb 27 07:51:49 crc kubenswrapper[4612]: I0227 07:51:49.583563 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-66487b7dd-r2cds" event={"ID":"458167a3-25b3-4cd1-8628-865ad916d3a6","Type":"ContainerStarted","Data":"47ac8b300890eb2dbe6d5ee7578f5f522e8820a79526ada246c4cb452b44b1ad"} Feb 27 07:51:49 crc kubenswrapper[4612]: I0227 07:51:49.583969 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-66487b7dd-r2cds" Feb 27 07:51:49 crc kubenswrapper[4612]: I0227 07:51:49.585870 4612 patch_prober.go:28] interesting pod/controller-manager-66487b7dd-r2cds container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.57:8443/healthz\": dial tcp 10.217.0.57:8443: connect: connection refused" start-of-body= Feb 27 07:51:49 crc kubenswrapper[4612]: I0227 07:51:49.585915 4612 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-66487b7dd-r2cds" podUID="458167a3-25b3-4cd1-8628-865ad916d3a6" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.57:8443/healthz\": dial tcp 10.217.0.57:8443: connect: connection refused" Feb 27 07:51:49 crc kubenswrapper[4612]: I0227 07:51:49.601527 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-7d66dbf864-cd8fc" podStartSLOduration=7.601497956 podStartE2EDuration="7.601497956s" podCreationTimestamp="2026-02-27 07:51:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 07:51:49.596788117 +0000 UTC m=+167.450718115" watchObservedRunningTime="2026-02-27 07:51:49.601497956 +0000 UTC m=+167.455427954" Feb 27 07:51:49 crc kubenswrapper[4612]: I0227 07:51:49.627378 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-66487b7dd-r2cds" podStartSLOduration=7.627357579 podStartE2EDuration="7.627357579s" podCreationTimestamp="2026-02-27 07:51:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 07:51:49.624735657 +0000 UTC m=+167.478665655" watchObservedRunningTime="2026-02-27 07:51:49.627357579 +0000 UTC m=+167.481287577" Feb 27 07:51:49 crc kubenswrapper[4612]: I0227 07:51:49.790307 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 27 07:51:49 crc kubenswrapper[4612]: I0227 07:51:49.839472 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a00e512e-69e7-441f-93ca-028ef6c14bec-kubelet-dir\") pod \"a00e512e-69e7-441f-93ca-028ef6c14bec\" (UID: \"a00e512e-69e7-441f-93ca-028ef6c14bec\") " Feb 27 07:51:49 crc kubenswrapper[4612]: I0227 07:51:49.839665 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a00e512e-69e7-441f-93ca-028ef6c14bec-kube-api-access\") pod \"a00e512e-69e7-441f-93ca-028ef6c14bec\" (UID: \"a00e512e-69e7-441f-93ca-028ef6c14bec\") " Feb 27 07:51:49 crc kubenswrapper[4612]: I0227 07:51:49.840644 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a00e512e-69e7-441f-93ca-028ef6c14bec-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "a00e512e-69e7-441f-93ca-028ef6c14bec" (UID: "a00e512e-69e7-441f-93ca-028ef6c14bec"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 27 07:51:49 crc kubenswrapper[4612]: I0227 07:51:49.844339 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a00e512e-69e7-441f-93ca-028ef6c14bec-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "a00e512e-69e7-441f-93ca-028ef6c14bec" (UID: "a00e512e-69e7-441f-93ca-028ef6c14bec"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 07:51:49 crc kubenswrapper[4612]: I0227 07:51:49.941074 4612 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a00e512e-69e7-441f-93ca-028ef6c14bec-kubelet-dir\") on node \"crc\" DevicePath \"\"" Feb 27 07:51:49 crc kubenswrapper[4612]: I0227 07:51:49.941391 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a00e512e-69e7-441f-93ca-028ef6c14bec-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 27 07:51:49 crc kubenswrapper[4612]: I0227 07:51:49.990833 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-7d66dbf864-cd8fc" Feb 27 07:51:50 crc kubenswrapper[4612]: I0227 07:51:50.589457 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"a00e512e-69e7-441f-93ca-028ef6c14bec","Type":"ContainerDied","Data":"d6a3331e707fab87e1c3ba00cb1ccc8d926e52553b30072b9c2d8122c2687127"} Feb 27 07:51:50 crc kubenswrapper[4612]: I0227 07:51:50.590752 4612 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d6a3331e707fab87e1c3ba00cb1ccc8d926e52553b30072b9c2d8122c2687127" Feb 27 07:51:50 crc kubenswrapper[4612]: I0227 07:51:50.589492 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 27 07:51:50 crc kubenswrapper[4612]: I0227 07:51:50.592029 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-66487b7dd-r2cds" event={"ID":"458167a3-25b3-4cd1-8628-865ad916d3a6","Type":"ContainerStarted","Data":"09c3364c1066ea4c2143584d4fb76486d124f024e1ebdfd3b2362d9396abe8c9"} Feb 27 07:51:50 crc kubenswrapper[4612]: I0227 07:51:50.596497 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-66487b7dd-r2cds" Feb 27 07:51:51 crc kubenswrapper[4612]: I0227 07:51:51.117892 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-nkbg2"] Feb 27 07:51:52 crc kubenswrapper[4612]: I0227 07:51:52.409020 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-trzgg" Feb 27 07:51:52 crc kubenswrapper[4612]: I0227 07:51:52.409303 4612 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-trzgg" Feb 27 07:51:52 crc kubenswrapper[4612]: I0227 07:51:52.590483 4612 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-trzgg" Feb 27 07:51:52 crc kubenswrapper[4612]: I0227 07:51:52.676912 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-trzgg" Feb 27 07:51:53 crc kubenswrapper[4612]: I0227 07:51:53.308741 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Feb 27 07:51:53 crc kubenswrapper[4612]: E0227 07:51:53.308936 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a00e512e-69e7-441f-93ca-028ef6c14bec" containerName="pruner" Feb 27 07:51:53 crc kubenswrapper[4612]: I0227 07:51:53.308948 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="a00e512e-69e7-441f-93ca-028ef6c14bec" containerName="pruner" Feb 27 07:51:53 crc kubenswrapper[4612]: I0227 07:51:53.309036 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="a00e512e-69e7-441f-93ca-028ef6c14bec" containerName="pruner" Feb 27 07:51:53 crc kubenswrapper[4612]: I0227 07:51:53.309403 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Feb 27 07:51:53 crc kubenswrapper[4612]: I0227 07:51:53.309774 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Feb 27 07:51:53 crc kubenswrapper[4612]: I0227 07:51:53.311346 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Feb 27 07:51:53 crc kubenswrapper[4612]: I0227 07:51:53.311417 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Feb 27 07:51:53 crc kubenswrapper[4612]: I0227 07:51:53.487113 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/c13ed974-66aa-467d-9d11-4b375e53b8bc-kubelet-dir\") pod \"installer-9-crc\" (UID: \"c13ed974-66aa-467d-9d11-4b375e53b8bc\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 27 07:51:53 crc kubenswrapper[4612]: I0227 07:51:53.487158 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c13ed974-66aa-467d-9d11-4b375e53b8bc-kube-api-access\") pod \"installer-9-crc\" (UID: \"c13ed974-66aa-467d-9d11-4b375e53b8bc\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 27 07:51:53 crc kubenswrapper[4612]: I0227 07:51:53.487221 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/c13ed974-66aa-467d-9d11-4b375e53b8bc-var-lock\") pod \"installer-9-crc\" (UID: \"c13ed974-66aa-467d-9d11-4b375e53b8bc\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 27 07:51:53 crc kubenswrapper[4612]: I0227 07:51:53.588923 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c13ed974-66aa-467d-9d11-4b375e53b8bc-kube-api-access\") pod \"installer-9-crc\" (UID: \"c13ed974-66aa-467d-9d11-4b375e53b8bc\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 27 07:51:53 crc kubenswrapper[4612]: I0227 07:51:53.589052 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/c13ed974-66aa-467d-9d11-4b375e53b8bc-var-lock\") pod \"installer-9-crc\" (UID: \"c13ed974-66aa-467d-9d11-4b375e53b8bc\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 27 07:51:53 crc kubenswrapper[4612]: I0227 07:51:53.589088 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/c13ed974-66aa-467d-9d11-4b375e53b8bc-kubelet-dir\") pod \"installer-9-crc\" (UID: \"c13ed974-66aa-467d-9d11-4b375e53b8bc\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 27 07:51:53 crc kubenswrapper[4612]: I0227 07:51:53.589166 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/c13ed974-66aa-467d-9d11-4b375e53b8bc-kubelet-dir\") pod \"installer-9-crc\" (UID: \"c13ed974-66aa-467d-9d11-4b375e53b8bc\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 27 07:51:53 crc kubenswrapper[4612]: I0227 07:51:53.589208 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/c13ed974-66aa-467d-9d11-4b375e53b8bc-var-lock\") pod \"installer-9-crc\" (UID: \"c13ed974-66aa-467d-9d11-4b375e53b8bc\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 27 07:51:53 crc kubenswrapper[4612]: I0227 07:51:53.607426 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c13ed974-66aa-467d-9d11-4b375e53b8bc-kube-api-access\") pod \"installer-9-crc\" (UID: \"c13ed974-66aa-467d-9d11-4b375e53b8bc\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 27 07:51:53 crc kubenswrapper[4612]: I0227 07:51:53.635440 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Feb 27 07:51:53 crc kubenswrapper[4612]: I0227 07:51:53.766557 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-trzgg"] Feb 27 07:51:53 crc kubenswrapper[4612]: I0227 07:51:53.821019 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Feb 27 07:51:54 crc kubenswrapper[4612]: I0227 07:51:54.612239 4612 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-trzgg" podUID="3d35510b-1105-415a-b866-3c3fd63a646b" containerName="registry-server" containerID="cri-o://31441dacb6e0cd39876cc66c5fabf128dd6cabe206b87c0af0b9e46aed2c2d53" gracePeriod=2 Feb 27 07:51:54 crc kubenswrapper[4612]: I0227 07:51:54.612911 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"c13ed974-66aa-467d-9d11-4b375e53b8bc","Type":"ContainerStarted","Data":"5e4fd3279c97258bdd7509bed769c9e8ec75b544f388b59f8ed8b6f42e19ad71"} Feb 27 07:51:54 crc kubenswrapper[4612]: I0227 07:51:54.612958 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"c13ed974-66aa-467d-9d11-4b375e53b8bc","Type":"ContainerStarted","Data":"17264508b2043f15cff030f0af3aec3ce4093f5cb6025efd2e53ec84bc7ad125"} Feb 27 07:51:54 crc kubenswrapper[4612]: I0227 07:51:54.640722 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/installer-9-crc" podStartSLOduration=1.64070581 podStartE2EDuration="1.64070581s" podCreationTimestamp="2026-02-27 07:51:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 07:51:54.63706643 +0000 UTC m=+172.490996438" watchObservedRunningTime="2026-02-27 07:51:54.64070581 +0000 UTC m=+172.494635808" Feb 27 07:51:54 crc kubenswrapper[4612]: E0227 07:51:54.758314 4612 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3d35510b_1105_415a_b866_3c3fd63a646b.slice/crio-conmon-31441dacb6e0cd39876cc66c5fabf128dd6cabe206b87c0af0b9e46aed2c2d53.scope\": RecentStats: unable to find data in memory cache]" Feb 27 07:51:54 crc kubenswrapper[4612]: I0227 07:51:54.957057 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-trzgg" Feb 27 07:51:55 crc kubenswrapper[4612]: I0227 07:51:55.106227 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3d35510b-1105-415a-b866-3c3fd63a646b-catalog-content\") pod \"3d35510b-1105-415a-b866-3c3fd63a646b\" (UID: \"3d35510b-1105-415a-b866-3c3fd63a646b\") " Feb 27 07:51:55 crc kubenswrapper[4612]: I0227 07:51:55.106290 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3d35510b-1105-415a-b866-3c3fd63a646b-utilities\") pod \"3d35510b-1105-415a-b866-3c3fd63a646b\" (UID: \"3d35510b-1105-415a-b866-3c3fd63a646b\") " Feb 27 07:51:55 crc kubenswrapper[4612]: I0227 07:51:55.106323 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4kwtg\" (UniqueName: \"kubernetes.io/projected/3d35510b-1105-415a-b866-3c3fd63a646b-kube-api-access-4kwtg\") pod \"3d35510b-1105-415a-b866-3c3fd63a646b\" (UID: \"3d35510b-1105-415a-b866-3c3fd63a646b\") " Feb 27 07:51:55 crc kubenswrapper[4612]: I0227 07:51:55.107483 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3d35510b-1105-415a-b866-3c3fd63a646b-utilities" (OuterVolumeSpecName: "utilities") pod "3d35510b-1105-415a-b866-3c3fd63a646b" (UID: "3d35510b-1105-415a-b866-3c3fd63a646b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 07:51:55 crc kubenswrapper[4612]: I0227 07:51:55.112930 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3d35510b-1105-415a-b866-3c3fd63a646b-kube-api-access-4kwtg" (OuterVolumeSpecName: "kube-api-access-4kwtg") pod "3d35510b-1105-415a-b866-3c3fd63a646b" (UID: "3d35510b-1105-415a-b866-3c3fd63a646b"). InnerVolumeSpecName "kube-api-access-4kwtg". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 07:51:55 crc kubenswrapper[4612]: I0227 07:51:55.133992 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3d35510b-1105-415a-b866-3c3fd63a646b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3d35510b-1105-415a-b866-3c3fd63a646b" (UID: "3d35510b-1105-415a-b866-3c3fd63a646b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 07:51:55 crc kubenswrapper[4612]: I0227 07:51:55.207837 4612 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3d35510b-1105-415a-b866-3c3fd63a646b-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 27 07:51:55 crc kubenswrapper[4612]: I0227 07:51:55.207876 4612 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3d35510b-1105-415a-b866-3c3fd63a646b-utilities\") on node \"crc\" DevicePath \"\"" Feb 27 07:51:55 crc kubenswrapper[4612]: I0227 07:51:55.207892 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4kwtg\" (UniqueName: \"kubernetes.io/projected/3d35510b-1105-415a-b866-3c3fd63a646b-kube-api-access-4kwtg\") on node \"crc\" DevicePath \"\"" Feb 27 07:51:55 crc kubenswrapper[4612]: I0227 07:51:55.618361 4612 generic.go:334] "Generic (PLEG): container finished" podID="3d35510b-1105-415a-b866-3c3fd63a646b" containerID="31441dacb6e0cd39876cc66c5fabf128dd6cabe206b87c0af0b9e46aed2c2d53" exitCode=0 Feb 27 07:51:55 crc kubenswrapper[4612]: I0227 07:51:55.618439 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-trzgg" Feb 27 07:51:55 crc kubenswrapper[4612]: I0227 07:51:55.618643 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-trzgg" event={"ID":"3d35510b-1105-415a-b866-3c3fd63a646b","Type":"ContainerDied","Data":"31441dacb6e0cd39876cc66c5fabf128dd6cabe206b87c0af0b9e46aed2c2d53"} Feb 27 07:51:55 crc kubenswrapper[4612]: I0227 07:51:55.618677 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-trzgg" event={"ID":"3d35510b-1105-415a-b866-3c3fd63a646b","Type":"ContainerDied","Data":"8096dd2568edca1253a3d02cb9bad54dd90684eed31e4a52a6fc8808e0c88293"} Feb 27 07:51:55 crc kubenswrapper[4612]: I0227 07:51:55.618720 4612 scope.go:117] "RemoveContainer" containerID="31441dacb6e0cd39876cc66c5fabf128dd6cabe206b87c0af0b9e46aed2c2d53" Feb 27 07:51:55 crc kubenswrapper[4612]: I0227 07:51:55.638483 4612 scope.go:117] "RemoveContainer" containerID="b246e89ac50d94a23e23872ec9fe20d3f113cf379791aeaa2921f51441508fb3" Feb 27 07:51:55 crc kubenswrapper[4612]: I0227 07:51:55.648273 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-trzgg"] Feb 27 07:51:55 crc kubenswrapper[4612]: I0227 07:51:55.652484 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-trzgg"] Feb 27 07:51:55 crc kubenswrapper[4612]: I0227 07:51:55.665375 4612 scope.go:117] "RemoveContainer" containerID="ec3552c14e5b1f1fcc3cb85c4168c6bcdcf7d462e6c062a92794d15c9bbdcd7f" Feb 27 07:51:55 crc kubenswrapper[4612]: I0227 07:51:55.682388 4612 scope.go:117] "RemoveContainer" containerID="31441dacb6e0cd39876cc66c5fabf128dd6cabe206b87c0af0b9e46aed2c2d53" Feb 27 07:51:55 crc kubenswrapper[4612]: E0227 07:51:55.683017 4612 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"31441dacb6e0cd39876cc66c5fabf128dd6cabe206b87c0af0b9e46aed2c2d53\": container with ID starting with 31441dacb6e0cd39876cc66c5fabf128dd6cabe206b87c0af0b9e46aed2c2d53 not found: ID does not exist" containerID="31441dacb6e0cd39876cc66c5fabf128dd6cabe206b87c0af0b9e46aed2c2d53" Feb 27 07:51:55 crc kubenswrapper[4612]: I0227 07:51:55.683085 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"31441dacb6e0cd39876cc66c5fabf128dd6cabe206b87c0af0b9e46aed2c2d53"} err="failed to get container status \"31441dacb6e0cd39876cc66c5fabf128dd6cabe206b87c0af0b9e46aed2c2d53\": rpc error: code = NotFound desc = could not find container \"31441dacb6e0cd39876cc66c5fabf128dd6cabe206b87c0af0b9e46aed2c2d53\": container with ID starting with 31441dacb6e0cd39876cc66c5fabf128dd6cabe206b87c0af0b9e46aed2c2d53 not found: ID does not exist" Feb 27 07:51:55 crc kubenswrapper[4612]: I0227 07:51:55.683113 4612 scope.go:117] "RemoveContainer" containerID="b246e89ac50d94a23e23872ec9fe20d3f113cf379791aeaa2921f51441508fb3" Feb 27 07:51:55 crc kubenswrapper[4612]: E0227 07:51:55.683547 4612 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b246e89ac50d94a23e23872ec9fe20d3f113cf379791aeaa2921f51441508fb3\": container with ID starting with b246e89ac50d94a23e23872ec9fe20d3f113cf379791aeaa2921f51441508fb3 not found: ID does not exist" containerID="b246e89ac50d94a23e23872ec9fe20d3f113cf379791aeaa2921f51441508fb3" Feb 27 07:51:55 crc kubenswrapper[4612]: I0227 07:51:55.683576 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b246e89ac50d94a23e23872ec9fe20d3f113cf379791aeaa2921f51441508fb3"} err="failed to get container status \"b246e89ac50d94a23e23872ec9fe20d3f113cf379791aeaa2921f51441508fb3\": rpc error: code = NotFound desc = could not find container \"b246e89ac50d94a23e23872ec9fe20d3f113cf379791aeaa2921f51441508fb3\": container with ID starting with b246e89ac50d94a23e23872ec9fe20d3f113cf379791aeaa2921f51441508fb3 not found: ID does not exist" Feb 27 07:51:55 crc kubenswrapper[4612]: I0227 07:51:55.683591 4612 scope.go:117] "RemoveContainer" containerID="ec3552c14e5b1f1fcc3cb85c4168c6bcdcf7d462e6c062a92794d15c9bbdcd7f" Feb 27 07:51:55 crc kubenswrapper[4612]: E0227 07:51:55.683857 4612 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ec3552c14e5b1f1fcc3cb85c4168c6bcdcf7d462e6c062a92794d15c9bbdcd7f\": container with ID starting with ec3552c14e5b1f1fcc3cb85c4168c6bcdcf7d462e6c062a92794d15c9bbdcd7f not found: ID does not exist" containerID="ec3552c14e5b1f1fcc3cb85c4168c6bcdcf7d462e6c062a92794d15c9bbdcd7f" Feb 27 07:51:55 crc kubenswrapper[4612]: I0227 07:51:55.683888 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ec3552c14e5b1f1fcc3cb85c4168c6bcdcf7d462e6c062a92794d15c9bbdcd7f"} err="failed to get container status \"ec3552c14e5b1f1fcc3cb85c4168c6bcdcf7d462e6c062a92794d15c9bbdcd7f\": rpc error: code = NotFound desc = could not find container \"ec3552c14e5b1f1fcc3cb85c4168c6bcdcf7d462e6c062a92794d15c9bbdcd7f\": container with ID starting with ec3552c14e5b1f1fcc3cb85c4168c6bcdcf7d462e6c062a92794d15c9bbdcd7f not found: ID does not exist" Feb 27 07:51:56 crc kubenswrapper[4612]: I0227 07:51:56.861430 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3d35510b-1105-415a-b866-3c3fd63a646b" path="/var/lib/kubelet/pods/3d35510b-1105-415a-b866-3c3fd63a646b/volumes" Feb 27 07:51:58 crc kubenswrapper[4612]: I0227 07:51:58.640108 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-w4brj" event={"ID":"0bed0a85-ad8c-4670-8193-0e9a90e88d78","Type":"ContainerStarted","Data":"db569021060effab133a7d67b77b4bf9c38c1f256f73e7d7ec963c63bd23482d"} Feb 27 07:51:58 crc kubenswrapper[4612]: I0227 07:51:58.641499 4612 generic.go:334] "Generic (PLEG): container finished" podID="b4a05b89-0ef2-4f07-b1de-f558a36415a3" containerID="ebd705cf2e79783dc207f1bd1f410ff3ed40b1b60e3eaa30a6dd2942e0dd3980" exitCode=0 Feb 27 07:51:58 crc kubenswrapper[4612]: I0227 07:51:58.641543 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-99cd5" event={"ID":"b4a05b89-0ef2-4f07-b1de-f558a36415a3","Type":"ContainerDied","Data":"ebd705cf2e79783dc207f1bd1f410ff3ed40b1b60e3eaa30a6dd2942e0dd3980"} Feb 27 07:51:59 crc kubenswrapper[4612]: I0227 07:51:59.650135 4612 generic.go:334] "Generic (PLEG): container finished" podID="201de829-c0b7-4d50-9a3e-ef65d3c06916" containerID="e7d182d8908ed3789a5ec0e42333b53967e23f78c8f66f9498c6a40440374b5a" exitCode=0 Feb 27 07:51:59 crc kubenswrapper[4612]: I0227 07:51:59.650212 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p5fp8" event={"ID":"201de829-c0b7-4d50-9a3e-ef65d3c06916","Type":"ContainerDied","Data":"e7d182d8908ed3789a5ec0e42333b53967e23f78c8f66f9498c6a40440374b5a"} Feb 27 07:51:59 crc kubenswrapper[4612]: I0227 07:51:59.660072 4612 generic.go:334] "Generic (PLEG): container finished" podID="0bed0a85-ad8c-4670-8193-0e9a90e88d78" containerID="db569021060effab133a7d67b77b4bf9c38c1f256f73e7d7ec963c63bd23482d" exitCode=0 Feb 27 07:51:59 crc kubenswrapper[4612]: I0227 07:51:59.660158 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-w4brj" event={"ID":"0bed0a85-ad8c-4670-8193-0e9a90e88d78","Type":"ContainerDied","Data":"db569021060effab133a7d67b77b4bf9c38c1f256f73e7d7ec963c63bd23482d"} Feb 27 07:51:59 crc kubenswrapper[4612]: I0227 07:51:59.664916 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-99cd5" event={"ID":"b4a05b89-0ef2-4f07-b1de-f558a36415a3","Type":"ContainerStarted","Data":"b9fd648455a1f23c8fe4b0657e89196db987d450104c1ad5ea1c2b53f4c669bd"} Feb 27 07:51:59 crc kubenswrapper[4612]: I0227 07:51:59.714887 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-99cd5" podStartSLOduration=3.4723289 podStartE2EDuration="49.714870237s" podCreationTimestamp="2026-02-27 07:51:10 +0000 UTC" firstStartedPulling="2026-02-27 07:51:12.824054266 +0000 UTC m=+130.677984264" lastFinishedPulling="2026-02-27 07:51:59.066595603 +0000 UTC m=+176.920525601" observedRunningTime="2026-02-27 07:51:59.713303144 +0000 UTC m=+177.567233142" watchObservedRunningTime="2026-02-27 07:51:59.714870237 +0000 UTC m=+177.568800235" Feb 27 07:52:00 crc kubenswrapper[4612]: I0227 07:52:00.140381 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29536312-tjtt6"] Feb 27 07:52:00 crc kubenswrapper[4612]: E0227 07:52:00.140639 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3d35510b-1105-415a-b866-3c3fd63a646b" containerName="registry-server" Feb 27 07:52:00 crc kubenswrapper[4612]: I0227 07:52:00.140658 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="3d35510b-1105-415a-b866-3c3fd63a646b" containerName="registry-server" Feb 27 07:52:00 crc kubenswrapper[4612]: E0227 07:52:00.140678 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3d35510b-1105-415a-b866-3c3fd63a646b" containerName="extract-content" Feb 27 07:52:00 crc kubenswrapper[4612]: I0227 07:52:00.140702 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="3d35510b-1105-415a-b866-3c3fd63a646b" containerName="extract-content" Feb 27 07:52:00 crc kubenswrapper[4612]: E0227 07:52:00.140716 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3d35510b-1105-415a-b866-3c3fd63a646b" containerName="extract-utilities" Feb 27 07:52:00 crc kubenswrapper[4612]: I0227 07:52:00.140725 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="3d35510b-1105-415a-b866-3c3fd63a646b" containerName="extract-utilities" Feb 27 07:52:00 crc kubenswrapper[4612]: I0227 07:52:00.140837 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="3d35510b-1105-415a-b866-3c3fd63a646b" containerName="registry-server" Feb 27 07:52:00 crc kubenswrapper[4612]: I0227 07:52:00.141257 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536312-tjtt6" Feb 27 07:52:00 crc kubenswrapper[4612]: I0227 07:52:00.144011 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-5zwfx" Feb 27 07:52:00 crc kubenswrapper[4612]: I0227 07:52:00.146106 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 27 07:52:00 crc kubenswrapper[4612]: I0227 07:52:00.147188 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 27 07:52:00 crc kubenswrapper[4612]: I0227 07:52:00.155303 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536312-tjtt6"] Feb 27 07:52:00 crc kubenswrapper[4612]: I0227 07:52:00.178795 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cdpst\" (UniqueName: \"kubernetes.io/projected/396f1b18-f225-4d67-a98b-87b2605406d4-kube-api-access-cdpst\") pod \"auto-csr-approver-29536312-tjtt6\" (UID: \"396f1b18-f225-4d67-a98b-87b2605406d4\") " pod="openshift-infra/auto-csr-approver-29536312-tjtt6" Feb 27 07:52:00 crc kubenswrapper[4612]: I0227 07:52:00.280247 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cdpst\" (UniqueName: \"kubernetes.io/projected/396f1b18-f225-4d67-a98b-87b2605406d4-kube-api-access-cdpst\") pod \"auto-csr-approver-29536312-tjtt6\" (UID: \"396f1b18-f225-4d67-a98b-87b2605406d4\") " pod="openshift-infra/auto-csr-approver-29536312-tjtt6" Feb 27 07:52:00 crc kubenswrapper[4612]: I0227 07:52:00.313120 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cdpst\" (UniqueName: \"kubernetes.io/projected/396f1b18-f225-4d67-a98b-87b2605406d4-kube-api-access-cdpst\") pod \"auto-csr-approver-29536312-tjtt6\" (UID: \"396f1b18-f225-4d67-a98b-87b2605406d4\") " pod="openshift-infra/auto-csr-approver-29536312-tjtt6" Feb 27 07:52:00 crc kubenswrapper[4612]: I0227 07:52:00.375591 4612 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-99cd5" Feb 27 07:52:00 crc kubenswrapper[4612]: I0227 07:52:00.375728 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-99cd5" Feb 27 07:52:00 crc kubenswrapper[4612]: I0227 07:52:00.457054 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536312-tjtt6" Feb 27 07:52:00 crc kubenswrapper[4612]: I0227 07:52:00.726220 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fwkxc" event={"ID":"a321b7e1-dbb6-4839-b112-14d7674639c0","Type":"ContainerStarted","Data":"07539f761b9f4d1ae3d06f040cbd6f19e2153f9ebf9c0cfd92de620b1132d10d"} Feb 27 07:52:00 crc kubenswrapper[4612]: I0227 07:52:00.744184 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-w4brj" event={"ID":"0bed0a85-ad8c-4670-8193-0e9a90e88d78","Type":"ContainerStarted","Data":"d9b35ac517f10a4d23b4a3238292b60f5e90ded7b9e5bd0bb0cecd7ba88e541b"} Feb 27 07:52:00 crc kubenswrapper[4612]: I0227 07:52:00.749244 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fjg45" event={"ID":"7676c2d7-9df2-4dc8-a2bd-cde398845d26","Type":"ContainerStarted","Data":"9412e39c97cfdbf2ecdea3e1130977db6d47a3576f38d0de4ca96e6dfdf8d3ed"} Feb 27 07:52:00 crc kubenswrapper[4612]: I0227 07:52:00.759792 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p5fp8" event={"ID":"201de829-c0b7-4d50-9a3e-ef65d3c06916","Type":"ContainerStarted","Data":"a93c39714c2e5eda46c61ea58eae7a9e3f1ba38218d4c700a99e6cc7b1c55c31"} Feb 27 07:52:00 crc kubenswrapper[4612]: I0227 07:52:00.797700 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-w4brj" podStartSLOduration=3.715894187 podStartE2EDuration="49.797651374s" podCreationTimestamp="2026-02-27 07:51:11 +0000 UTC" firstStartedPulling="2026-02-27 07:51:13.946084344 +0000 UTC m=+131.800014342" lastFinishedPulling="2026-02-27 07:52:00.027841531 +0000 UTC m=+177.881771529" observedRunningTime="2026-02-27 07:52:00.77426985 +0000 UTC m=+178.628199848" watchObservedRunningTime="2026-02-27 07:52:00.797651374 +0000 UTC m=+178.651581372" Feb 27 07:52:00 crc kubenswrapper[4612]: I0227 07:52:00.798442 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536312-tjtt6"] Feb 27 07:52:00 crc kubenswrapper[4612]: W0227 07:52:00.827020 4612 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod396f1b18_f225_4d67_a98b_87b2605406d4.slice/crio-e08fa62afab00c7952b1722ceda0a294f75a2b6c33cdb3d25481ef9313fbde90 WatchSource:0}: Error finding container e08fa62afab00c7952b1722ceda0a294f75a2b6c33cdb3d25481ef9313fbde90: Status 404 returned error can't find the container with id e08fa62afab00c7952b1722ceda0a294f75a2b6c33cdb3d25481ef9313fbde90 Feb 27 07:52:00 crc kubenswrapper[4612]: I0227 07:52:00.831856 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-p5fp8" podStartSLOduration=4.556989543 podStartE2EDuration="51.831841266s" podCreationTimestamp="2026-02-27 07:51:09 +0000 UTC" firstStartedPulling="2026-02-27 07:51:12.848424527 +0000 UTC m=+130.702354525" lastFinishedPulling="2026-02-27 07:52:00.12327624 +0000 UTC m=+177.977206248" observedRunningTime="2026-02-27 07:52:00.829176163 +0000 UTC m=+178.683106151" watchObservedRunningTime="2026-02-27 07:52:00.831841266 +0000 UTC m=+178.685771264" Feb 27 07:52:01 crc kubenswrapper[4612]: I0227 07:52:01.482519 4612 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-99cd5" podUID="b4a05b89-0ef2-4f07-b1de-f558a36415a3" containerName="registry-server" probeResult="failure" output=< Feb 27 07:52:01 crc kubenswrapper[4612]: timeout: failed to connect service ":50051" within 1s Feb 27 07:52:01 crc kubenswrapper[4612]: > Feb 27 07:52:01 crc kubenswrapper[4612]: I0227 07:52:01.764520 4612 generic.go:334] "Generic (PLEG): container finished" podID="a321b7e1-dbb6-4839-b112-14d7674639c0" containerID="07539f761b9f4d1ae3d06f040cbd6f19e2153f9ebf9c0cfd92de620b1132d10d" exitCode=0 Feb 27 07:52:01 crc kubenswrapper[4612]: I0227 07:52:01.764574 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fwkxc" event={"ID":"a321b7e1-dbb6-4839-b112-14d7674639c0","Type":"ContainerDied","Data":"07539f761b9f4d1ae3d06f040cbd6f19e2153f9ebf9c0cfd92de620b1132d10d"} Feb 27 07:52:01 crc kubenswrapper[4612]: I0227 07:52:01.767683 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536312-tjtt6" event={"ID":"396f1b18-f225-4d67-a98b-87b2605406d4","Type":"ContainerStarted","Data":"e08fa62afab00c7952b1722ceda0a294f75a2b6c33cdb3d25481ef9313fbde90"} Feb 27 07:52:01 crc kubenswrapper[4612]: I0227 07:52:01.769021 4612 generic.go:334] "Generic (PLEG): container finished" podID="7676c2d7-9df2-4dc8-a2bd-cde398845d26" containerID="9412e39c97cfdbf2ecdea3e1130977db6d47a3576f38d0de4ca96e6dfdf8d3ed" exitCode=0 Feb 27 07:52:01 crc kubenswrapper[4612]: I0227 07:52:01.769684 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fjg45" event={"ID":"7676c2d7-9df2-4dc8-a2bd-cde398845d26","Type":"ContainerDied","Data":"9412e39c97cfdbf2ecdea3e1130977db6d47a3576f38d0de4ca96e6dfdf8d3ed"} Feb 27 07:52:02 crc kubenswrapper[4612]: I0227 07:52:02.009830 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-w4brj" Feb 27 07:52:02 crc kubenswrapper[4612]: I0227 07:52:02.010197 4612 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-w4brj" Feb 27 07:52:02 crc kubenswrapper[4612]: I0227 07:52:02.160818 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-66487b7dd-r2cds"] Feb 27 07:52:02 crc kubenswrapper[4612]: I0227 07:52:02.161007 4612 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-66487b7dd-r2cds" podUID="458167a3-25b3-4cd1-8628-865ad916d3a6" containerName="controller-manager" containerID="cri-o://09c3364c1066ea4c2143584d4fb76486d124f024e1ebdfd3b2362d9396abe8c9" gracePeriod=30 Feb 27 07:52:02 crc kubenswrapper[4612]: I0227 07:52:02.179126 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7d66dbf864-cd8fc"] Feb 27 07:52:02 crc kubenswrapper[4612]: I0227 07:52:02.179319 4612 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-7d66dbf864-cd8fc" podUID="cc6862b7-a2ab-4051-9d83-17774bfc5ac7" containerName="route-controller-manager" containerID="cri-o://80134093f8f5fbc54255812e2817474d95b0563912dc50683b7a6a006cc7eecf" gracePeriod=30 Feb 27 07:52:02 crc kubenswrapper[4612]: I0227 07:52:02.625861 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-66487b7dd-r2cds" Feb 27 07:52:02 crc kubenswrapper[4612]: I0227 07:52:02.694635 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7d66dbf864-cd8fc" Feb 27 07:52:02 crc kubenswrapper[4612]: I0227 07:52:02.798317 4612 generic.go:334] "Generic (PLEG): container finished" podID="458167a3-25b3-4cd1-8628-865ad916d3a6" containerID="09c3364c1066ea4c2143584d4fb76486d124f024e1ebdfd3b2362d9396abe8c9" exitCode=0 Feb 27 07:52:02 crc kubenswrapper[4612]: I0227 07:52:02.798372 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-66487b7dd-r2cds" event={"ID":"458167a3-25b3-4cd1-8628-865ad916d3a6","Type":"ContainerDied","Data":"09c3364c1066ea4c2143584d4fb76486d124f024e1ebdfd3b2362d9396abe8c9"} Feb 27 07:52:02 crc kubenswrapper[4612]: I0227 07:52:02.798396 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-66487b7dd-r2cds" event={"ID":"458167a3-25b3-4cd1-8628-865ad916d3a6","Type":"ContainerDied","Data":"47ac8b300890eb2dbe6d5ee7578f5f522e8820a79526ada246c4cb452b44b1ad"} Feb 27 07:52:02 crc kubenswrapper[4612]: I0227 07:52:02.798412 4612 scope.go:117] "RemoveContainer" containerID="09c3364c1066ea4c2143584d4fb76486d124f024e1ebdfd3b2362d9396abe8c9" Feb 27 07:52:02 crc kubenswrapper[4612]: I0227 07:52:02.798495 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-66487b7dd-r2cds" Feb 27 07:52:02 crc kubenswrapper[4612]: I0227 07:52:02.835513 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/458167a3-25b3-4cd1-8628-865ad916d3a6-serving-cert\") pod \"458167a3-25b3-4cd1-8628-865ad916d3a6\" (UID: \"458167a3-25b3-4cd1-8628-865ad916d3a6\") " Feb 27 07:52:02 crc kubenswrapper[4612]: I0227 07:52:02.843944 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fjg45" event={"ID":"7676c2d7-9df2-4dc8-a2bd-cde398845d26","Type":"ContainerStarted","Data":"56496367db4cbd067458a7e859e26159545b5d2a06fb533402516e4f5ab15d97"} Feb 27 07:52:02 crc kubenswrapper[4612]: I0227 07:52:02.851523 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/458167a3-25b3-4cd1-8628-865ad916d3a6-proxy-ca-bundles\") pod \"458167a3-25b3-4cd1-8628-865ad916d3a6\" (UID: \"458167a3-25b3-4cd1-8628-865ad916d3a6\") " Feb 27 07:52:02 crc kubenswrapper[4612]: I0227 07:52:02.851769 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lstjj\" (UniqueName: \"kubernetes.io/projected/cc6862b7-a2ab-4051-9d83-17774bfc5ac7-kube-api-access-lstjj\") pod \"cc6862b7-a2ab-4051-9d83-17774bfc5ac7\" (UID: \"cc6862b7-a2ab-4051-9d83-17774bfc5ac7\") " Feb 27 07:52:02 crc kubenswrapper[4612]: I0227 07:52:02.856759 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/458167a3-25b3-4cd1-8628-865ad916d3a6-client-ca\") pod \"458167a3-25b3-4cd1-8628-865ad916d3a6\" (UID: \"458167a3-25b3-4cd1-8628-865ad916d3a6\") " Feb 27 07:52:02 crc kubenswrapper[4612]: I0227 07:52:02.856866 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cc6862b7-a2ab-4051-9d83-17774bfc5ac7-config\") pod \"cc6862b7-a2ab-4051-9d83-17774bfc5ac7\" (UID: \"cc6862b7-a2ab-4051-9d83-17774bfc5ac7\") " Feb 27 07:52:02 crc kubenswrapper[4612]: I0227 07:52:02.857007 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cc6862b7-a2ab-4051-9d83-17774bfc5ac7-serving-cert\") pod \"cc6862b7-a2ab-4051-9d83-17774bfc5ac7\" (UID: \"cc6862b7-a2ab-4051-9d83-17774bfc5ac7\") " Feb 27 07:52:02 crc kubenswrapper[4612]: I0227 07:52:02.857095 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4wjzq\" (UniqueName: \"kubernetes.io/projected/458167a3-25b3-4cd1-8628-865ad916d3a6-kube-api-access-4wjzq\") pod \"458167a3-25b3-4cd1-8628-865ad916d3a6\" (UID: \"458167a3-25b3-4cd1-8628-865ad916d3a6\") " Feb 27 07:52:02 crc kubenswrapper[4612]: I0227 07:52:02.857194 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/cc6862b7-a2ab-4051-9d83-17774bfc5ac7-client-ca\") pod \"cc6862b7-a2ab-4051-9d83-17774bfc5ac7\" (UID: \"cc6862b7-a2ab-4051-9d83-17774bfc5ac7\") " Feb 27 07:52:02 crc kubenswrapper[4612]: I0227 07:52:02.854756 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7d66dbf864-cd8fc" Feb 27 07:52:02 crc kubenswrapper[4612]: I0227 07:52:02.854630 4612 generic.go:334] "Generic (PLEG): container finished" podID="cc6862b7-a2ab-4051-9d83-17774bfc5ac7" containerID="80134093f8f5fbc54255812e2817474d95b0563912dc50683b7a6a006cc7eecf" exitCode=0 Feb 27 07:52:02 crc kubenswrapper[4612]: I0227 07:52:02.857531 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/458167a3-25b3-4cd1-8628-865ad916d3a6-config\") pod \"458167a3-25b3-4cd1-8628-865ad916d3a6\" (UID: \"458167a3-25b3-4cd1-8628-865ad916d3a6\") " Feb 27 07:52:02 crc kubenswrapper[4612]: I0227 07:52:02.852740 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/458167a3-25b3-4cd1-8628-865ad916d3a6-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "458167a3-25b3-4cd1-8628-865ad916d3a6" (UID: "458167a3-25b3-4cd1-8628-865ad916d3a6"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 07:52:02 crc kubenswrapper[4612]: I0227 07:52:02.853031 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/458167a3-25b3-4cd1-8628-865ad916d3a6-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "458167a3-25b3-4cd1-8628-865ad916d3a6" (UID: "458167a3-25b3-4cd1-8628-865ad916d3a6"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 07:52:02 crc kubenswrapper[4612]: I0227 07:52:02.857229 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/458167a3-25b3-4cd1-8628-865ad916d3a6-client-ca" (OuterVolumeSpecName: "client-ca") pod "458167a3-25b3-4cd1-8628-865ad916d3a6" (UID: "458167a3-25b3-4cd1-8628-865ad916d3a6"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 07:52:02 crc kubenswrapper[4612]: I0227 07:52:02.857363 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cc6862b7-a2ab-4051-9d83-17774bfc5ac7-kube-api-access-lstjj" (OuterVolumeSpecName: "kube-api-access-lstjj") pod "cc6862b7-a2ab-4051-9d83-17774bfc5ac7" (UID: "cc6862b7-a2ab-4051-9d83-17774bfc5ac7"). InnerVolumeSpecName "kube-api-access-lstjj". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 07:52:02 crc kubenswrapper[4612]: I0227 07:52:02.858297 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cc6862b7-a2ab-4051-9d83-17774bfc5ac7-config" (OuterVolumeSpecName: "config") pod "cc6862b7-a2ab-4051-9d83-17774bfc5ac7" (UID: "cc6862b7-a2ab-4051-9d83-17774bfc5ac7"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 07:52:02 crc kubenswrapper[4612]: I0227 07:52:02.859588 4612 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/458167a3-25b3-4cd1-8628-865ad916d3a6-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 27 07:52:02 crc kubenswrapper[4612]: I0227 07:52:02.862395 4612 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/458167a3-25b3-4cd1-8628-865ad916d3a6-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Feb 27 07:52:02 crc kubenswrapper[4612]: I0227 07:52:02.863305 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lstjj\" (UniqueName: \"kubernetes.io/projected/cc6862b7-a2ab-4051-9d83-17774bfc5ac7-kube-api-access-lstjj\") on node \"crc\" DevicePath \"\"" Feb 27 07:52:02 crc kubenswrapper[4612]: I0227 07:52:02.863415 4612 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/458167a3-25b3-4cd1-8628-865ad916d3a6-client-ca\") on node \"crc\" DevicePath \"\"" Feb 27 07:52:02 crc kubenswrapper[4612]: I0227 07:52:02.863481 4612 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cc6862b7-a2ab-4051-9d83-17774bfc5ac7-config\") on node \"crc\" DevicePath \"\"" Feb 27 07:52:02 crc kubenswrapper[4612]: I0227 07:52:02.863851 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cc6862b7-a2ab-4051-9d83-17774bfc5ac7-client-ca" (OuterVolumeSpecName: "client-ca") pod "cc6862b7-a2ab-4051-9d83-17774bfc5ac7" (UID: "cc6862b7-a2ab-4051-9d83-17774bfc5ac7"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 07:52:02 crc kubenswrapper[4612]: I0227 07:52:02.864793 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cc6862b7-a2ab-4051-9d83-17774bfc5ac7-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "cc6862b7-a2ab-4051-9d83-17774bfc5ac7" (UID: "cc6862b7-a2ab-4051-9d83-17774bfc5ac7"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 07:52:02 crc kubenswrapper[4612]: I0227 07:52:02.865386 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/458167a3-25b3-4cd1-8628-865ad916d3a6-config" (OuterVolumeSpecName: "config") pod "458167a3-25b3-4cd1-8628-865ad916d3a6" (UID: "458167a3-25b3-4cd1-8628-865ad916d3a6"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 07:52:02 crc kubenswrapper[4612]: I0227 07:52:02.867409 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/458167a3-25b3-4cd1-8628-865ad916d3a6-kube-api-access-4wjzq" (OuterVolumeSpecName: "kube-api-access-4wjzq") pod "458167a3-25b3-4cd1-8628-865ad916d3a6" (UID: "458167a3-25b3-4cd1-8628-865ad916d3a6"). InnerVolumeSpecName "kube-api-access-4wjzq". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 07:52:02 crc kubenswrapper[4612]: I0227 07:52:02.886133 4612 scope.go:117] "RemoveContainer" containerID="09c3364c1066ea4c2143584d4fb76486d124f024e1ebdfd3b2362d9396abe8c9" Feb 27 07:52:02 crc kubenswrapper[4612]: E0227 07:52:02.896357 4612 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"09c3364c1066ea4c2143584d4fb76486d124f024e1ebdfd3b2362d9396abe8c9\": container with ID starting with 09c3364c1066ea4c2143584d4fb76486d124f024e1ebdfd3b2362d9396abe8c9 not found: ID does not exist" containerID="09c3364c1066ea4c2143584d4fb76486d124f024e1ebdfd3b2362d9396abe8c9" Feb 27 07:52:02 crc kubenswrapper[4612]: I0227 07:52:02.896410 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"09c3364c1066ea4c2143584d4fb76486d124f024e1ebdfd3b2362d9396abe8c9"} err="failed to get container status \"09c3364c1066ea4c2143584d4fb76486d124f024e1ebdfd3b2362d9396abe8c9\": rpc error: code = NotFound desc = could not find container \"09c3364c1066ea4c2143584d4fb76486d124f024e1ebdfd3b2362d9396abe8c9\": container with ID starting with 09c3364c1066ea4c2143584d4fb76486d124f024e1ebdfd3b2362d9396abe8c9 not found: ID does not exist" Feb 27 07:52:02 crc kubenswrapper[4612]: I0227 07:52:02.918406 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7d66dbf864-cd8fc" event={"ID":"cc6862b7-a2ab-4051-9d83-17774bfc5ac7","Type":"ContainerDied","Data":"80134093f8f5fbc54255812e2817474d95b0563912dc50683b7a6a006cc7eecf"} Feb 27 07:52:02 crc kubenswrapper[4612]: I0227 07:52:02.918612 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7d66dbf864-cd8fc" event={"ID":"cc6862b7-a2ab-4051-9d83-17774bfc5ac7","Type":"ContainerDied","Data":"be2d1a4f1ec68403bc6af09057d10b297350c07592ffd1cd2d6dd0a8a713522a"} Feb 27 07:52:02 crc kubenswrapper[4612]: I0227 07:52:02.918744 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qcsjx" event={"ID":"f167bfef-3a04-4363-aa30-7f75a88a35e9","Type":"ContainerStarted","Data":"9af8303237b79e77bd66b699c83411679cb8e1b91e317bf7370a2c3bab78e75f"} Feb 27 07:52:02 crc kubenswrapper[4612]: I0227 07:52:02.918851 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fwkxc" event={"ID":"a321b7e1-dbb6-4839-b112-14d7674639c0","Type":"ContainerStarted","Data":"5957f89e071a4c4332d217cff26adc7774782198f5c9c6fb6298b97137125ca0"} Feb 27 07:52:02 crc kubenswrapper[4612]: I0227 07:52:02.918925 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qlhr9" event={"ID":"7faf5006-23b1-4ef4-ad29-e0e676340a7c","Type":"ContainerStarted","Data":"a2e0af366c381d79253fa4c1ba8882fa2b8590183c06bfe972f9171eca234e7a"} Feb 27 07:52:02 crc kubenswrapper[4612]: I0227 07:52:02.920739 4612 scope.go:117] "RemoveContainer" containerID="80134093f8f5fbc54255812e2817474d95b0563912dc50683b7a6a006cc7eecf" Feb 27 07:52:02 crc kubenswrapper[4612]: I0227 07:52:02.968586 4612 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/458167a3-25b3-4cd1-8628-865ad916d3a6-config\") on node \"crc\" DevicePath \"\"" Feb 27 07:52:02 crc kubenswrapper[4612]: I0227 07:52:02.968874 4612 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cc6862b7-a2ab-4051-9d83-17774bfc5ac7-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 27 07:52:02 crc kubenswrapper[4612]: I0227 07:52:02.968937 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4wjzq\" (UniqueName: \"kubernetes.io/projected/458167a3-25b3-4cd1-8628-865ad916d3a6-kube-api-access-4wjzq\") on node \"crc\" DevicePath \"\"" Feb 27 07:52:02 crc kubenswrapper[4612]: I0227 07:52:02.968950 4612 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/cc6862b7-a2ab-4051-9d83-17774bfc5ac7-client-ca\") on node \"crc\" DevicePath \"\"" Feb 27 07:52:02 crc kubenswrapper[4612]: I0227 07:52:02.978958 4612 scope.go:117] "RemoveContainer" containerID="80134093f8f5fbc54255812e2817474d95b0563912dc50683b7a6a006cc7eecf" Feb 27 07:52:02 crc kubenswrapper[4612]: E0227 07:52:02.981095 4612 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"80134093f8f5fbc54255812e2817474d95b0563912dc50683b7a6a006cc7eecf\": container with ID starting with 80134093f8f5fbc54255812e2817474d95b0563912dc50683b7a6a006cc7eecf not found: ID does not exist" containerID="80134093f8f5fbc54255812e2817474d95b0563912dc50683b7a6a006cc7eecf" Feb 27 07:52:02 crc kubenswrapper[4612]: I0227 07:52:02.981179 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"80134093f8f5fbc54255812e2817474d95b0563912dc50683b7a6a006cc7eecf"} err="failed to get container status \"80134093f8f5fbc54255812e2817474d95b0563912dc50683b7a6a006cc7eecf\": rpc error: code = NotFound desc = could not find container \"80134093f8f5fbc54255812e2817474d95b0563912dc50683b7a6a006cc7eecf\": container with ID starting with 80134093f8f5fbc54255812e2817474d95b0563912dc50683b7a6a006cc7eecf not found: ID does not exist" Feb 27 07:52:03 crc kubenswrapper[4612]: I0227 07:52:03.004668 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7d66dbf864-cd8fc"] Feb 27 07:52:03 crc kubenswrapper[4612]: I0227 07:52:03.005271 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7d66dbf864-cd8fc"] Feb 27 07:52:03 crc kubenswrapper[4612]: I0227 07:52:03.024503 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-fwkxc" podStartSLOduration=3.6828512939999998 podStartE2EDuration="51.024488819s" podCreationTimestamp="2026-02-27 07:51:12 +0000 UTC" firstStartedPulling="2026-02-27 07:51:15.032711048 +0000 UTC m=+132.886641046" lastFinishedPulling="2026-02-27 07:52:02.374348573 +0000 UTC m=+180.228278571" observedRunningTime="2026-02-27 07:52:03.021468836 +0000 UTC m=+180.875398834" watchObservedRunningTime="2026-02-27 07:52:03.024488819 +0000 UTC m=+180.878418817" Feb 27 07:52:03 crc kubenswrapper[4612]: I0227 07:52:03.046807 4612 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-fwkxc" Feb 27 07:52:03 crc kubenswrapper[4612]: I0227 07:52:03.046868 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-fwkxc" Feb 27 07:52:03 crc kubenswrapper[4612]: I0227 07:52:03.065763 4612 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-marketplace-w4brj" podUID="0bed0a85-ad8c-4670-8193-0e9a90e88d78" containerName="registry-server" probeResult="failure" output=< Feb 27 07:52:03 crc kubenswrapper[4612]: timeout: failed to connect service ":50051" within 1s Feb 27 07:52:03 crc kubenswrapper[4612]: > Feb 27 07:52:03 crc kubenswrapper[4612]: I0227 07:52:03.137681 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-66487b7dd-r2cds"] Feb 27 07:52:03 crc kubenswrapper[4612]: I0227 07:52:03.147163 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-66487b7dd-r2cds"] Feb 27 07:52:03 crc kubenswrapper[4612]: I0227 07:52:03.714889 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-59f7dcc9c8-9b7dk"] Feb 27 07:52:03 crc kubenswrapper[4612]: E0227 07:52:03.715381 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc6862b7-a2ab-4051-9d83-17774bfc5ac7" containerName="route-controller-manager" Feb 27 07:52:03 crc kubenswrapper[4612]: I0227 07:52:03.715414 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc6862b7-a2ab-4051-9d83-17774bfc5ac7" containerName="route-controller-manager" Feb 27 07:52:03 crc kubenswrapper[4612]: E0227 07:52:03.715432 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="458167a3-25b3-4cd1-8628-865ad916d3a6" containerName="controller-manager" Feb 27 07:52:03 crc kubenswrapper[4612]: I0227 07:52:03.715441 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="458167a3-25b3-4cd1-8628-865ad916d3a6" containerName="controller-manager" Feb 27 07:52:03 crc kubenswrapper[4612]: I0227 07:52:03.715815 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="458167a3-25b3-4cd1-8628-865ad916d3a6" containerName="controller-manager" Feb 27 07:52:03 crc kubenswrapper[4612]: I0227 07:52:03.715881 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="cc6862b7-a2ab-4051-9d83-17774bfc5ac7" containerName="route-controller-manager" Feb 27 07:52:03 crc kubenswrapper[4612]: I0227 07:52:03.717092 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-59f7dcc9c8-9b7dk" Feb 27 07:52:03 crc kubenswrapper[4612]: I0227 07:52:03.719173 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5969dcdf8-bbhhq"] Feb 27 07:52:03 crc kubenswrapper[4612]: I0227 07:52:03.721209 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Feb 27 07:52:03 crc kubenswrapper[4612]: I0227 07:52:03.721410 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Feb 27 07:52:03 crc kubenswrapper[4612]: I0227 07:52:03.721499 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Feb 27 07:52:03 crc kubenswrapper[4612]: I0227 07:52:03.721545 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Feb 27 07:52:03 crc kubenswrapper[4612]: I0227 07:52:03.722481 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Feb 27 07:52:03 crc kubenswrapper[4612]: I0227 07:52:03.722895 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5969dcdf8-bbhhq" Feb 27 07:52:03 crc kubenswrapper[4612]: I0227 07:52:03.723611 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Feb 27 07:52:03 crc kubenswrapper[4612]: I0227 07:52:03.726857 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-59f7dcc9c8-9b7dk"] Feb 27 07:52:03 crc kubenswrapper[4612]: I0227 07:52:03.731155 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Feb 27 07:52:03 crc kubenswrapper[4612]: I0227 07:52:03.731732 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Feb 27 07:52:03 crc kubenswrapper[4612]: I0227 07:52:03.731891 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Feb 27 07:52:03 crc kubenswrapper[4612]: I0227 07:52:03.732107 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Feb 27 07:52:03 crc kubenswrapper[4612]: I0227 07:52:03.732989 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Feb 27 07:52:03 crc kubenswrapper[4612]: I0227 07:52:03.733583 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Feb 27 07:52:03 crc kubenswrapper[4612]: I0227 07:52:03.733654 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Feb 27 07:52:03 crc kubenswrapper[4612]: I0227 07:52:03.738680 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5969dcdf8-bbhhq"] Feb 27 07:52:03 crc kubenswrapper[4612]: I0227 07:52:03.778328 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f04702a5-b6df-431b-9f93-a37d9bf5dc96-serving-cert\") pod \"route-controller-manager-5969dcdf8-bbhhq\" (UID: \"f04702a5-b6df-431b-9f93-a37d9bf5dc96\") " pod="openshift-route-controller-manager/route-controller-manager-5969dcdf8-bbhhq" Feb 27 07:52:03 crc kubenswrapper[4612]: I0227 07:52:03.778380 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c07650a7-6d9b-43da-ae78-e8d8dbe342f4-serving-cert\") pod \"controller-manager-59f7dcc9c8-9b7dk\" (UID: \"c07650a7-6d9b-43da-ae78-e8d8dbe342f4\") " pod="openshift-controller-manager/controller-manager-59f7dcc9c8-9b7dk" Feb 27 07:52:03 crc kubenswrapper[4612]: I0227 07:52:03.778444 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8k8z9\" (UniqueName: \"kubernetes.io/projected/f04702a5-b6df-431b-9f93-a37d9bf5dc96-kube-api-access-8k8z9\") pod \"route-controller-manager-5969dcdf8-bbhhq\" (UID: \"f04702a5-b6df-431b-9f93-a37d9bf5dc96\") " pod="openshift-route-controller-manager/route-controller-manager-5969dcdf8-bbhhq" Feb 27 07:52:03 crc kubenswrapper[4612]: I0227 07:52:03.778519 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f04702a5-b6df-431b-9f93-a37d9bf5dc96-client-ca\") pod \"route-controller-manager-5969dcdf8-bbhhq\" (UID: \"f04702a5-b6df-431b-9f93-a37d9bf5dc96\") " pod="openshift-route-controller-manager/route-controller-manager-5969dcdf8-bbhhq" Feb 27 07:52:03 crc kubenswrapper[4612]: I0227 07:52:03.778554 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c07650a7-6d9b-43da-ae78-e8d8dbe342f4-config\") pod \"controller-manager-59f7dcc9c8-9b7dk\" (UID: \"c07650a7-6d9b-43da-ae78-e8d8dbe342f4\") " pod="openshift-controller-manager/controller-manager-59f7dcc9c8-9b7dk" Feb 27 07:52:03 crc kubenswrapper[4612]: I0227 07:52:03.778580 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c07650a7-6d9b-43da-ae78-e8d8dbe342f4-client-ca\") pod \"controller-manager-59f7dcc9c8-9b7dk\" (UID: \"c07650a7-6d9b-43da-ae78-e8d8dbe342f4\") " pod="openshift-controller-manager/controller-manager-59f7dcc9c8-9b7dk" Feb 27 07:52:03 crc kubenswrapper[4612]: I0227 07:52:03.778602 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c07650a7-6d9b-43da-ae78-e8d8dbe342f4-proxy-ca-bundles\") pod \"controller-manager-59f7dcc9c8-9b7dk\" (UID: \"c07650a7-6d9b-43da-ae78-e8d8dbe342f4\") " pod="openshift-controller-manager/controller-manager-59f7dcc9c8-9b7dk" Feb 27 07:52:03 crc kubenswrapper[4612]: I0227 07:52:03.778629 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jd5vk\" (UniqueName: \"kubernetes.io/projected/c07650a7-6d9b-43da-ae78-e8d8dbe342f4-kube-api-access-jd5vk\") pod \"controller-manager-59f7dcc9c8-9b7dk\" (UID: \"c07650a7-6d9b-43da-ae78-e8d8dbe342f4\") " pod="openshift-controller-manager/controller-manager-59f7dcc9c8-9b7dk" Feb 27 07:52:03 crc kubenswrapper[4612]: I0227 07:52:03.778651 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f04702a5-b6df-431b-9f93-a37d9bf5dc96-config\") pod \"route-controller-manager-5969dcdf8-bbhhq\" (UID: \"f04702a5-b6df-431b-9f93-a37d9bf5dc96\") " pod="openshift-route-controller-manager/route-controller-manager-5969dcdf8-bbhhq" Feb 27 07:52:03 crc kubenswrapper[4612]: I0227 07:52:03.879929 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8k8z9\" (UniqueName: \"kubernetes.io/projected/f04702a5-b6df-431b-9f93-a37d9bf5dc96-kube-api-access-8k8z9\") pod \"route-controller-manager-5969dcdf8-bbhhq\" (UID: \"f04702a5-b6df-431b-9f93-a37d9bf5dc96\") " pod="openshift-route-controller-manager/route-controller-manager-5969dcdf8-bbhhq" Feb 27 07:52:03 crc kubenswrapper[4612]: I0227 07:52:03.880022 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f04702a5-b6df-431b-9f93-a37d9bf5dc96-client-ca\") pod \"route-controller-manager-5969dcdf8-bbhhq\" (UID: \"f04702a5-b6df-431b-9f93-a37d9bf5dc96\") " pod="openshift-route-controller-manager/route-controller-manager-5969dcdf8-bbhhq" Feb 27 07:52:03 crc kubenswrapper[4612]: I0227 07:52:03.880066 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c07650a7-6d9b-43da-ae78-e8d8dbe342f4-config\") pod \"controller-manager-59f7dcc9c8-9b7dk\" (UID: \"c07650a7-6d9b-43da-ae78-e8d8dbe342f4\") " pod="openshift-controller-manager/controller-manager-59f7dcc9c8-9b7dk" Feb 27 07:52:03 crc kubenswrapper[4612]: I0227 07:52:03.880994 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f04702a5-b6df-431b-9f93-a37d9bf5dc96-client-ca\") pod \"route-controller-manager-5969dcdf8-bbhhq\" (UID: \"f04702a5-b6df-431b-9f93-a37d9bf5dc96\") " pod="openshift-route-controller-manager/route-controller-manager-5969dcdf8-bbhhq" Feb 27 07:52:03 crc kubenswrapper[4612]: I0227 07:52:03.881657 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c07650a7-6d9b-43da-ae78-e8d8dbe342f4-config\") pod \"controller-manager-59f7dcc9c8-9b7dk\" (UID: \"c07650a7-6d9b-43da-ae78-e8d8dbe342f4\") " pod="openshift-controller-manager/controller-manager-59f7dcc9c8-9b7dk" Feb 27 07:52:03 crc kubenswrapper[4612]: I0227 07:52:03.881768 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c07650a7-6d9b-43da-ae78-e8d8dbe342f4-client-ca\") pod \"controller-manager-59f7dcc9c8-9b7dk\" (UID: \"c07650a7-6d9b-43da-ae78-e8d8dbe342f4\") " pod="openshift-controller-manager/controller-manager-59f7dcc9c8-9b7dk" Feb 27 07:52:03 crc kubenswrapper[4612]: I0227 07:52:03.881787 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c07650a7-6d9b-43da-ae78-e8d8dbe342f4-proxy-ca-bundles\") pod \"controller-manager-59f7dcc9c8-9b7dk\" (UID: \"c07650a7-6d9b-43da-ae78-e8d8dbe342f4\") " pod="openshift-controller-manager/controller-manager-59f7dcc9c8-9b7dk" Feb 27 07:52:03 crc kubenswrapper[4612]: I0227 07:52:03.881825 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jd5vk\" (UniqueName: \"kubernetes.io/projected/c07650a7-6d9b-43da-ae78-e8d8dbe342f4-kube-api-access-jd5vk\") pod \"controller-manager-59f7dcc9c8-9b7dk\" (UID: \"c07650a7-6d9b-43da-ae78-e8d8dbe342f4\") " pod="openshift-controller-manager/controller-manager-59f7dcc9c8-9b7dk" Feb 27 07:52:03 crc kubenswrapper[4612]: I0227 07:52:03.881844 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f04702a5-b6df-431b-9f93-a37d9bf5dc96-config\") pod \"route-controller-manager-5969dcdf8-bbhhq\" (UID: \"f04702a5-b6df-431b-9f93-a37d9bf5dc96\") " pod="openshift-route-controller-manager/route-controller-manager-5969dcdf8-bbhhq" Feb 27 07:52:03 crc kubenswrapper[4612]: I0227 07:52:03.882595 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f04702a5-b6df-431b-9f93-a37d9bf5dc96-serving-cert\") pod \"route-controller-manager-5969dcdf8-bbhhq\" (UID: \"f04702a5-b6df-431b-9f93-a37d9bf5dc96\") " pod="openshift-route-controller-manager/route-controller-manager-5969dcdf8-bbhhq" Feb 27 07:52:03 crc kubenswrapper[4612]: I0227 07:52:03.882390 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c07650a7-6d9b-43da-ae78-e8d8dbe342f4-client-ca\") pod \"controller-manager-59f7dcc9c8-9b7dk\" (UID: \"c07650a7-6d9b-43da-ae78-e8d8dbe342f4\") " pod="openshift-controller-manager/controller-manager-59f7dcc9c8-9b7dk" Feb 27 07:52:03 crc kubenswrapper[4612]: I0227 07:52:03.882670 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c07650a7-6d9b-43da-ae78-e8d8dbe342f4-serving-cert\") pod \"controller-manager-59f7dcc9c8-9b7dk\" (UID: \"c07650a7-6d9b-43da-ae78-e8d8dbe342f4\") " pod="openshift-controller-manager/controller-manager-59f7dcc9c8-9b7dk" Feb 27 07:52:03 crc kubenswrapper[4612]: I0227 07:52:03.882540 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c07650a7-6d9b-43da-ae78-e8d8dbe342f4-proxy-ca-bundles\") pod \"controller-manager-59f7dcc9c8-9b7dk\" (UID: \"c07650a7-6d9b-43da-ae78-e8d8dbe342f4\") " pod="openshift-controller-manager/controller-manager-59f7dcc9c8-9b7dk" Feb 27 07:52:03 crc kubenswrapper[4612]: I0227 07:52:03.882935 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f04702a5-b6df-431b-9f93-a37d9bf5dc96-config\") pod \"route-controller-manager-5969dcdf8-bbhhq\" (UID: \"f04702a5-b6df-431b-9f93-a37d9bf5dc96\") " pod="openshift-route-controller-manager/route-controller-manager-5969dcdf8-bbhhq" Feb 27 07:52:03 crc kubenswrapper[4612]: I0227 07:52:03.890218 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c07650a7-6d9b-43da-ae78-e8d8dbe342f4-serving-cert\") pod \"controller-manager-59f7dcc9c8-9b7dk\" (UID: \"c07650a7-6d9b-43da-ae78-e8d8dbe342f4\") " pod="openshift-controller-manager/controller-manager-59f7dcc9c8-9b7dk" Feb 27 07:52:03 crc kubenswrapper[4612]: I0227 07:52:03.893438 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f04702a5-b6df-431b-9f93-a37d9bf5dc96-serving-cert\") pod \"route-controller-manager-5969dcdf8-bbhhq\" (UID: \"f04702a5-b6df-431b-9f93-a37d9bf5dc96\") " pod="openshift-route-controller-manager/route-controller-manager-5969dcdf8-bbhhq" Feb 27 07:52:03 crc kubenswrapper[4612]: I0227 07:52:03.896811 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8k8z9\" (UniqueName: \"kubernetes.io/projected/f04702a5-b6df-431b-9f93-a37d9bf5dc96-kube-api-access-8k8z9\") pod \"route-controller-manager-5969dcdf8-bbhhq\" (UID: \"f04702a5-b6df-431b-9f93-a37d9bf5dc96\") " pod="openshift-route-controller-manager/route-controller-manager-5969dcdf8-bbhhq" Feb 27 07:52:03 crc kubenswrapper[4612]: I0227 07:52:03.905757 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jd5vk\" (UniqueName: \"kubernetes.io/projected/c07650a7-6d9b-43da-ae78-e8d8dbe342f4-kube-api-access-jd5vk\") pod \"controller-manager-59f7dcc9c8-9b7dk\" (UID: \"c07650a7-6d9b-43da-ae78-e8d8dbe342f4\") " pod="openshift-controller-manager/controller-manager-59f7dcc9c8-9b7dk" Feb 27 07:52:03 crc kubenswrapper[4612]: I0227 07:52:03.918011 4612 generic.go:334] "Generic (PLEG): container finished" podID="7faf5006-23b1-4ef4-ad29-e0e676340a7c" containerID="a2e0af366c381d79253fa4c1ba8882fa2b8590183c06bfe972f9171eca234e7a" exitCode=0 Feb 27 07:52:03 crc kubenswrapper[4612]: I0227 07:52:03.918118 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qlhr9" event={"ID":"7faf5006-23b1-4ef4-ad29-e0e676340a7c","Type":"ContainerDied","Data":"a2e0af366c381d79253fa4c1ba8882fa2b8590183c06bfe972f9171eca234e7a"} Feb 27 07:52:03 crc kubenswrapper[4612]: I0227 07:52:03.928641 4612 generic.go:334] "Generic (PLEG): container finished" podID="f167bfef-3a04-4363-aa30-7f75a88a35e9" containerID="9af8303237b79e77bd66b699c83411679cb8e1b91e317bf7370a2c3bab78e75f" exitCode=0 Feb 27 07:52:03 crc kubenswrapper[4612]: I0227 07:52:03.930488 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qcsjx" event={"ID":"f167bfef-3a04-4363-aa30-7f75a88a35e9","Type":"ContainerDied","Data":"9af8303237b79e77bd66b699c83411679cb8e1b91e317bf7370a2c3bab78e75f"} Feb 27 07:52:04 crc kubenswrapper[4612]: I0227 07:52:04.044566 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-59f7dcc9c8-9b7dk" Feb 27 07:52:04 crc kubenswrapper[4612]: I0227 07:52:04.049880 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5969dcdf8-bbhhq" Feb 27 07:52:04 crc kubenswrapper[4612]: I0227 07:52:04.101609 4612 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-fwkxc" podUID="a321b7e1-dbb6-4839-b112-14d7674639c0" containerName="registry-server" probeResult="failure" output=< Feb 27 07:52:04 crc kubenswrapper[4612]: timeout: failed to connect service ":50051" within 1s Feb 27 07:52:04 crc kubenswrapper[4612]: > Feb 27 07:52:04 crc kubenswrapper[4612]: I0227 07:52:04.371331 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-fjg45" podStartSLOduration=3.987962238 podStartE2EDuration="51.371309692s" podCreationTimestamp="2026-02-27 07:51:13 +0000 UTC" firstStartedPulling="2026-02-27 07:51:15.012261145 +0000 UTC m=+132.866191143" lastFinishedPulling="2026-02-27 07:52:02.395608589 +0000 UTC m=+180.249538597" observedRunningTime="2026-02-27 07:52:03.990645452 +0000 UTC m=+181.844575460" watchObservedRunningTime="2026-02-27 07:52:04.371309692 +0000 UTC m=+182.225239690" Feb 27 07:52:04 crc kubenswrapper[4612]: I0227 07:52:04.386269 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-59f7dcc9c8-9b7dk"] Feb 27 07:52:04 crc kubenswrapper[4612]: I0227 07:52:04.408310 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5969dcdf8-bbhhq"] Feb 27 07:52:04 crc kubenswrapper[4612]: I0227 07:52:04.859916 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="458167a3-25b3-4cd1-8628-865ad916d3a6" path="/var/lib/kubelet/pods/458167a3-25b3-4cd1-8628-865ad916d3a6/volumes" Feb 27 07:52:04 crc kubenswrapper[4612]: I0227 07:52:04.860812 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cc6862b7-a2ab-4051-9d83-17774bfc5ac7" path="/var/lib/kubelet/pods/cc6862b7-a2ab-4051-9d83-17774bfc5ac7/volumes" Feb 27 07:52:04 crc kubenswrapper[4612]: E0227 07:52:04.946886 4612 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcc6862b7_a2ab_4051_9d83_17774bfc5ac7.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod458167a3_25b3_4cd1_8628_865ad916d3a6.slice/crio-47ac8b300890eb2dbe6d5ee7578f5f522e8820a79526ada246c4cb452b44b1ad\": RecentStats: unable to find data in memory cache]" Feb 27 07:52:04 crc kubenswrapper[4612]: I0227 07:52:04.955615 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5969dcdf8-bbhhq" event={"ID":"f04702a5-b6df-431b-9f93-a37d9bf5dc96","Type":"ContainerStarted","Data":"e67f8de92a83ce39fcf3b74a51fadf0ab4e216d2f00209fd9b8cc92ab0c9e46a"} Feb 27 07:52:04 crc kubenswrapper[4612]: I0227 07:52:04.955653 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5969dcdf8-bbhhq" event={"ID":"f04702a5-b6df-431b-9f93-a37d9bf5dc96","Type":"ContainerStarted","Data":"84b5510c599bc98767d5a231f39eed73c857e7009563960fa82229dc78a38de5"} Feb 27 07:52:04 crc kubenswrapper[4612]: I0227 07:52:04.956110 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-5969dcdf8-bbhhq" Feb 27 07:52:04 crc kubenswrapper[4612]: I0227 07:52:04.962577 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qcsjx" event={"ID":"f167bfef-3a04-4363-aa30-7f75a88a35e9","Type":"ContainerStarted","Data":"da9a4aee21ae34fe8a6ee2c2a89a45943fcf57d05dec12aeadfc759f1fe2d89a"} Feb 27 07:52:04 crc kubenswrapper[4612]: I0227 07:52:04.964942 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-59f7dcc9c8-9b7dk" event={"ID":"c07650a7-6d9b-43da-ae78-e8d8dbe342f4","Type":"ContainerStarted","Data":"8f23ab18cb7541e20fb46e1015f363ad7d720d9b768f2f2fb0d6a861887e5df5"} Feb 27 07:52:04 crc kubenswrapper[4612]: I0227 07:52:04.964974 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-59f7dcc9c8-9b7dk" event={"ID":"c07650a7-6d9b-43da-ae78-e8d8dbe342f4","Type":"ContainerStarted","Data":"c3a1755593c4fc24b8296fe431ce2152d07bb1d724dd53bf31d3881fb8986642"} Feb 27 07:52:04 crc kubenswrapper[4612]: I0227 07:52:04.965712 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-59f7dcc9c8-9b7dk" Feb 27 07:52:04 crc kubenswrapper[4612]: I0227 07:52:04.967189 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qlhr9" event={"ID":"7faf5006-23b1-4ef4-ad29-e0e676340a7c","Type":"ContainerStarted","Data":"bc3d2a71143833b070c3078c2c2694503dc694341f36a9dad42134afa1a0c8c9"} Feb 27 07:52:04 crc kubenswrapper[4612]: I0227 07:52:04.972453 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-5969dcdf8-bbhhq" podStartSLOduration=2.972438808 podStartE2EDuration="2.972438808s" podCreationTimestamp="2026-02-27 07:52:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 07:52:04.971307146 +0000 UTC m=+182.825237144" watchObservedRunningTime="2026-02-27 07:52:04.972438808 +0000 UTC m=+182.826368806" Feb 27 07:52:04 crc kubenswrapper[4612]: I0227 07:52:04.974384 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-59f7dcc9c8-9b7dk" Feb 27 07:52:04 crc kubenswrapper[4612]: I0227 07:52:04.996831 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-qlhr9" podStartSLOduration=4.081689955 podStartE2EDuration="55.996813139s" podCreationTimestamp="2026-02-27 07:51:09 +0000 UTC" firstStartedPulling="2026-02-27 07:51:12.762805578 +0000 UTC m=+130.616735566" lastFinishedPulling="2026-02-27 07:52:04.677928752 +0000 UTC m=+182.531858750" observedRunningTime="2026-02-27 07:52:04.995281237 +0000 UTC m=+182.849211235" watchObservedRunningTime="2026-02-27 07:52:04.996813139 +0000 UTC m=+182.850743137" Feb 27 07:52:05 crc kubenswrapper[4612]: I0227 07:52:05.018910 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-59f7dcc9c8-9b7dk" podStartSLOduration=3.018896108 podStartE2EDuration="3.018896108s" podCreationTimestamp="2026-02-27 07:52:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 07:52:05.01826191 +0000 UTC m=+182.872191908" watchObservedRunningTime="2026-02-27 07:52:05.018896108 +0000 UTC m=+182.872826106" Feb 27 07:52:05 crc kubenswrapper[4612]: I0227 07:52:05.047191 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-qcsjx" podStartSLOduration=4.775849825 podStartE2EDuration="56.047172037s" podCreationTimestamp="2026-02-27 07:51:09 +0000 UTC" firstStartedPulling="2026-02-27 07:51:12.744246797 +0000 UTC m=+130.598176795" lastFinishedPulling="2026-02-27 07:52:04.015569009 +0000 UTC m=+181.869499007" observedRunningTime="2026-02-27 07:52:05.042146848 +0000 UTC m=+182.896076856" watchObservedRunningTime="2026-02-27 07:52:05.047172037 +0000 UTC m=+182.901102035" Feb 27 07:52:05 crc kubenswrapper[4612]: I0227 07:52:05.336256 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-5969dcdf8-bbhhq" Feb 27 07:52:08 crc kubenswrapper[4612]: I0227 07:52:08.715448 4612 ???:1] "http: TLS handshake error from 192.168.126.11:46668: no serving certificate available for the kubelet" Feb 27 07:52:09 crc kubenswrapper[4612]: I0227 07:52:09.835730 4612 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-p5fp8" Feb 27 07:52:09 crc kubenswrapper[4612]: I0227 07:52:09.835780 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-p5fp8" Feb 27 07:52:09 crc kubenswrapper[4612]: I0227 07:52:09.879252 4612 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-p5fp8" Feb 27 07:52:10 crc kubenswrapper[4612]: I0227 07:52:10.009781 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-qlhr9" Feb 27 07:52:10 crc kubenswrapper[4612]: I0227 07:52:10.010129 4612 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-qlhr9" Feb 27 07:52:10 crc kubenswrapper[4612]: I0227 07:52:10.057490 4612 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-qlhr9" Feb 27 07:52:10 crc kubenswrapper[4612]: I0227 07:52:10.068161 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-p5fp8" Feb 27 07:52:10 crc kubenswrapper[4612]: I0227 07:52:10.255892 4612 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-qcsjx" Feb 27 07:52:10 crc kubenswrapper[4612]: I0227 07:52:10.255948 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-qcsjx" Feb 27 07:52:10 crc kubenswrapper[4612]: I0227 07:52:10.314930 4612 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-qcsjx" Feb 27 07:52:10 crc kubenswrapper[4612]: I0227 07:52:10.424327 4612 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-99cd5" Feb 27 07:52:10 crc kubenswrapper[4612]: I0227 07:52:10.466472 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-99cd5" Feb 27 07:52:11 crc kubenswrapper[4612]: I0227 07:52:11.039265 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-qlhr9" Feb 27 07:52:11 crc kubenswrapper[4612]: I0227 07:52:11.053892 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-qcsjx" Feb 27 07:52:12 crc kubenswrapper[4612]: I0227 07:52:12.087154 4612 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-w4brj" Feb 27 07:52:12 crc kubenswrapper[4612]: I0227 07:52:12.139375 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-w4brj" Feb 27 07:52:12 crc kubenswrapper[4612]: I0227 07:52:12.170025 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-qcsjx"] Feb 27 07:52:12 crc kubenswrapper[4612]: I0227 07:52:12.367661 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-99cd5"] Feb 27 07:52:12 crc kubenswrapper[4612]: I0227 07:52:12.367910 4612 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-99cd5" podUID="b4a05b89-0ef2-4f07-b1de-f558a36415a3" containerName="registry-server" containerID="cri-o://b9fd648455a1f23c8fe4b0657e89196db987d450104c1ad5ea1c2b53f4c669bd" gracePeriod=2 Feb 27 07:52:13 crc kubenswrapper[4612]: I0227 07:52:13.018217 4612 generic.go:334] "Generic (PLEG): container finished" podID="b4a05b89-0ef2-4f07-b1de-f558a36415a3" containerID="b9fd648455a1f23c8fe4b0657e89196db987d450104c1ad5ea1c2b53f4c669bd" exitCode=0 Feb 27 07:52:13 crc kubenswrapper[4612]: I0227 07:52:13.018676 4612 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-qcsjx" podUID="f167bfef-3a04-4363-aa30-7f75a88a35e9" containerName="registry-server" containerID="cri-o://da9a4aee21ae34fe8a6ee2c2a89a45943fcf57d05dec12aeadfc759f1fe2d89a" gracePeriod=2 Feb 27 07:52:13 crc kubenswrapper[4612]: I0227 07:52:13.018253 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-99cd5" event={"ID":"b4a05b89-0ef2-4f07-b1de-f558a36415a3","Type":"ContainerDied","Data":"b9fd648455a1f23c8fe4b0657e89196db987d450104c1ad5ea1c2b53f4c669bd"} Feb 27 07:52:13 crc kubenswrapper[4612]: I0227 07:52:13.082283 4612 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-fwkxc" Feb 27 07:52:13 crc kubenswrapper[4612]: I0227 07:52:13.125949 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-fwkxc" Feb 27 07:52:13 crc kubenswrapper[4612]: I0227 07:52:13.413277 4612 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-fjg45" Feb 27 07:52:13 crc kubenswrapper[4612]: I0227 07:52:13.413330 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-fjg45" Feb 27 07:52:13 crc kubenswrapper[4612]: I0227 07:52:13.457625 4612 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-fjg45" Feb 27 07:52:13 crc kubenswrapper[4612]: I0227 07:52:13.812641 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-qcsjx" Feb 27 07:52:13 crc kubenswrapper[4612]: I0227 07:52:13.814932 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-99cd5" Feb 27 07:52:13 crc kubenswrapper[4612]: I0227 07:52:13.932197 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qwhq4\" (UniqueName: \"kubernetes.io/projected/f167bfef-3a04-4363-aa30-7f75a88a35e9-kube-api-access-qwhq4\") pod \"f167bfef-3a04-4363-aa30-7f75a88a35e9\" (UID: \"f167bfef-3a04-4363-aa30-7f75a88a35e9\") " Feb 27 07:52:13 crc kubenswrapper[4612]: I0227 07:52:13.932567 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f167bfef-3a04-4363-aa30-7f75a88a35e9-catalog-content\") pod \"f167bfef-3a04-4363-aa30-7f75a88a35e9\" (UID: \"f167bfef-3a04-4363-aa30-7f75a88a35e9\") " Feb 27 07:52:13 crc kubenswrapper[4612]: I0227 07:52:13.932596 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b4a05b89-0ef2-4f07-b1de-f558a36415a3-catalog-content\") pod \"b4a05b89-0ef2-4f07-b1de-f558a36415a3\" (UID: \"b4a05b89-0ef2-4f07-b1de-f558a36415a3\") " Feb 27 07:52:13 crc kubenswrapper[4612]: I0227 07:52:13.932636 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b4a05b89-0ef2-4f07-b1de-f558a36415a3-utilities\") pod \"b4a05b89-0ef2-4f07-b1de-f558a36415a3\" (UID: \"b4a05b89-0ef2-4f07-b1de-f558a36415a3\") " Feb 27 07:52:13 crc kubenswrapper[4612]: I0227 07:52:13.932734 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rkh9l\" (UniqueName: \"kubernetes.io/projected/b4a05b89-0ef2-4f07-b1de-f558a36415a3-kube-api-access-rkh9l\") pod \"b4a05b89-0ef2-4f07-b1de-f558a36415a3\" (UID: \"b4a05b89-0ef2-4f07-b1de-f558a36415a3\") " Feb 27 07:52:13 crc kubenswrapper[4612]: I0227 07:52:13.932758 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f167bfef-3a04-4363-aa30-7f75a88a35e9-utilities\") pod \"f167bfef-3a04-4363-aa30-7f75a88a35e9\" (UID: \"f167bfef-3a04-4363-aa30-7f75a88a35e9\") " Feb 27 07:52:13 crc kubenswrapper[4612]: I0227 07:52:13.933676 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b4a05b89-0ef2-4f07-b1de-f558a36415a3-utilities" (OuterVolumeSpecName: "utilities") pod "b4a05b89-0ef2-4f07-b1de-f558a36415a3" (UID: "b4a05b89-0ef2-4f07-b1de-f558a36415a3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 07:52:13 crc kubenswrapper[4612]: I0227 07:52:13.933711 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f167bfef-3a04-4363-aa30-7f75a88a35e9-utilities" (OuterVolumeSpecName: "utilities") pod "f167bfef-3a04-4363-aa30-7f75a88a35e9" (UID: "f167bfef-3a04-4363-aa30-7f75a88a35e9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 07:52:13 crc kubenswrapper[4612]: I0227 07:52:13.943877 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f167bfef-3a04-4363-aa30-7f75a88a35e9-kube-api-access-qwhq4" (OuterVolumeSpecName: "kube-api-access-qwhq4") pod "f167bfef-3a04-4363-aa30-7f75a88a35e9" (UID: "f167bfef-3a04-4363-aa30-7f75a88a35e9"). InnerVolumeSpecName "kube-api-access-qwhq4". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 07:52:13 crc kubenswrapper[4612]: I0227 07:52:13.943915 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b4a05b89-0ef2-4f07-b1de-f558a36415a3-kube-api-access-rkh9l" (OuterVolumeSpecName: "kube-api-access-rkh9l") pod "b4a05b89-0ef2-4f07-b1de-f558a36415a3" (UID: "b4a05b89-0ef2-4f07-b1de-f558a36415a3"). InnerVolumeSpecName "kube-api-access-rkh9l". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 07:52:13 crc kubenswrapper[4612]: I0227 07:52:13.989480 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f167bfef-3a04-4363-aa30-7f75a88a35e9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f167bfef-3a04-4363-aa30-7f75a88a35e9" (UID: "f167bfef-3a04-4363-aa30-7f75a88a35e9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 07:52:13 crc kubenswrapper[4612]: I0227 07:52:13.998967 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b4a05b89-0ef2-4f07-b1de-f558a36415a3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b4a05b89-0ef2-4f07-b1de-f558a36415a3" (UID: "b4a05b89-0ef2-4f07-b1de-f558a36415a3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 07:52:14 crc kubenswrapper[4612]: I0227 07:52:14.025106 4612 generic.go:334] "Generic (PLEG): container finished" podID="f167bfef-3a04-4363-aa30-7f75a88a35e9" containerID="da9a4aee21ae34fe8a6ee2c2a89a45943fcf57d05dec12aeadfc759f1fe2d89a" exitCode=0 Feb 27 07:52:14 crc kubenswrapper[4612]: I0227 07:52:14.025168 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qcsjx" event={"ID":"f167bfef-3a04-4363-aa30-7f75a88a35e9","Type":"ContainerDied","Data":"da9a4aee21ae34fe8a6ee2c2a89a45943fcf57d05dec12aeadfc759f1fe2d89a"} Feb 27 07:52:14 crc kubenswrapper[4612]: I0227 07:52:14.025194 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qcsjx" event={"ID":"f167bfef-3a04-4363-aa30-7f75a88a35e9","Type":"ContainerDied","Data":"95e30b402c583c3375346400e6dd10960d5f7f6b2bab0ceac8b32fa3e35e5ad8"} Feb 27 07:52:14 crc kubenswrapper[4612]: I0227 07:52:14.025211 4612 scope.go:117] "RemoveContainer" containerID="da9a4aee21ae34fe8a6ee2c2a89a45943fcf57d05dec12aeadfc759f1fe2d89a" Feb 27 07:52:14 crc kubenswrapper[4612]: I0227 07:52:14.025311 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-qcsjx" Feb 27 07:52:14 crc kubenswrapper[4612]: I0227 07:52:14.031083 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536312-tjtt6" event={"ID":"396f1b18-f225-4d67-a98b-87b2605406d4","Type":"ContainerStarted","Data":"fe48465b2a2f1e8415bd353b3cca066d922ee14bd35e47161e5458f20f4f96d3"} Feb 27 07:52:14 crc kubenswrapper[4612]: I0227 07:52:14.035036 4612 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b4a05b89-0ef2-4f07-b1de-f558a36415a3-utilities\") on node \"crc\" DevicePath \"\"" Feb 27 07:52:14 crc kubenswrapper[4612]: I0227 07:52:14.035072 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rkh9l\" (UniqueName: \"kubernetes.io/projected/b4a05b89-0ef2-4f07-b1de-f558a36415a3-kube-api-access-rkh9l\") on node \"crc\" DevicePath \"\"" Feb 27 07:52:14 crc kubenswrapper[4612]: I0227 07:52:14.035087 4612 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f167bfef-3a04-4363-aa30-7f75a88a35e9-utilities\") on node \"crc\" DevicePath \"\"" Feb 27 07:52:14 crc kubenswrapper[4612]: I0227 07:52:14.035097 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qwhq4\" (UniqueName: \"kubernetes.io/projected/f167bfef-3a04-4363-aa30-7f75a88a35e9-kube-api-access-qwhq4\") on node \"crc\" DevicePath \"\"" Feb 27 07:52:14 crc kubenswrapper[4612]: I0227 07:52:14.035105 4612 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f167bfef-3a04-4363-aa30-7f75a88a35e9-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 27 07:52:14 crc kubenswrapper[4612]: I0227 07:52:14.035114 4612 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b4a05b89-0ef2-4f07-b1de-f558a36415a3-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 27 07:52:14 crc kubenswrapper[4612]: I0227 07:52:14.038161 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-99cd5" Feb 27 07:52:14 crc kubenswrapper[4612]: I0227 07:52:14.038421 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-99cd5" event={"ID":"b4a05b89-0ef2-4f07-b1de-f558a36415a3","Type":"ContainerDied","Data":"00c5fe7ac50ab88468386ef0cc47f1da9a51457ee9170cc7305347b0d2b3a8cb"} Feb 27 07:52:14 crc kubenswrapper[4612]: I0227 07:52:14.047745 4612 scope.go:117] "RemoveContainer" containerID="9af8303237b79e77bd66b699c83411679cb8e1b91e317bf7370a2c3bab78e75f" Feb 27 07:52:14 crc kubenswrapper[4612]: I0227 07:52:14.050560 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29536312-tjtt6" podStartSLOduration=1.2688280889999999 podStartE2EDuration="14.05054508s" podCreationTimestamp="2026-02-27 07:52:00 +0000 UTC" firstStartedPulling="2026-02-27 07:52:00.832791733 +0000 UTC m=+178.686721731" lastFinishedPulling="2026-02-27 07:52:13.614508724 +0000 UTC m=+191.468438722" observedRunningTime="2026-02-27 07:52:14.048763341 +0000 UTC m=+191.902693349" watchObservedRunningTime="2026-02-27 07:52:14.05054508 +0000 UTC m=+191.904475068" Feb 27 07:52:14 crc kubenswrapper[4612]: I0227 07:52:14.068615 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-qcsjx"] Feb 27 07:52:14 crc kubenswrapper[4612]: I0227 07:52:14.069663 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-qcsjx"] Feb 27 07:52:14 crc kubenswrapper[4612]: I0227 07:52:14.075626 4612 scope.go:117] "RemoveContainer" containerID="28c5a083b3d3ea88b579e2dec6c6a702efbf9dbaa452442b29f513993fe5cd55" Feb 27 07:52:14 crc kubenswrapper[4612]: I0227 07:52:14.090413 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-99cd5"] Feb 27 07:52:14 crc kubenswrapper[4612]: I0227 07:52:14.092102 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-99cd5"] Feb 27 07:52:14 crc kubenswrapper[4612]: I0227 07:52:14.099278 4612 scope.go:117] "RemoveContainer" containerID="da9a4aee21ae34fe8a6ee2c2a89a45943fcf57d05dec12aeadfc759f1fe2d89a" Feb 27 07:52:14 crc kubenswrapper[4612]: E0227 07:52:14.099803 4612 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"da9a4aee21ae34fe8a6ee2c2a89a45943fcf57d05dec12aeadfc759f1fe2d89a\": container with ID starting with da9a4aee21ae34fe8a6ee2c2a89a45943fcf57d05dec12aeadfc759f1fe2d89a not found: ID does not exist" containerID="da9a4aee21ae34fe8a6ee2c2a89a45943fcf57d05dec12aeadfc759f1fe2d89a" Feb 27 07:52:14 crc kubenswrapper[4612]: I0227 07:52:14.099844 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"da9a4aee21ae34fe8a6ee2c2a89a45943fcf57d05dec12aeadfc759f1fe2d89a"} err="failed to get container status \"da9a4aee21ae34fe8a6ee2c2a89a45943fcf57d05dec12aeadfc759f1fe2d89a\": rpc error: code = NotFound desc = could not find container \"da9a4aee21ae34fe8a6ee2c2a89a45943fcf57d05dec12aeadfc759f1fe2d89a\": container with ID starting with da9a4aee21ae34fe8a6ee2c2a89a45943fcf57d05dec12aeadfc759f1fe2d89a not found: ID does not exist" Feb 27 07:52:14 crc kubenswrapper[4612]: I0227 07:52:14.099874 4612 scope.go:117] "RemoveContainer" containerID="9af8303237b79e77bd66b699c83411679cb8e1b91e317bf7370a2c3bab78e75f" Feb 27 07:52:14 crc kubenswrapper[4612]: E0227 07:52:14.100231 4612 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9af8303237b79e77bd66b699c83411679cb8e1b91e317bf7370a2c3bab78e75f\": container with ID starting with 9af8303237b79e77bd66b699c83411679cb8e1b91e317bf7370a2c3bab78e75f not found: ID does not exist" containerID="9af8303237b79e77bd66b699c83411679cb8e1b91e317bf7370a2c3bab78e75f" Feb 27 07:52:14 crc kubenswrapper[4612]: I0227 07:52:14.100256 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-fjg45" Feb 27 07:52:14 crc kubenswrapper[4612]: I0227 07:52:14.100256 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9af8303237b79e77bd66b699c83411679cb8e1b91e317bf7370a2c3bab78e75f"} err="failed to get container status \"9af8303237b79e77bd66b699c83411679cb8e1b91e317bf7370a2c3bab78e75f\": rpc error: code = NotFound desc = could not find container \"9af8303237b79e77bd66b699c83411679cb8e1b91e317bf7370a2c3bab78e75f\": container with ID starting with 9af8303237b79e77bd66b699c83411679cb8e1b91e317bf7370a2c3bab78e75f not found: ID does not exist" Feb 27 07:52:14 crc kubenswrapper[4612]: I0227 07:52:14.100280 4612 scope.go:117] "RemoveContainer" containerID="28c5a083b3d3ea88b579e2dec6c6a702efbf9dbaa452442b29f513993fe5cd55" Feb 27 07:52:14 crc kubenswrapper[4612]: E0227 07:52:14.100742 4612 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"28c5a083b3d3ea88b579e2dec6c6a702efbf9dbaa452442b29f513993fe5cd55\": container with ID starting with 28c5a083b3d3ea88b579e2dec6c6a702efbf9dbaa452442b29f513993fe5cd55 not found: ID does not exist" containerID="28c5a083b3d3ea88b579e2dec6c6a702efbf9dbaa452442b29f513993fe5cd55" Feb 27 07:52:14 crc kubenswrapper[4612]: I0227 07:52:14.100802 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"28c5a083b3d3ea88b579e2dec6c6a702efbf9dbaa452442b29f513993fe5cd55"} err="failed to get container status \"28c5a083b3d3ea88b579e2dec6c6a702efbf9dbaa452442b29f513993fe5cd55\": rpc error: code = NotFound desc = could not find container \"28c5a083b3d3ea88b579e2dec6c6a702efbf9dbaa452442b29f513993fe5cd55\": container with ID starting with 28c5a083b3d3ea88b579e2dec6c6a702efbf9dbaa452442b29f513993fe5cd55 not found: ID does not exist" Feb 27 07:52:14 crc kubenswrapper[4612]: I0227 07:52:14.100825 4612 scope.go:117] "RemoveContainer" containerID="b9fd648455a1f23c8fe4b0657e89196db987d450104c1ad5ea1c2b53f4c669bd" Feb 27 07:52:14 crc kubenswrapper[4612]: I0227 07:52:14.117394 4612 scope.go:117] "RemoveContainer" containerID="ebd705cf2e79783dc207f1bd1f410ff3ed40b1b60e3eaa30a6dd2942e0dd3980" Feb 27 07:52:14 crc kubenswrapper[4612]: I0227 07:52:14.142791 4612 scope.go:117] "RemoveContainer" containerID="8ac8333061750bbd7f5d78a4ef26f5753f05fb46272b5a603b24767136bc54ac" Feb 27 07:52:14 crc kubenswrapper[4612]: I0227 07:52:14.304611 4612 csr.go:261] certificate signing request csr-x9x9f is approved, waiting to be issued Feb 27 07:52:14 crc kubenswrapper[4612]: I0227 07:52:14.310353 4612 csr.go:257] certificate signing request csr-x9x9f is issued Feb 27 07:52:14 crc kubenswrapper[4612]: I0227 07:52:14.860167 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b4a05b89-0ef2-4f07-b1de-f558a36415a3" path="/var/lib/kubelet/pods/b4a05b89-0ef2-4f07-b1de-f558a36415a3/volumes" Feb 27 07:52:14 crc kubenswrapper[4612]: I0227 07:52:14.860912 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f167bfef-3a04-4363-aa30-7f75a88a35e9" path="/var/lib/kubelet/pods/f167bfef-3a04-4363-aa30-7f75a88a35e9/volumes" Feb 27 07:52:15 crc kubenswrapper[4612]: I0227 07:52:15.045420 4612 generic.go:334] "Generic (PLEG): container finished" podID="396f1b18-f225-4d67-a98b-87b2605406d4" containerID="fe48465b2a2f1e8415bd353b3cca066d922ee14bd35e47161e5458f20f4f96d3" exitCode=0 Feb 27 07:52:15 crc kubenswrapper[4612]: I0227 07:52:15.046025 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536312-tjtt6" event={"ID":"396f1b18-f225-4d67-a98b-87b2605406d4","Type":"ContainerDied","Data":"fe48465b2a2f1e8415bd353b3cca066d922ee14bd35e47161e5458f20f4f96d3"} Feb 27 07:52:15 crc kubenswrapper[4612]: E0227 07:52:15.088335 4612 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcc6862b7_a2ab_4051_9d83_17774bfc5ac7.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod458167a3_25b3_4cd1_8628_865ad916d3a6.slice/crio-47ac8b300890eb2dbe6d5ee7578f5f522e8820a79526ada246c4cb452b44b1ad\": RecentStats: unable to find data in memory cache]" Feb 27 07:52:15 crc kubenswrapper[4612]: I0227 07:52:15.311507 4612 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2027-02-24 05:54:36 +0000 UTC, rotation deadline is 2026-12-10 22:44:08.35382704 +0000 UTC Feb 27 07:52:15 crc kubenswrapper[4612]: I0227 07:52:15.311542 4612 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 6878h51m53.042286952s for next certificate rotation Feb 27 07:52:16 crc kubenswrapper[4612]: I0227 07:52:16.151859 4612 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-nkbg2" podUID="6ece12fd-d439-4104-a7a9-a9d174e29b5a" containerName="oauth-openshift" containerID="cri-o://337d6f75cb558f81002522a58efc2f5296e04ef640a355c38af791aaf38647a3" gracePeriod=15 Feb 27 07:52:16 crc kubenswrapper[4612]: I0227 07:52:16.320756 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536312-tjtt6" Feb 27 07:52:16 crc kubenswrapper[4612]: I0227 07:52:16.375282 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cdpst\" (UniqueName: \"kubernetes.io/projected/396f1b18-f225-4d67-a98b-87b2605406d4-kube-api-access-cdpst\") pod \"396f1b18-f225-4d67-a98b-87b2605406d4\" (UID: \"396f1b18-f225-4d67-a98b-87b2605406d4\") " Feb 27 07:52:16 crc kubenswrapper[4612]: I0227 07:52:16.384267 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/396f1b18-f225-4d67-a98b-87b2605406d4-kube-api-access-cdpst" (OuterVolumeSpecName: "kube-api-access-cdpst") pod "396f1b18-f225-4d67-a98b-87b2605406d4" (UID: "396f1b18-f225-4d67-a98b-87b2605406d4"). InnerVolumeSpecName "kube-api-access-cdpst". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 07:52:16 crc kubenswrapper[4612]: I0227 07:52:16.476788 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cdpst\" (UniqueName: \"kubernetes.io/projected/396f1b18-f225-4d67-a98b-87b2605406d4-kube-api-access-cdpst\") on node \"crc\" DevicePath \"\"" Feb 27 07:52:16 crc kubenswrapper[4612]: I0227 07:52:16.505986 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-nkbg2" Feb 27 07:52:16 crc kubenswrapper[4612]: I0227 07:52:16.577287 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/6ece12fd-d439-4104-a7a9-a9d174e29b5a-v4-0-config-system-session\") pod \"6ece12fd-d439-4104-a7a9-a9d174e29b5a\" (UID: \"6ece12fd-d439-4104-a7a9-a9d174e29b5a\") " Feb 27 07:52:16 crc kubenswrapper[4612]: I0227 07:52:16.577332 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf2ls\" (UniqueName: \"kubernetes.io/projected/6ece12fd-d439-4104-a7a9-a9d174e29b5a-kube-api-access-gf2ls\") pod \"6ece12fd-d439-4104-a7a9-a9d174e29b5a\" (UID: \"6ece12fd-d439-4104-a7a9-a9d174e29b5a\") " Feb 27 07:52:16 crc kubenswrapper[4612]: I0227 07:52:16.577365 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/6ece12fd-d439-4104-a7a9-a9d174e29b5a-v4-0-config-system-service-ca\") pod \"6ece12fd-d439-4104-a7a9-a9d174e29b5a\" (UID: \"6ece12fd-d439-4104-a7a9-a9d174e29b5a\") " Feb 27 07:52:16 crc kubenswrapper[4612]: I0227 07:52:16.577392 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/6ece12fd-d439-4104-a7a9-a9d174e29b5a-v4-0-config-user-idp-0-file-data\") pod \"6ece12fd-d439-4104-a7a9-a9d174e29b5a\" (UID: \"6ece12fd-d439-4104-a7a9-a9d174e29b5a\") " Feb 27 07:52:16 crc kubenswrapper[4612]: I0227 07:52:16.577423 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/6ece12fd-d439-4104-a7a9-a9d174e29b5a-audit-policies\") pod \"6ece12fd-d439-4104-a7a9-a9d174e29b5a\" (UID: \"6ece12fd-d439-4104-a7a9-a9d174e29b5a\") " Feb 27 07:52:16 crc kubenswrapper[4612]: I0227 07:52:16.577440 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/6ece12fd-d439-4104-a7a9-a9d174e29b5a-v4-0-config-system-serving-cert\") pod \"6ece12fd-d439-4104-a7a9-a9d174e29b5a\" (UID: \"6ece12fd-d439-4104-a7a9-a9d174e29b5a\") " Feb 27 07:52:16 crc kubenswrapper[4612]: I0227 07:52:16.577457 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/6ece12fd-d439-4104-a7a9-a9d174e29b5a-v4-0-config-user-template-provider-selection\") pod \"6ece12fd-d439-4104-a7a9-a9d174e29b5a\" (UID: \"6ece12fd-d439-4104-a7a9-a9d174e29b5a\") " Feb 27 07:52:16 crc kubenswrapper[4612]: I0227 07:52:16.577475 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/6ece12fd-d439-4104-a7a9-a9d174e29b5a-audit-dir\") pod \"6ece12fd-d439-4104-a7a9-a9d174e29b5a\" (UID: \"6ece12fd-d439-4104-a7a9-a9d174e29b5a\") " Feb 27 07:52:16 crc kubenswrapper[4612]: I0227 07:52:16.577499 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/6ece12fd-d439-4104-a7a9-a9d174e29b5a-v4-0-config-system-router-certs\") pod \"6ece12fd-d439-4104-a7a9-a9d174e29b5a\" (UID: \"6ece12fd-d439-4104-a7a9-a9d174e29b5a\") " Feb 27 07:52:16 crc kubenswrapper[4612]: I0227 07:52:16.577515 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/6ece12fd-d439-4104-a7a9-a9d174e29b5a-v4-0-config-system-cliconfig\") pod \"6ece12fd-d439-4104-a7a9-a9d174e29b5a\" (UID: \"6ece12fd-d439-4104-a7a9-a9d174e29b5a\") " Feb 27 07:52:16 crc kubenswrapper[4612]: I0227 07:52:16.577541 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/6ece12fd-d439-4104-a7a9-a9d174e29b5a-v4-0-config-user-template-error\") pod \"6ece12fd-d439-4104-a7a9-a9d174e29b5a\" (UID: \"6ece12fd-d439-4104-a7a9-a9d174e29b5a\") " Feb 27 07:52:16 crc kubenswrapper[4612]: I0227 07:52:16.577569 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/6ece12fd-d439-4104-a7a9-a9d174e29b5a-v4-0-config-user-template-login\") pod \"6ece12fd-d439-4104-a7a9-a9d174e29b5a\" (UID: \"6ece12fd-d439-4104-a7a9-a9d174e29b5a\") " Feb 27 07:52:16 crc kubenswrapper[4612]: I0227 07:52:16.577596 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6ece12fd-d439-4104-a7a9-a9d174e29b5a-v4-0-config-system-trusted-ca-bundle\") pod \"6ece12fd-d439-4104-a7a9-a9d174e29b5a\" (UID: \"6ece12fd-d439-4104-a7a9-a9d174e29b5a\") " Feb 27 07:52:16 crc kubenswrapper[4612]: I0227 07:52:16.577627 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/6ece12fd-d439-4104-a7a9-a9d174e29b5a-v4-0-config-system-ocp-branding-template\") pod \"6ece12fd-d439-4104-a7a9-a9d174e29b5a\" (UID: \"6ece12fd-d439-4104-a7a9-a9d174e29b5a\") " Feb 27 07:52:16 crc kubenswrapper[4612]: I0227 07:52:16.578572 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6ece12fd-d439-4104-a7a9-a9d174e29b5a-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "6ece12fd-d439-4104-a7a9-a9d174e29b5a" (UID: "6ece12fd-d439-4104-a7a9-a9d174e29b5a"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 27 07:52:16 crc kubenswrapper[4612]: I0227 07:52:16.579075 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ece12fd-d439-4104-a7a9-a9d174e29b5a-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "6ece12fd-d439-4104-a7a9-a9d174e29b5a" (UID: "6ece12fd-d439-4104-a7a9-a9d174e29b5a"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 07:52:16 crc kubenswrapper[4612]: I0227 07:52:16.579167 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ece12fd-d439-4104-a7a9-a9d174e29b5a-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "6ece12fd-d439-4104-a7a9-a9d174e29b5a" (UID: "6ece12fd-d439-4104-a7a9-a9d174e29b5a"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 07:52:16 crc kubenswrapper[4612]: I0227 07:52:16.579237 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ece12fd-d439-4104-a7a9-a9d174e29b5a-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "6ece12fd-d439-4104-a7a9-a9d174e29b5a" (UID: "6ece12fd-d439-4104-a7a9-a9d174e29b5a"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 07:52:16 crc kubenswrapper[4612]: I0227 07:52:16.579417 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ece12fd-d439-4104-a7a9-a9d174e29b5a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "6ece12fd-d439-4104-a7a9-a9d174e29b5a" (UID: "6ece12fd-d439-4104-a7a9-a9d174e29b5a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 07:52:16 crc kubenswrapper[4612]: I0227 07:52:16.581480 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ece12fd-d439-4104-a7a9-a9d174e29b5a-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "6ece12fd-d439-4104-a7a9-a9d174e29b5a" (UID: "6ece12fd-d439-4104-a7a9-a9d174e29b5a"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 07:52:16 crc kubenswrapper[4612]: I0227 07:52:16.582011 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ece12fd-d439-4104-a7a9-a9d174e29b5a-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "6ece12fd-d439-4104-a7a9-a9d174e29b5a" (UID: "6ece12fd-d439-4104-a7a9-a9d174e29b5a"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 07:52:16 crc kubenswrapper[4612]: I0227 07:52:16.582634 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ece12fd-d439-4104-a7a9-a9d174e29b5a-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "6ece12fd-d439-4104-a7a9-a9d174e29b5a" (UID: "6ece12fd-d439-4104-a7a9-a9d174e29b5a"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 07:52:16 crc kubenswrapper[4612]: I0227 07:52:16.582850 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ece12fd-d439-4104-a7a9-a9d174e29b5a-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "6ece12fd-d439-4104-a7a9-a9d174e29b5a" (UID: "6ece12fd-d439-4104-a7a9-a9d174e29b5a"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 07:52:16 crc kubenswrapper[4612]: I0227 07:52:16.583018 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ece12fd-d439-4104-a7a9-a9d174e29b5a-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "6ece12fd-d439-4104-a7a9-a9d174e29b5a" (UID: "6ece12fd-d439-4104-a7a9-a9d174e29b5a"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 07:52:16 crc kubenswrapper[4612]: I0227 07:52:16.583604 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ece12fd-d439-4104-a7a9-a9d174e29b5a-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "6ece12fd-d439-4104-a7a9-a9d174e29b5a" (UID: "6ece12fd-d439-4104-a7a9-a9d174e29b5a"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 07:52:16 crc kubenswrapper[4612]: I0227 07:52:16.583833 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ece12fd-d439-4104-a7a9-a9d174e29b5a-kube-api-access-gf2ls" (OuterVolumeSpecName: "kube-api-access-gf2ls") pod "6ece12fd-d439-4104-a7a9-a9d174e29b5a" (UID: "6ece12fd-d439-4104-a7a9-a9d174e29b5a"). InnerVolumeSpecName "kube-api-access-gf2ls". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 07:52:16 crc kubenswrapper[4612]: I0227 07:52:16.583880 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ece12fd-d439-4104-a7a9-a9d174e29b5a-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "6ece12fd-d439-4104-a7a9-a9d174e29b5a" (UID: "6ece12fd-d439-4104-a7a9-a9d174e29b5a"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 07:52:16 crc kubenswrapper[4612]: I0227 07:52:16.584206 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ece12fd-d439-4104-a7a9-a9d174e29b5a-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "6ece12fd-d439-4104-a7a9-a9d174e29b5a" (UID: "6ece12fd-d439-4104-a7a9-a9d174e29b5a"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 07:52:16 crc kubenswrapper[4612]: I0227 07:52:16.679535 4612 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/6ece12fd-d439-4104-a7a9-a9d174e29b5a-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Feb 27 07:52:16 crc kubenswrapper[4612]: I0227 07:52:16.679581 4612 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/6ece12fd-d439-4104-a7a9-a9d174e29b5a-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Feb 27 07:52:16 crc kubenswrapper[4612]: I0227 07:52:16.679597 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf2ls\" (UniqueName: \"kubernetes.io/projected/6ece12fd-d439-4104-a7a9-a9d174e29b5a-kube-api-access-gf2ls\") on node \"crc\" DevicePath \"\"" Feb 27 07:52:16 crc kubenswrapper[4612]: I0227 07:52:16.679609 4612 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/6ece12fd-d439-4104-a7a9-a9d174e29b5a-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Feb 27 07:52:16 crc kubenswrapper[4612]: I0227 07:52:16.679622 4612 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/6ece12fd-d439-4104-a7a9-a9d174e29b5a-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Feb 27 07:52:16 crc kubenswrapper[4612]: I0227 07:52:16.679636 4612 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/6ece12fd-d439-4104-a7a9-a9d174e29b5a-audit-policies\") on node \"crc\" DevicePath \"\"" Feb 27 07:52:16 crc kubenswrapper[4612]: I0227 07:52:16.679736 4612 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/6ece12fd-d439-4104-a7a9-a9d174e29b5a-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 27 07:52:16 crc kubenswrapper[4612]: I0227 07:52:16.679754 4612 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/6ece12fd-d439-4104-a7a9-a9d174e29b5a-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Feb 27 07:52:16 crc kubenswrapper[4612]: I0227 07:52:16.679768 4612 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/6ece12fd-d439-4104-a7a9-a9d174e29b5a-audit-dir\") on node \"crc\" DevicePath \"\"" Feb 27 07:52:16 crc kubenswrapper[4612]: I0227 07:52:16.679779 4612 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/6ece12fd-d439-4104-a7a9-a9d174e29b5a-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Feb 27 07:52:16 crc kubenswrapper[4612]: I0227 07:52:16.679796 4612 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/6ece12fd-d439-4104-a7a9-a9d174e29b5a-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Feb 27 07:52:16 crc kubenswrapper[4612]: I0227 07:52:16.679809 4612 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/6ece12fd-d439-4104-a7a9-a9d174e29b5a-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Feb 27 07:52:16 crc kubenswrapper[4612]: I0227 07:52:16.679821 4612 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/6ece12fd-d439-4104-a7a9-a9d174e29b5a-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Feb 27 07:52:16 crc kubenswrapper[4612]: I0227 07:52:16.679833 4612 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6ece12fd-d439-4104-a7a9-a9d174e29b5a-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 07:52:16 crc kubenswrapper[4612]: I0227 07:52:16.765911 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-fjg45"] Feb 27 07:52:16 crc kubenswrapper[4612]: I0227 07:52:16.766168 4612 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-fjg45" podUID="7676c2d7-9df2-4dc8-a2bd-cde398845d26" containerName="registry-server" containerID="cri-o://56496367db4cbd067458a7e859e26159545b5d2a06fb533402516e4f5ab15d97" gracePeriod=2 Feb 27 07:52:17 crc kubenswrapper[4612]: I0227 07:52:17.056534 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536312-tjtt6" event={"ID":"396f1b18-f225-4d67-a98b-87b2605406d4","Type":"ContainerDied","Data":"e08fa62afab00c7952b1722ceda0a294f75a2b6c33cdb3d25481ef9313fbde90"} Feb 27 07:52:17 crc kubenswrapper[4612]: I0227 07:52:17.056585 4612 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e08fa62afab00c7952b1722ceda0a294f75a2b6c33cdb3d25481ef9313fbde90" Feb 27 07:52:17 crc kubenswrapper[4612]: I0227 07:52:17.056658 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536312-tjtt6" Feb 27 07:52:17 crc kubenswrapper[4612]: I0227 07:52:17.063977 4612 generic.go:334] "Generic (PLEG): container finished" podID="6ece12fd-d439-4104-a7a9-a9d174e29b5a" containerID="337d6f75cb558f81002522a58efc2f5296e04ef640a355c38af791aaf38647a3" exitCode=0 Feb 27 07:52:17 crc kubenswrapper[4612]: I0227 07:52:17.064056 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-nkbg2" Feb 27 07:52:17 crc kubenswrapper[4612]: I0227 07:52:17.064064 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-nkbg2" event={"ID":"6ece12fd-d439-4104-a7a9-a9d174e29b5a","Type":"ContainerDied","Data":"337d6f75cb558f81002522a58efc2f5296e04ef640a355c38af791aaf38647a3"} Feb 27 07:52:17 crc kubenswrapper[4612]: I0227 07:52:17.064167 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-nkbg2" event={"ID":"6ece12fd-d439-4104-a7a9-a9d174e29b5a","Type":"ContainerDied","Data":"2142c9ee66879280306a4cce8c69235d9b9acac78defb5151ef4f7e8961cd22e"} Feb 27 07:52:17 crc kubenswrapper[4612]: I0227 07:52:17.064185 4612 scope.go:117] "RemoveContainer" containerID="337d6f75cb558f81002522a58efc2f5296e04ef640a355c38af791aaf38647a3" Feb 27 07:52:17 crc kubenswrapper[4612]: I0227 07:52:17.067570 4612 generic.go:334] "Generic (PLEG): container finished" podID="7676c2d7-9df2-4dc8-a2bd-cde398845d26" containerID="56496367db4cbd067458a7e859e26159545b5d2a06fb533402516e4f5ab15d97" exitCode=0 Feb 27 07:52:17 crc kubenswrapper[4612]: I0227 07:52:17.067605 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fjg45" event={"ID":"7676c2d7-9df2-4dc8-a2bd-cde398845d26","Type":"ContainerDied","Data":"56496367db4cbd067458a7e859e26159545b5d2a06fb533402516e4f5ab15d97"} Feb 27 07:52:17 crc kubenswrapper[4612]: I0227 07:52:17.091385 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-nkbg2"] Feb 27 07:52:17 crc kubenswrapper[4612]: I0227 07:52:17.091964 4612 scope.go:117] "RemoveContainer" containerID="337d6f75cb558f81002522a58efc2f5296e04ef640a355c38af791aaf38647a3" Feb 27 07:52:17 crc kubenswrapper[4612]: E0227 07:52:17.092472 4612 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"337d6f75cb558f81002522a58efc2f5296e04ef640a355c38af791aaf38647a3\": container with ID starting with 337d6f75cb558f81002522a58efc2f5296e04ef640a355c38af791aaf38647a3 not found: ID does not exist" containerID="337d6f75cb558f81002522a58efc2f5296e04ef640a355c38af791aaf38647a3" Feb 27 07:52:17 crc kubenswrapper[4612]: I0227 07:52:17.092514 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"337d6f75cb558f81002522a58efc2f5296e04ef640a355c38af791aaf38647a3"} err="failed to get container status \"337d6f75cb558f81002522a58efc2f5296e04ef640a355c38af791aaf38647a3\": rpc error: code = NotFound desc = could not find container \"337d6f75cb558f81002522a58efc2f5296e04ef640a355c38af791aaf38647a3\": container with ID starting with 337d6f75cb558f81002522a58efc2f5296e04ef640a355c38af791aaf38647a3 not found: ID does not exist" Feb 27 07:52:17 crc kubenswrapper[4612]: I0227 07:52:17.095833 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-nkbg2"] Feb 27 07:52:17 crc kubenswrapper[4612]: I0227 07:52:17.112330 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-fjg45" Feb 27 07:52:17 crc kubenswrapper[4612]: I0227 07:52:17.189117 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7676c2d7-9df2-4dc8-a2bd-cde398845d26-utilities\") pod \"7676c2d7-9df2-4dc8-a2bd-cde398845d26\" (UID: \"7676c2d7-9df2-4dc8-a2bd-cde398845d26\") " Feb 27 07:52:17 crc kubenswrapper[4612]: I0227 07:52:17.189255 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-97g6s\" (UniqueName: \"kubernetes.io/projected/7676c2d7-9df2-4dc8-a2bd-cde398845d26-kube-api-access-97g6s\") pod \"7676c2d7-9df2-4dc8-a2bd-cde398845d26\" (UID: \"7676c2d7-9df2-4dc8-a2bd-cde398845d26\") " Feb 27 07:52:17 crc kubenswrapper[4612]: I0227 07:52:17.189309 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7676c2d7-9df2-4dc8-a2bd-cde398845d26-catalog-content\") pod \"7676c2d7-9df2-4dc8-a2bd-cde398845d26\" (UID: \"7676c2d7-9df2-4dc8-a2bd-cde398845d26\") " Feb 27 07:52:17 crc kubenswrapper[4612]: I0227 07:52:17.190006 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7676c2d7-9df2-4dc8-a2bd-cde398845d26-utilities" (OuterVolumeSpecName: "utilities") pod "7676c2d7-9df2-4dc8-a2bd-cde398845d26" (UID: "7676c2d7-9df2-4dc8-a2bd-cde398845d26"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 07:52:17 crc kubenswrapper[4612]: I0227 07:52:17.192585 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7676c2d7-9df2-4dc8-a2bd-cde398845d26-kube-api-access-97g6s" (OuterVolumeSpecName: "kube-api-access-97g6s") pod "7676c2d7-9df2-4dc8-a2bd-cde398845d26" (UID: "7676c2d7-9df2-4dc8-a2bd-cde398845d26"). InnerVolumeSpecName "kube-api-access-97g6s". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 07:52:17 crc kubenswrapper[4612]: I0227 07:52:17.290811 4612 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7676c2d7-9df2-4dc8-a2bd-cde398845d26-utilities\") on node \"crc\" DevicePath \"\"" Feb 27 07:52:17 crc kubenswrapper[4612]: I0227 07:52:17.291009 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-97g6s\" (UniqueName: \"kubernetes.io/projected/7676c2d7-9df2-4dc8-a2bd-cde398845d26-kube-api-access-97g6s\") on node \"crc\" DevicePath \"\"" Feb 27 07:52:17 crc kubenswrapper[4612]: I0227 07:52:17.301903 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7676c2d7-9df2-4dc8-a2bd-cde398845d26-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7676c2d7-9df2-4dc8-a2bd-cde398845d26" (UID: "7676c2d7-9df2-4dc8-a2bd-cde398845d26"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 07:52:17 crc kubenswrapper[4612]: I0227 07:52:17.392832 4612 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7676c2d7-9df2-4dc8-a2bd-cde398845d26-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 27 07:52:18 crc kubenswrapper[4612]: I0227 07:52:18.085659 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fjg45" event={"ID":"7676c2d7-9df2-4dc8-a2bd-cde398845d26","Type":"ContainerDied","Data":"5dd86e2dd7860913a54b819f2b93c1ae9b8d82b3e622ad3d886a97a478d55778"} Feb 27 07:52:18 crc kubenswrapper[4612]: I0227 07:52:18.085798 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-fjg45" Feb 27 07:52:18 crc kubenswrapper[4612]: I0227 07:52:18.086089 4612 scope.go:117] "RemoveContainer" containerID="56496367db4cbd067458a7e859e26159545b5d2a06fb533402516e4f5ab15d97" Feb 27 07:52:18 crc kubenswrapper[4612]: I0227 07:52:18.101334 4612 scope.go:117] "RemoveContainer" containerID="9412e39c97cfdbf2ecdea3e1130977db6d47a3576f38d0de4ca96e6dfdf8d3ed" Feb 27 07:52:18 crc kubenswrapper[4612]: I0227 07:52:18.117454 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-fjg45"] Feb 27 07:52:18 crc kubenswrapper[4612]: I0227 07:52:18.120680 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-fjg45"] Feb 27 07:52:18 crc kubenswrapper[4612]: I0227 07:52:18.121530 4612 scope.go:117] "RemoveContainer" containerID="f78d4b506918d6b031a0ab9625bd103a819e4528a72a33626159b8fc4213bfd8" Feb 27 07:52:18 crc kubenswrapper[4612]: I0227 07:52:18.861092 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ece12fd-d439-4104-a7a9-a9d174e29b5a" path="/var/lib/kubelet/pods/6ece12fd-d439-4104-a7a9-a9d174e29b5a/volumes" Feb 27 07:52:18 crc kubenswrapper[4612]: I0227 07:52:18.861784 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7676c2d7-9df2-4dc8-a2bd-cde398845d26" path="/var/lib/kubelet/pods/7676c2d7-9df2-4dc8-a2bd-cde398845d26/volumes" Feb 27 07:52:22 crc kubenswrapper[4612]: I0227 07:52:22.167747 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-59f7dcc9c8-9b7dk"] Feb 27 07:52:22 crc kubenswrapper[4612]: I0227 07:52:22.168313 4612 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-59f7dcc9c8-9b7dk" podUID="c07650a7-6d9b-43da-ae78-e8d8dbe342f4" containerName="controller-manager" containerID="cri-o://8f23ab18cb7541e20fb46e1015f363ad7d720d9b768f2f2fb0d6a861887e5df5" gracePeriod=30 Feb 27 07:52:22 crc kubenswrapper[4612]: I0227 07:52:22.263923 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5969dcdf8-bbhhq"] Feb 27 07:52:22 crc kubenswrapper[4612]: I0227 07:52:22.264176 4612 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-5969dcdf8-bbhhq" podUID="f04702a5-b6df-431b-9f93-a37d9bf5dc96" containerName="route-controller-manager" containerID="cri-o://e67f8de92a83ce39fcf3b74a51fadf0ab4e216d2f00209fd9b8cc92ab0c9e46a" gracePeriod=30 Feb 27 07:52:22 crc kubenswrapper[4612]: I0227 07:52:22.501965 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-59f7dcc9c8-9b7dk" Feb 27 07:52:22 crc kubenswrapper[4612]: I0227 07:52:22.563641 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c07650a7-6d9b-43da-ae78-e8d8dbe342f4-client-ca\") pod \"c07650a7-6d9b-43da-ae78-e8d8dbe342f4\" (UID: \"c07650a7-6d9b-43da-ae78-e8d8dbe342f4\") " Feb 27 07:52:22 crc kubenswrapper[4612]: I0227 07:52:22.563711 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c07650a7-6d9b-43da-ae78-e8d8dbe342f4-proxy-ca-bundles\") pod \"c07650a7-6d9b-43da-ae78-e8d8dbe342f4\" (UID: \"c07650a7-6d9b-43da-ae78-e8d8dbe342f4\") " Feb 27 07:52:22 crc kubenswrapper[4612]: I0227 07:52:22.563809 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c07650a7-6d9b-43da-ae78-e8d8dbe342f4-config\") pod \"c07650a7-6d9b-43da-ae78-e8d8dbe342f4\" (UID: \"c07650a7-6d9b-43da-ae78-e8d8dbe342f4\") " Feb 27 07:52:22 crc kubenswrapper[4612]: I0227 07:52:22.563828 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jd5vk\" (UniqueName: \"kubernetes.io/projected/c07650a7-6d9b-43da-ae78-e8d8dbe342f4-kube-api-access-jd5vk\") pod \"c07650a7-6d9b-43da-ae78-e8d8dbe342f4\" (UID: \"c07650a7-6d9b-43da-ae78-e8d8dbe342f4\") " Feb 27 07:52:22 crc kubenswrapper[4612]: I0227 07:52:22.563870 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c07650a7-6d9b-43da-ae78-e8d8dbe342f4-serving-cert\") pod \"c07650a7-6d9b-43da-ae78-e8d8dbe342f4\" (UID: \"c07650a7-6d9b-43da-ae78-e8d8dbe342f4\") " Feb 27 07:52:22 crc kubenswrapper[4612]: I0227 07:52:22.564485 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c07650a7-6d9b-43da-ae78-e8d8dbe342f4-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "c07650a7-6d9b-43da-ae78-e8d8dbe342f4" (UID: "c07650a7-6d9b-43da-ae78-e8d8dbe342f4"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 07:52:22 crc kubenswrapper[4612]: I0227 07:52:22.564814 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c07650a7-6d9b-43da-ae78-e8d8dbe342f4-client-ca" (OuterVolumeSpecName: "client-ca") pod "c07650a7-6d9b-43da-ae78-e8d8dbe342f4" (UID: "c07650a7-6d9b-43da-ae78-e8d8dbe342f4"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 07:52:22 crc kubenswrapper[4612]: I0227 07:52:22.565617 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c07650a7-6d9b-43da-ae78-e8d8dbe342f4-config" (OuterVolumeSpecName: "config") pod "c07650a7-6d9b-43da-ae78-e8d8dbe342f4" (UID: "c07650a7-6d9b-43da-ae78-e8d8dbe342f4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 07:52:22 crc kubenswrapper[4612]: I0227 07:52:22.569042 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c07650a7-6d9b-43da-ae78-e8d8dbe342f4-kube-api-access-jd5vk" (OuterVolumeSpecName: "kube-api-access-jd5vk") pod "c07650a7-6d9b-43da-ae78-e8d8dbe342f4" (UID: "c07650a7-6d9b-43da-ae78-e8d8dbe342f4"). InnerVolumeSpecName "kube-api-access-jd5vk". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 07:52:22 crc kubenswrapper[4612]: I0227 07:52:22.577724 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c07650a7-6d9b-43da-ae78-e8d8dbe342f4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "c07650a7-6d9b-43da-ae78-e8d8dbe342f4" (UID: "c07650a7-6d9b-43da-ae78-e8d8dbe342f4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 07:52:22 crc kubenswrapper[4612]: I0227 07:52:22.596004 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5969dcdf8-bbhhq" Feb 27 07:52:22 crc kubenswrapper[4612]: I0227 07:52:22.664744 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f04702a5-b6df-431b-9f93-a37d9bf5dc96-config\") pod \"f04702a5-b6df-431b-9f93-a37d9bf5dc96\" (UID: \"f04702a5-b6df-431b-9f93-a37d9bf5dc96\") " Feb 27 07:52:22 crc kubenswrapper[4612]: I0227 07:52:22.664880 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f04702a5-b6df-431b-9f93-a37d9bf5dc96-serving-cert\") pod \"f04702a5-b6df-431b-9f93-a37d9bf5dc96\" (UID: \"f04702a5-b6df-431b-9f93-a37d9bf5dc96\") " Feb 27 07:52:22 crc kubenswrapper[4612]: I0227 07:52:22.664911 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8k8z9\" (UniqueName: \"kubernetes.io/projected/f04702a5-b6df-431b-9f93-a37d9bf5dc96-kube-api-access-8k8z9\") pod \"f04702a5-b6df-431b-9f93-a37d9bf5dc96\" (UID: \"f04702a5-b6df-431b-9f93-a37d9bf5dc96\") " Feb 27 07:52:22 crc kubenswrapper[4612]: I0227 07:52:22.664944 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f04702a5-b6df-431b-9f93-a37d9bf5dc96-client-ca\") pod \"f04702a5-b6df-431b-9f93-a37d9bf5dc96\" (UID: \"f04702a5-b6df-431b-9f93-a37d9bf5dc96\") " Feb 27 07:52:22 crc kubenswrapper[4612]: I0227 07:52:22.665126 4612 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c07650a7-6d9b-43da-ae78-e8d8dbe342f4-config\") on node \"crc\" DevicePath \"\"" Feb 27 07:52:22 crc kubenswrapper[4612]: I0227 07:52:22.665144 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jd5vk\" (UniqueName: \"kubernetes.io/projected/c07650a7-6d9b-43da-ae78-e8d8dbe342f4-kube-api-access-jd5vk\") on node \"crc\" DevicePath \"\"" Feb 27 07:52:22 crc kubenswrapper[4612]: I0227 07:52:22.665153 4612 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c07650a7-6d9b-43da-ae78-e8d8dbe342f4-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 27 07:52:22 crc kubenswrapper[4612]: I0227 07:52:22.665163 4612 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c07650a7-6d9b-43da-ae78-e8d8dbe342f4-client-ca\") on node \"crc\" DevicePath \"\"" Feb 27 07:52:22 crc kubenswrapper[4612]: I0227 07:52:22.665171 4612 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c07650a7-6d9b-43da-ae78-e8d8dbe342f4-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Feb 27 07:52:22 crc kubenswrapper[4612]: I0227 07:52:22.665580 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f04702a5-b6df-431b-9f93-a37d9bf5dc96-config" (OuterVolumeSpecName: "config") pod "f04702a5-b6df-431b-9f93-a37d9bf5dc96" (UID: "f04702a5-b6df-431b-9f93-a37d9bf5dc96"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 07:52:22 crc kubenswrapper[4612]: I0227 07:52:22.665900 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f04702a5-b6df-431b-9f93-a37d9bf5dc96-client-ca" (OuterVolumeSpecName: "client-ca") pod "f04702a5-b6df-431b-9f93-a37d9bf5dc96" (UID: "f04702a5-b6df-431b-9f93-a37d9bf5dc96"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 07:52:22 crc kubenswrapper[4612]: I0227 07:52:22.667802 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f04702a5-b6df-431b-9f93-a37d9bf5dc96-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "f04702a5-b6df-431b-9f93-a37d9bf5dc96" (UID: "f04702a5-b6df-431b-9f93-a37d9bf5dc96"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 07:52:22 crc kubenswrapper[4612]: I0227 07:52:22.667913 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f04702a5-b6df-431b-9f93-a37d9bf5dc96-kube-api-access-8k8z9" (OuterVolumeSpecName: "kube-api-access-8k8z9") pod "f04702a5-b6df-431b-9f93-a37d9bf5dc96" (UID: "f04702a5-b6df-431b-9f93-a37d9bf5dc96"). InnerVolumeSpecName "kube-api-access-8k8z9". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 07:52:22 crc kubenswrapper[4612]: I0227 07:52:22.765846 4612 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f04702a5-b6df-431b-9f93-a37d9bf5dc96-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 27 07:52:22 crc kubenswrapper[4612]: I0227 07:52:22.765872 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8k8z9\" (UniqueName: \"kubernetes.io/projected/f04702a5-b6df-431b-9f93-a37d9bf5dc96-kube-api-access-8k8z9\") on node \"crc\" DevicePath \"\"" Feb 27 07:52:22 crc kubenswrapper[4612]: I0227 07:52:22.765882 4612 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f04702a5-b6df-431b-9f93-a37d9bf5dc96-client-ca\") on node \"crc\" DevicePath \"\"" Feb 27 07:52:22 crc kubenswrapper[4612]: I0227 07:52:22.765890 4612 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f04702a5-b6df-431b-9f93-a37d9bf5dc96-config\") on node \"crc\" DevicePath \"\"" Feb 27 07:52:23 crc kubenswrapper[4612]: I0227 07:52:23.116624 4612 generic.go:334] "Generic (PLEG): container finished" podID="f04702a5-b6df-431b-9f93-a37d9bf5dc96" containerID="e67f8de92a83ce39fcf3b74a51fadf0ab4e216d2f00209fd9b8cc92ab0c9e46a" exitCode=0 Feb 27 07:52:23 crc kubenswrapper[4612]: I0227 07:52:23.116679 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5969dcdf8-bbhhq" event={"ID":"f04702a5-b6df-431b-9f93-a37d9bf5dc96","Type":"ContainerDied","Data":"e67f8de92a83ce39fcf3b74a51fadf0ab4e216d2f00209fd9b8cc92ab0c9e46a"} Feb 27 07:52:23 crc kubenswrapper[4612]: I0227 07:52:23.116723 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5969dcdf8-bbhhq" event={"ID":"f04702a5-b6df-431b-9f93-a37d9bf5dc96","Type":"ContainerDied","Data":"84b5510c599bc98767d5a231f39eed73c857e7009563960fa82229dc78a38de5"} Feb 27 07:52:23 crc kubenswrapper[4612]: I0227 07:52:23.116741 4612 scope.go:117] "RemoveContainer" containerID="e67f8de92a83ce39fcf3b74a51fadf0ab4e216d2f00209fd9b8cc92ab0c9e46a" Feb 27 07:52:23 crc kubenswrapper[4612]: I0227 07:52:23.116868 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5969dcdf8-bbhhq" Feb 27 07:52:23 crc kubenswrapper[4612]: I0227 07:52:23.122411 4612 generic.go:334] "Generic (PLEG): container finished" podID="c07650a7-6d9b-43da-ae78-e8d8dbe342f4" containerID="8f23ab18cb7541e20fb46e1015f363ad7d720d9b768f2f2fb0d6a861887e5df5" exitCode=0 Feb 27 07:52:23 crc kubenswrapper[4612]: I0227 07:52:23.122470 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-59f7dcc9c8-9b7dk" event={"ID":"c07650a7-6d9b-43da-ae78-e8d8dbe342f4","Type":"ContainerDied","Data":"8f23ab18cb7541e20fb46e1015f363ad7d720d9b768f2f2fb0d6a861887e5df5"} Feb 27 07:52:23 crc kubenswrapper[4612]: I0227 07:52:23.122507 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-59f7dcc9c8-9b7dk" event={"ID":"c07650a7-6d9b-43da-ae78-e8d8dbe342f4","Type":"ContainerDied","Data":"c3a1755593c4fc24b8296fe431ce2152d07bb1d724dd53bf31d3881fb8986642"} Feb 27 07:52:23 crc kubenswrapper[4612]: I0227 07:52:23.123601 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-59f7dcc9c8-9b7dk" Feb 27 07:52:23 crc kubenswrapper[4612]: I0227 07:52:23.146635 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5969dcdf8-bbhhq"] Feb 27 07:52:23 crc kubenswrapper[4612]: I0227 07:52:23.153452 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5969dcdf8-bbhhq"] Feb 27 07:52:23 crc kubenswrapper[4612]: I0227 07:52:23.157913 4612 scope.go:117] "RemoveContainer" containerID="e67f8de92a83ce39fcf3b74a51fadf0ab4e216d2f00209fd9b8cc92ab0c9e46a" Feb 27 07:52:23 crc kubenswrapper[4612]: I0227 07:52:23.158034 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-59f7dcc9c8-9b7dk"] Feb 27 07:52:23 crc kubenswrapper[4612]: E0227 07:52:23.158332 4612 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e67f8de92a83ce39fcf3b74a51fadf0ab4e216d2f00209fd9b8cc92ab0c9e46a\": container with ID starting with e67f8de92a83ce39fcf3b74a51fadf0ab4e216d2f00209fd9b8cc92ab0c9e46a not found: ID does not exist" containerID="e67f8de92a83ce39fcf3b74a51fadf0ab4e216d2f00209fd9b8cc92ab0c9e46a" Feb 27 07:52:23 crc kubenswrapper[4612]: I0227 07:52:23.158372 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e67f8de92a83ce39fcf3b74a51fadf0ab4e216d2f00209fd9b8cc92ab0c9e46a"} err="failed to get container status \"e67f8de92a83ce39fcf3b74a51fadf0ab4e216d2f00209fd9b8cc92ab0c9e46a\": rpc error: code = NotFound desc = could not find container \"e67f8de92a83ce39fcf3b74a51fadf0ab4e216d2f00209fd9b8cc92ab0c9e46a\": container with ID starting with e67f8de92a83ce39fcf3b74a51fadf0ab4e216d2f00209fd9b8cc92ab0c9e46a not found: ID does not exist" Feb 27 07:52:23 crc kubenswrapper[4612]: I0227 07:52:23.158409 4612 scope.go:117] "RemoveContainer" containerID="8f23ab18cb7541e20fb46e1015f363ad7d720d9b768f2f2fb0d6a861887e5df5" Feb 27 07:52:23 crc kubenswrapper[4612]: I0227 07:52:23.162465 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-59f7dcc9c8-9b7dk"] Feb 27 07:52:23 crc kubenswrapper[4612]: I0227 07:52:23.179922 4612 scope.go:117] "RemoveContainer" containerID="8f23ab18cb7541e20fb46e1015f363ad7d720d9b768f2f2fb0d6a861887e5df5" Feb 27 07:52:23 crc kubenswrapper[4612]: E0227 07:52:23.180394 4612 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8f23ab18cb7541e20fb46e1015f363ad7d720d9b768f2f2fb0d6a861887e5df5\": container with ID starting with 8f23ab18cb7541e20fb46e1015f363ad7d720d9b768f2f2fb0d6a861887e5df5 not found: ID does not exist" containerID="8f23ab18cb7541e20fb46e1015f363ad7d720d9b768f2f2fb0d6a861887e5df5" Feb 27 07:52:23 crc kubenswrapper[4612]: I0227 07:52:23.180447 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8f23ab18cb7541e20fb46e1015f363ad7d720d9b768f2f2fb0d6a861887e5df5"} err="failed to get container status \"8f23ab18cb7541e20fb46e1015f363ad7d720d9b768f2f2fb0d6a861887e5df5\": rpc error: code = NotFound desc = could not find container \"8f23ab18cb7541e20fb46e1015f363ad7d720d9b768f2f2fb0d6a861887e5df5\": container with ID starting with 8f23ab18cb7541e20fb46e1015f363ad7d720d9b768f2f2fb0d6a861887e5df5 not found: ID does not exist" Feb 27 07:52:23 crc kubenswrapper[4612]: I0227 07:52:23.731364 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-77df9fd6fb-6kkpl"] Feb 27 07:52:23 crc kubenswrapper[4612]: E0227 07:52:23.731827 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f167bfef-3a04-4363-aa30-7f75a88a35e9" containerName="extract-utilities" Feb 27 07:52:23 crc kubenswrapper[4612]: I0227 07:52:23.731866 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="f167bfef-3a04-4363-aa30-7f75a88a35e9" containerName="extract-utilities" Feb 27 07:52:23 crc kubenswrapper[4612]: E0227 07:52:23.731893 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="396f1b18-f225-4d67-a98b-87b2605406d4" containerName="oc" Feb 27 07:52:23 crc kubenswrapper[4612]: I0227 07:52:23.731911 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="396f1b18-f225-4d67-a98b-87b2605406d4" containerName="oc" Feb 27 07:52:23 crc kubenswrapper[4612]: E0227 07:52:23.731937 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b4a05b89-0ef2-4f07-b1de-f558a36415a3" containerName="extract-utilities" Feb 27 07:52:23 crc kubenswrapper[4612]: I0227 07:52:23.731975 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="b4a05b89-0ef2-4f07-b1de-f558a36415a3" containerName="extract-utilities" Feb 27 07:52:23 crc kubenswrapper[4612]: E0227 07:52:23.731995 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7676c2d7-9df2-4dc8-a2bd-cde398845d26" containerName="registry-server" Feb 27 07:52:23 crc kubenswrapper[4612]: I0227 07:52:23.732013 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="7676c2d7-9df2-4dc8-a2bd-cde398845d26" containerName="registry-server" Feb 27 07:52:23 crc kubenswrapper[4612]: E0227 07:52:23.732037 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f167bfef-3a04-4363-aa30-7f75a88a35e9" containerName="extract-content" Feb 27 07:52:23 crc kubenswrapper[4612]: I0227 07:52:23.732054 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="f167bfef-3a04-4363-aa30-7f75a88a35e9" containerName="extract-content" Feb 27 07:52:23 crc kubenswrapper[4612]: E0227 07:52:23.732083 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f04702a5-b6df-431b-9f93-a37d9bf5dc96" containerName="route-controller-manager" Feb 27 07:52:23 crc kubenswrapper[4612]: I0227 07:52:23.732101 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="f04702a5-b6df-431b-9f93-a37d9bf5dc96" containerName="route-controller-manager" Feb 27 07:52:23 crc kubenswrapper[4612]: E0227 07:52:23.732130 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b4a05b89-0ef2-4f07-b1de-f558a36415a3" containerName="extract-content" Feb 27 07:52:23 crc kubenswrapper[4612]: I0227 07:52:23.732146 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="b4a05b89-0ef2-4f07-b1de-f558a36415a3" containerName="extract-content" Feb 27 07:52:23 crc kubenswrapper[4612]: E0227 07:52:23.732162 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b4a05b89-0ef2-4f07-b1de-f558a36415a3" containerName="registry-server" Feb 27 07:52:23 crc kubenswrapper[4612]: I0227 07:52:23.732178 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="b4a05b89-0ef2-4f07-b1de-f558a36415a3" containerName="registry-server" Feb 27 07:52:23 crc kubenswrapper[4612]: E0227 07:52:23.732205 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7676c2d7-9df2-4dc8-a2bd-cde398845d26" containerName="extract-utilities" Feb 27 07:52:23 crc kubenswrapper[4612]: I0227 07:52:23.732225 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="7676c2d7-9df2-4dc8-a2bd-cde398845d26" containerName="extract-utilities" Feb 27 07:52:23 crc kubenswrapper[4612]: E0227 07:52:23.732250 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f167bfef-3a04-4363-aa30-7f75a88a35e9" containerName="registry-server" Feb 27 07:52:23 crc kubenswrapper[4612]: I0227 07:52:23.732266 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="f167bfef-3a04-4363-aa30-7f75a88a35e9" containerName="registry-server" Feb 27 07:52:23 crc kubenswrapper[4612]: E0227 07:52:23.732291 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7676c2d7-9df2-4dc8-a2bd-cde398845d26" containerName="extract-content" Feb 27 07:52:23 crc kubenswrapper[4612]: I0227 07:52:23.732307 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="7676c2d7-9df2-4dc8-a2bd-cde398845d26" containerName="extract-content" Feb 27 07:52:23 crc kubenswrapper[4612]: E0227 07:52:23.732328 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c07650a7-6d9b-43da-ae78-e8d8dbe342f4" containerName="controller-manager" Feb 27 07:52:23 crc kubenswrapper[4612]: I0227 07:52:23.732344 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="c07650a7-6d9b-43da-ae78-e8d8dbe342f4" containerName="controller-manager" Feb 27 07:52:23 crc kubenswrapper[4612]: E0227 07:52:23.732364 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ece12fd-d439-4104-a7a9-a9d174e29b5a" containerName="oauth-openshift" Feb 27 07:52:23 crc kubenswrapper[4612]: I0227 07:52:23.732383 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ece12fd-d439-4104-a7a9-a9d174e29b5a" containerName="oauth-openshift" Feb 27 07:52:23 crc kubenswrapper[4612]: I0227 07:52:23.732615 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="f04702a5-b6df-431b-9f93-a37d9bf5dc96" containerName="route-controller-manager" Feb 27 07:52:23 crc kubenswrapper[4612]: I0227 07:52:23.732643 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="7676c2d7-9df2-4dc8-a2bd-cde398845d26" containerName="registry-server" Feb 27 07:52:23 crc kubenswrapper[4612]: I0227 07:52:23.732670 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="c07650a7-6d9b-43da-ae78-e8d8dbe342f4" containerName="controller-manager" Feb 27 07:52:23 crc kubenswrapper[4612]: I0227 07:52:23.732737 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="f167bfef-3a04-4363-aa30-7f75a88a35e9" containerName="registry-server" Feb 27 07:52:23 crc kubenswrapper[4612]: I0227 07:52:23.732762 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="396f1b18-f225-4d67-a98b-87b2605406d4" containerName="oc" Feb 27 07:52:23 crc kubenswrapper[4612]: I0227 07:52:23.732784 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="b4a05b89-0ef2-4f07-b1de-f558a36415a3" containerName="registry-server" Feb 27 07:52:23 crc kubenswrapper[4612]: I0227 07:52:23.732810 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="6ece12fd-d439-4104-a7a9-a9d174e29b5a" containerName="oauth-openshift" Feb 27 07:52:23 crc kubenswrapper[4612]: I0227 07:52:23.733580 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-77df9fd6fb-6kkpl" Feb 27 07:52:23 crc kubenswrapper[4612]: I0227 07:52:23.737992 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Feb 27 07:52:23 crc kubenswrapper[4612]: I0227 07:52:23.738148 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Feb 27 07:52:23 crc kubenswrapper[4612]: I0227 07:52:23.740027 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Feb 27 07:52:23 crc kubenswrapper[4612]: I0227 07:52:23.740158 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Feb 27 07:52:23 crc kubenswrapper[4612]: I0227 07:52:23.740462 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Feb 27 07:52:23 crc kubenswrapper[4612]: I0227 07:52:23.741650 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Feb 27 07:52:23 crc kubenswrapper[4612]: I0227 07:52:23.742536 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-f9b49f89c-mxx7s"] Feb 27 07:52:23 crc kubenswrapper[4612]: I0227 07:52:23.743551 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-f9b49f89c-mxx7s" Feb 27 07:52:23 crc kubenswrapper[4612]: I0227 07:52:23.751011 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Feb 27 07:52:23 crc kubenswrapper[4612]: I0227 07:52:23.751764 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Feb 27 07:52:23 crc kubenswrapper[4612]: I0227 07:52:23.751935 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Feb 27 07:52:23 crc kubenswrapper[4612]: I0227 07:52:23.752612 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Feb 27 07:52:23 crc kubenswrapper[4612]: I0227 07:52:23.757014 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-77df9fd6fb-6kkpl"] Feb 27 07:52:23 crc kubenswrapper[4612]: I0227 07:52:23.757556 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Feb 27 07:52:23 crc kubenswrapper[4612]: I0227 07:52:23.758876 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Feb 27 07:52:23 crc kubenswrapper[4612]: I0227 07:52:23.761664 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-f9b49f89c-mxx7s"] Feb 27 07:52:23 crc kubenswrapper[4612]: I0227 07:52:23.767302 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Feb 27 07:52:23 crc kubenswrapper[4612]: I0227 07:52:23.784792 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c243dedd-5585-4353-9303-6e42ebc82f69-config\") pod \"controller-manager-f9b49f89c-mxx7s\" (UID: \"c243dedd-5585-4353-9303-6e42ebc82f69\") " pod="openshift-controller-manager/controller-manager-f9b49f89c-mxx7s" Feb 27 07:52:23 crc kubenswrapper[4612]: I0227 07:52:23.784870 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a7b8e4c7-3bf3-4ef0-8aca-96037f1eea24-client-ca\") pod \"route-controller-manager-77df9fd6fb-6kkpl\" (UID: \"a7b8e4c7-3bf3-4ef0-8aca-96037f1eea24\") " pod="openshift-route-controller-manager/route-controller-manager-77df9fd6fb-6kkpl" Feb 27 07:52:23 crc kubenswrapper[4612]: I0227 07:52:23.784910 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c243dedd-5585-4353-9303-6e42ebc82f69-proxy-ca-bundles\") pod \"controller-manager-f9b49f89c-mxx7s\" (UID: \"c243dedd-5585-4353-9303-6e42ebc82f69\") " pod="openshift-controller-manager/controller-manager-f9b49f89c-mxx7s" Feb 27 07:52:23 crc kubenswrapper[4612]: I0227 07:52:23.785017 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c243dedd-5585-4353-9303-6e42ebc82f69-serving-cert\") pod \"controller-manager-f9b49f89c-mxx7s\" (UID: \"c243dedd-5585-4353-9303-6e42ebc82f69\") " pod="openshift-controller-manager/controller-manager-f9b49f89c-mxx7s" Feb 27 07:52:23 crc kubenswrapper[4612]: I0227 07:52:23.785111 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a7b8e4c7-3bf3-4ef0-8aca-96037f1eea24-config\") pod \"route-controller-manager-77df9fd6fb-6kkpl\" (UID: \"a7b8e4c7-3bf3-4ef0-8aca-96037f1eea24\") " pod="openshift-route-controller-manager/route-controller-manager-77df9fd6fb-6kkpl" Feb 27 07:52:23 crc kubenswrapper[4612]: I0227 07:52:23.785156 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a7b8e4c7-3bf3-4ef0-8aca-96037f1eea24-serving-cert\") pod \"route-controller-manager-77df9fd6fb-6kkpl\" (UID: \"a7b8e4c7-3bf3-4ef0-8aca-96037f1eea24\") " pod="openshift-route-controller-manager/route-controller-manager-77df9fd6fb-6kkpl" Feb 27 07:52:23 crc kubenswrapper[4612]: I0227 07:52:23.785204 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c243dedd-5585-4353-9303-6e42ebc82f69-client-ca\") pod \"controller-manager-f9b49f89c-mxx7s\" (UID: \"c243dedd-5585-4353-9303-6e42ebc82f69\") " pod="openshift-controller-manager/controller-manager-f9b49f89c-mxx7s" Feb 27 07:52:23 crc kubenswrapper[4612]: I0227 07:52:23.785252 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qpq69\" (UniqueName: \"kubernetes.io/projected/a7b8e4c7-3bf3-4ef0-8aca-96037f1eea24-kube-api-access-qpq69\") pod \"route-controller-manager-77df9fd6fb-6kkpl\" (UID: \"a7b8e4c7-3bf3-4ef0-8aca-96037f1eea24\") " pod="openshift-route-controller-manager/route-controller-manager-77df9fd6fb-6kkpl" Feb 27 07:52:23 crc kubenswrapper[4612]: I0227 07:52:23.785300 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ql99m\" (UniqueName: \"kubernetes.io/projected/c243dedd-5585-4353-9303-6e42ebc82f69-kube-api-access-ql99m\") pod \"controller-manager-f9b49f89c-mxx7s\" (UID: \"c243dedd-5585-4353-9303-6e42ebc82f69\") " pod="openshift-controller-manager/controller-manager-f9b49f89c-mxx7s" Feb 27 07:52:23 crc kubenswrapper[4612]: I0227 07:52:23.886532 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c243dedd-5585-4353-9303-6e42ebc82f69-config\") pod \"controller-manager-f9b49f89c-mxx7s\" (UID: \"c243dedd-5585-4353-9303-6e42ebc82f69\") " pod="openshift-controller-manager/controller-manager-f9b49f89c-mxx7s" Feb 27 07:52:23 crc kubenswrapper[4612]: I0227 07:52:23.886579 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a7b8e4c7-3bf3-4ef0-8aca-96037f1eea24-client-ca\") pod \"route-controller-manager-77df9fd6fb-6kkpl\" (UID: \"a7b8e4c7-3bf3-4ef0-8aca-96037f1eea24\") " pod="openshift-route-controller-manager/route-controller-manager-77df9fd6fb-6kkpl" Feb 27 07:52:23 crc kubenswrapper[4612]: I0227 07:52:23.886602 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c243dedd-5585-4353-9303-6e42ebc82f69-proxy-ca-bundles\") pod \"controller-manager-f9b49f89c-mxx7s\" (UID: \"c243dedd-5585-4353-9303-6e42ebc82f69\") " pod="openshift-controller-manager/controller-manager-f9b49f89c-mxx7s" Feb 27 07:52:23 crc kubenswrapper[4612]: I0227 07:52:23.886655 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c243dedd-5585-4353-9303-6e42ebc82f69-serving-cert\") pod \"controller-manager-f9b49f89c-mxx7s\" (UID: \"c243dedd-5585-4353-9303-6e42ebc82f69\") " pod="openshift-controller-manager/controller-manager-f9b49f89c-mxx7s" Feb 27 07:52:23 crc kubenswrapper[4612]: I0227 07:52:23.886672 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a7b8e4c7-3bf3-4ef0-8aca-96037f1eea24-serving-cert\") pod \"route-controller-manager-77df9fd6fb-6kkpl\" (UID: \"a7b8e4c7-3bf3-4ef0-8aca-96037f1eea24\") " pod="openshift-route-controller-manager/route-controller-manager-77df9fd6fb-6kkpl" Feb 27 07:52:23 crc kubenswrapper[4612]: I0227 07:52:23.886688 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a7b8e4c7-3bf3-4ef0-8aca-96037f1eea24-config\") pod \"route-controller-manager-77df9fd6fb-6kkpl\" (UID: \"a7b8e4c7-3bf3-4ef0-8aca-96037f1eea24\") " pod="openshift-route-controller-manager/route-controller-manager-77df9fd6fb-6kkpl" Feb 27 07:52:23 crc kubenswrapper[4612]: I0227 07:52:23.886726 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c243dedd-5585-4353-9303-6e42ebc82f69-client-ca\") pod \"controller-manager-f9b49f89c-mxx7s\" (UID: \"c243dedd-5585-4353-9303-6e42ebc82f69\") " pod="openshift-controller-manager/controller-manager-f9b49f89c-mxx7s" Feb 27 07:52:23 crc kubenswrapper[4612]: I0227 07:52:23.886752 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qpq69\" (UniqueName: \"kubernetes.io/projected/a7b8e4c7-3bf3-4ef0-8aca-96037f1eea24-kube-api-access-qpq69\") pod \"route-controller-manager-77df9fd6fb-6kkpl\" (UID: \"a7b8e4c7-3bf3-4ef0-8aca-96037f1eea24\") " pod="openshift-route-controller-manager/route-controller-manager-77df9fd6fb-6kkpl" Feb 27 07:52:23 crc kubenswrapper[4612]: I0227 07:52:23.886770 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ql99m\" (UniqueName: \"kubernetes.io/projected/c243dedd-5585-4353-9303-6e42ebc82f69-kube-api-access-ql99m\") pod \"controller-manager-f9b49f89c-mxx7s\" (UID: \"c243dedd-5585-4353-9303-6e42ebc82f69\") " pod="openshift-controller-manager/controller-manager-f9b49f89c-mxx7s" Feb 27 07:52:23 crc kubenswrapper[4612]: I0227 07:52:23.888238 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a7b8e4c7-3bf3-4ef0-8aca-96037f1eea24-config\") pod \"route-controller-manager-77df9fd6fb-6kkpl\" (UID: \"a7b8e4c7-3bf3-4ef0-8aca-96037f1eea24\") " pod="openshift-route-controller-manager/route-controller-manager-77df9fd6fb-6kkpl" Feb 27 07:52:23 crc kubenswrapper[4612]: I0227 07:52:23.888275 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c243dedd-5585-4353-9303-6e42ebc82f69-client-ca\") pod \"controller-manager-f9b49f89c-mxx7s\" (UID: \"c243dedd-5585-4353-9303-6e42ebc82f69\") " pod="openshift-controller-manager/controller-manager-f9b49f89c-mxx7s" Feb 27 07:52:23 crc kubenswrapper[4612]: I0227 07:52:23.889376 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a7b8e4c7-3bf3-4ef0-8aca-96037f1eea24-client-ca\") pod \"route-controller-manager-77df9fd6fb-6kkpl\" (UID: \"a7b8e4c7-3bf3-4ef0-8aca-96037f1eea24\") " pod="openshift-route-controller-manager/route-controller-manager-77df9fd6fb-6kkpl" Feb 27 07:52:23 crc kubenswrapper[4612]: I0227 07:52:23.889863 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c243dedd-5585-4353-9303-6e42ebc82f69-proxy-ca-bundles\") pod \"controller-manager-f9b49f89c-mxx7s\" (UID: \"c243dedd-5585-4353-9303-6e42ebc82f69\") " pod="openshift-controller-manager/controller-manager-f9b49f89c-mxx7s" Feb 27 07:52:23 crc kubenswrapper[4612]: I0227 07:52:23.889614 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c243dedd-5585-4353-9303-6e42ebc82f69-config\") pod \"controller-manager-f9b49f89c-mxx7s\" (UID: \"c243dedd-5585-4353-9303-6e42ebc82f69\") " pod="openshift-controller-manager/controller-manager-f9b49f89c-mxx7s" Feb 27 07:52:23 crc kubenswrapper[4612]: I0227 07:52:23.897450 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a7b8e4c7-3bf3-4ef0-8aca-96037f1eea24-serving-cert\") pod \"route-controller-manager-77df9fd6fb-6kkpl\" (UID: \"a7b8e4c7-3bf3-4ef0-8aca-96037f1eea24\") " pod="openshift-route-controller-manager/route-controller-manager-77df9fd6fb-6kkpl" Feb 27 07:52:23 crc kubenswrapper[4612]: I0227 07:52:23.903886 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c243dedd-5585-4353-9303-6e42ebc82f69-serving-cert\") pod \"controller-manager-f9b49f89c-mxx7s\" (UID: \"c243dedd-5585-4353-9303-6e42ebc82f69\") " pod="openshift-controller-manager/controller-manager-f9b49f89c-mxx7s" Feb 27 07:52:23 crc kubenswrapper[4612]: I0227 07:52:23.905802 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qpq69\" (UniqueName: \"kubernetes.io/projected/a7b8e4c7-3bf3-4ef0-8aca-96037f1eea24-kube-api-access-qpq69\") pod \"route-controller-manager-77df9fd6fb-6kkpl\" (UID: \"a7b8e4c7-3bf3-4ef0-8aca-96037f1eea24\") " pod="openshift-route-controller-manager/route-controller-manager-77df9fd6fb-6kkpl" Feb 27 07:52:23 crc kubenswrapper[4612]: I0227 07:52:23.909466 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ql99m\" (UniqueName: \"kubernetes.io/projected/c243dedd-5585-4353-9303-6e42ebc82f69-kube-api-access-ql99m\") pod \"controller-manager-f9b49f89c-mxx7s\" (UID: \"c243dedd-5585-4353-9303-6e42ebc82f69\") " pod="openshift-controller-manager/controller-manager-f9b49f89c-mxx7s" Feb 27 07:52:24 crc kubenswrapper[4612]: I0227 07:52:24.069820 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-77df9fd6fb-6kkpl" Feb 27 07:52:24 crc kubenswrapper[4612]: I0227 07:52:24.102946 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-f9b49f89c-mxx7s" Feb 27 07:52:24 crc kubenswrapper[4612]: I0227 07:52:24.307587 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-77df9fd6fb-6kkpl"] Feb 27 07:52:24 crc kubenswrapper[4612]: I0227 07:52:24.339968 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-f9b49f89c-mxx7s"] Feb 27 07:52:24 crc kubenswrapper[4612]: I0227 07:52:24.727717 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-5fbdbc85dc-8qzqj"] Feb 27 07:52:24 crc kubenswrapper[4612]: I0227 07:52:24.728946 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-5fbdbc85dc-8qzqj" Feb 27 07:52:24 crc kubenswrapper[4612]: I0227 07:52:24.735143 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Feb 27 07:52:24 crc kubenswrapper[4612]: I0227 07:52:24.735191 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Feb 27 07:52:24 crc kubenswrapper[4612]: I0227 07:52:24.735301 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Feb 27 07:52:24 crc kubenswrapper[4612]: I0227 07:52:24.735568 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Feb 27 07:52:24 crc kubenswrapper[4612]: I0227 07:52:24.735580 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Feb 27 07:52:24 crc kubenswrapper[4612]: I0227 07:52:24.735670 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Feb 27 07:52:24 crc kubenswrapper[4612]: I0227 07:52:24.735730 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Feb 27 07:52:24 crc kubenswrapper[4612]: I0227 07:52:24.736091 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Feb 27 07:52:24 crc kubenswrapper[4612]: I0227 07:52:24.736867 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Feb 27 07:52:24 crc kubenswrapper[4612]: I0227 07:52:24.736926 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Feb 27 07:52:24 crc kubenswrapper[4612]: I0227 07:52:24.740375 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Feb 27 07:52:24 crc kubenswrapper[4612]: I0227 07:52:24.740590 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Feb 27 07:52:24 crc kubenswrapper[4612]: I0227 07:52:24.744566 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Feb 27 07:52:24 crc kubenswrapper[4612]: I0227 07:52:24.745220 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Feb 27 07:52:24 crc kubenswrapper[4612]: I0227 07:52:24.748794 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Feb 27 07:52:24 crc kubenswrapper[4612]: I0227 07:52:24.785420 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-5fbdbc85dc-8qzqj"] Feb 27 07:52:24 crc kubenswrapper[4612]: I0227 07:52:24.797235 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/e34d5825-a10e-49cc-9a35-6a873a1a3b41-audit-policies\") pod \"oauth-openshift-5fbdbc85dc-8qzqj\" (UID: \"e34d5825-a10e-49cc-9a35-6a873a1a3b41\") " pod="openshift-authentication/oauth-openshift-5fbdbc85dc-8qzqj" Feb 27 07:52:24 crc kubenswrapper[4612]: I0227 07:52:24.797308 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e34d5825-a10e-49cc-9a35-6a873a1a3b41-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-5fbdbc85dc-8qzqj\" (UID: \"e34d5825-a10e-49cc-9a35-6a873a1a3b41\") " pod="openshift-authentication/oauth-openshift-5fbdbc85dc-8qzqj" Feb 27 07:52:24 crc kubenswrapper[4612]: I0227 07:52:24.797355 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-76r9t\" (UniqueName: \"kubernetes.io/projected/e34d5825-a10e-49cc-9a35-6a873a1a3b41-kube-api-access-76r9t\") pod \"oauth-openshift-5fbdbc85dc-8qzqj\" (UID: \"e34d5825-a10e-49cc-9a35-6a873a1a3b41\") " pod="openshift-authentication/oauth-openshift-5fbdbc85dc-8qzqj" Feb 27 07:52:24 crc kubenswrapper[4612]: I0227 07:52:24.797390 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/e34d5825-a10e-49cc-9a35-6a873a1a3b41-v4-0-config-system-serving-cert\") pod \"oauth-openshift-5fbdbc85dc-8qzqj\" (UID: \"e34d5825-a10e-49cc-9a35-6a873a1a3b41\") " pod="openshift-authentication/oauth-openshift-5fbdbc85dc-8qzqj" Feb 27 07:52:24 crc kubenswrapper[4612]: I0227 07:52:24.797426 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/e34d5825-a10e-49cc-9a35-6a873a1a3b41-v4-0-config-user-template-login\") pod \"oauth-openshift-5fbdbc85dc-8qzqj\" (UID: \"e34d5825-a10e-49cc-9a35-6a873a1a3b41\") " pod="openshift-authentication/oauth-openshift-5fbdbc85dc-8qzqj" Feb 27 07:52:24 crc kubenswrapper[4612]: I0227 07:52:24.797471 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/e34d5825-a10e-49cc-9a35-6a873a1a3b41-v4-0-config-system-cliconfig\") pod \"oauth-openshift-5fbdbc85dc-8qzqj\" (UID: \"e34d5825-a10e-49cc-9a35-6a873a1a3b41\") " pod="openshift-authentication/oauth-openshift-5fbdbc85dc-8qzqj" Feb 27 07:52:24 crc kubenswrapper[4612]: I0227 07:52:24.797499 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/e34d5825-a10e-49cc-9a35-6a873a1a3b41-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-5fbdbc85dc-8qzqj\" (UID: \"e34d5825-a10e-49cc-9a35-6a873a1a3b41\") " pod="openshift-authentication/oauth-openshift-5fbdbc85dc-8qzqj" Feb 27 07:52:24 crc kubenswrapper[4612]: I0227 07:52:24.797525 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/e34d5825-a10e-49cc-9a35-6a873a1a3b41-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-5fbdbc85dc-8qzqj\" (UID: \"e34d5825-a10e-49cc-9a35-6a873a1a3b41\") " pod="openshift-authentication/oauth-openshift-5fbdbc85dc-8qzqj" Feb 27 07:52:24 crc kubenswrapper[4612]: I0227 07:52:24.797546 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/e34d5825-a10e-49cc-9a35-6a873a1a3b41-v4-0-config-system-service-ca\") pod \"oauth-openshift-5fbdbc85dc-8qzqj\" (UID: \"e34d5825-a10e-49cc-9a35-6a873a1a3b41\") " pod="openshift-authentication/oauth-openshift-5fbdbc85dc-8qzqj" Feb 27 07:52:24 crc kubenswrapper[4612]: I0227 07:52:24.797573 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/e34d5825-a10e-49cc-9a35-6a873a1a3b41-v4-0-config-system-session\") pod \"oauth-openshift-5fbdbc85dc-8qzqj\" (UID: \"e34d5825-a10e-49cc-9a35-6a873a1a3b41\") " pod="openshift-authentication/oauth-openshift-5fbdbc85dc-8qzqj" Feb 27 07:52:24 crc kubenswrapper[4612]: I0227 07:52:24.797605 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/e34d5825-a10e-49cc-9a35-6a873a1a3b41-audit-dir\") pod \"oauth-openshift-5fbdbc85dc-8qzqj\" (UID: \"e34d5825-a10e-49cc-9a35-6a873a1a3b41\") " pod="openshift-authentication/oauth-openshift-5fbdbc85dc-8qzqj" Feb 27 07:52:24 crc kubenswrapper[4612]: I0227 07:52:24.797642 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/e34d5825-a10e-49cc-9a35-6a873a1a3b41-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-5fbdbc85dc-8qzqj\" (UID: \"e34d5825-a10e-49cc-9a35-6a873a1a3b41\") " pod="openshift-authentication/oauth-openshift-5fbdbc85dc-8qzqj" Feb 27 07:52:24 crc kubenswrapper[4612]: I0227 07:52:24.797679 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/e34d5825-a10e-49cc-9a35-6a873a1a3b41-v4-0-config-user-template-error\") pod \"oauth-openshift-5fbdbc85dc-8qzqj\" (UID: \"e34d5825-a10e-49cc-9a35-6a873a1a3b41\") " pod="openshift-authentication/oauth-openshift-5fbdbc85dc-8qzqj" Feb 27 07:52:24 crc kubenswrapper[4612]: I0227 07:52:24.797742 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/e34d5825-a10e-49cc-9a35-6a873a1a3b41-v4-0-config-system-router-certs\") pod \"oauth-openshift-5fbdbc85dc-8qzqj\" (UID: \"e34d5825-a10e-49cc-9a35-6a873a1a3b41\") " pod="openshift-authentication/oauth-openshift-5fbdbc85dc-8qzqj" Feb 27 07:52:24 crc kubenswrapper[4612]: I0227 07:52:24.858575 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c07650a7-6d9b-43da-ae78-e8d8dbe342f4" path="/var/lib/kubelet/pods/c07650a7-6d9b-43da-ae78-e8d8dbe342f4/volumes" Feb 27 07:52:24 crc kubenswrapper[4612]: I0227 07:52:24.859502 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f04702a5-b6df-431b-9f93-a37d9bf5dc96" path="/var/lib/kubelet/pods/f04702a5-b6df-431b-9f93-a37d9bf5dc96/volumes" Feb 27 07:52:24 crc kubenswrapper[4612]: I0227 07:52:24.898408 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/e34d5825-a10e-49cc-9a35-6a873a1a3b41-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-5fbdbc85dc-8qzqj\" (UID: \"e34d5825-a10e-49cc-9a35-6a873a1a3b41\") " pod="openshift-authentication/oauth-openshift-5fbdbc85dc-8qzqj" Feb 27 07:52:24 crc kubenswrapper[4612]: I0227 07:52:24.898673 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/e34d5825-a10e-49cc-9a35-6a873a1a3b41-v4-0-config-user-template-error\") pod \"oauth-openshift-5fbdbc85dc-8qzqj\" (UID: \"e34d5825-a10e-49cc-9a35-6a873a1a3b41\") " pod="openshift-authentication/oauth-openshift-5fbdbc85dc-8qzqj" Feb 27 07:52:24 crc kubenswrapper[4612]: I0227 07:52:24.898799 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/e34d5825-a10e-49cc-9a35-6a873a1a3b41-v4-0-config-system-router-certs\") pod \"oauth-openshift-5fbdbc85dc-8qzqj\" (UID: \"e34d5825-a10e-49cc-9a35-6a873a1a3b41\") " pod="openshift-authentication/oauth-openshift-5fbdbc85dc-8qzqj" Feb 27 07:52:24 crc kubenswrapper[4612]: I0227 07:52:24.898960 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/e34d5825-a10e-49cc-9a35-6a873a1a3b41-audit-policies\") pod \"oauth-openshift-5fbdbc85dc-8qzqj\" (UID: \"e34d5825-a10e-49cc-9a35-6a873a1a3b41\") " pod="openshift-authentication/oauth-openshift-5fbdbc85dc-8qzqj" Feb 27 07:52:24 crc kubenswrapper[4612]: I0227 07:52:24.899041 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e34d5825-a10e-49cc-9a35-6a873a1a3b41-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-5fbdbc85dc-8qzqj\" (UID: \"e34d5825-a10e-49cc-9a35-6a873a1a3b41\") " pod="openshift-authentication/oauth-openshift-5fbdbc85dc-8qzqj" Feb 27 07:52:24 crc kubenswrapper[4612]: I0227 07:52:24.899186 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-76r9t\" (UniqueName: \"kubernetes.io/projected/e34d5825-a10e-49cc-9a35-6a873a1a3b41-kube-api-access-76r9t\") pod \"oauth-openshift-5fbdbc85dc-8qzqj\" (UID: \"e34d5825-a10e-49cc-9a35-6a873a1a3b41\") " pod="openshift-authentication/oauth-openshift-5fbdbc85dc-8qzqj" Feb 27 07:52:24 crc kubenswrapper[4612]: I0227 07:52:24.899265 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/e34d5825-a10e-49cc-9a35-6a873a1a3b41-v4-0-config-system-serving-cert\") pod \"oauth-openshift-5fbdbc85dc-8qzqj\" (UID: \"e34d5825-a10e-49cc-9a35-6a873a1a3b41\") " pod="openshift-authentication/oauth-openshift-5fbdbc85dc-8qzqj" Feb 27 07:52:24 crc kubenswrapper[4612]: I0227 07:52:24.899304 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/e34d5825-a10e-49cc-9a35-6a873a1a3b41-v4-0-config-user-template-login\") pod \"oauth-openshift-5fbdbc85dc-8qzqj\" (UID: \"e34d5825-a10e-49cc-9a35-6a873a1a3b41\") " pod="openshift-authentication/oauth-openshift-5fbdbc85dc-8qzqj" Feb 27 07:52:24 crc kubenswrapper[4612]: I0227 07:52:24.899331 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/e34d5825-a10e-49cc-9a35-6a873a1a3b41-v4-0-config-system-cliconfig\") pod \"oauth-openshift-5fbdbc85dc-8qzqj\" (UID: \"e34d5825-a10e-49cc-9a35-6a873a1a3b41\") " pod="openshift-authentication/oauth-openshift-5fbdbc85dc-8qzqj" Feb 27 07:52:24 crc kubenswrapper[4612]: I0227 07:52:24.900202 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/e34d5825-a10e-49cc-9a35-6a873a1a3b41-audit-policies\") pod \"oauth-openshift-5fbdbc85dc-8qzqj\" (UID: \"e34d5825-a10e-49cc-9a35-6a873a1a3b41\") " pod="openshift-authentication/oauth-openshift-5fbdbc85dc-8qzqj" Feb 27 07:52:24 crc kubenswrapper[4612]: I0227 07:52:24.900312 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e34d5825-a10e-49cc-9a35-6a873a1a3b41-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-5fbdbc85dc-8qzqj\" (UID: \"e34d5825-a10e-49cc-9a35-6a873a1a3b41\") " pod="openshift-authentication/oauth-openshift-5fbdbc85dc-8qzqj" Feb 27 07:52:24 crc kubenswrapper[4612]: I0227 07:52:24.900455 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/e34d5825-a10e-49cc-9a35-6a873a1a3b41-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-5fbdbc85dc-8qzqj\" (UID: \"e34d5825-a10e-49cc-9a35-6a873a1a3b41\") " pod="openshift-authentication/oauth-openshift-5fbdbc85dc-8qzqj" Feb 27 07:52:24 crc kubenswrapper[4612]: I0227 07:52:24.900706 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/e34d5825-a10e-49cc-9a35-6a873a1a3b41-v4-0-config-system-cliconfig\") pod \"oauth-openshift-5fbdbc85dc-8qzqj\" (UID: \"e34d5825-a10e-49cc-9a35-6a873a1a3b41\") " pod="openshift-authentication/oauth-openshift-5fbdbc85dc-8qzqj" Feb 27 07:52:24 crc kubenswrapper[4612]: I0227 07:52:24.901129 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/e34d5825-a10e-49cc-9a35-6a873a1a3b41-v4-0-config-system-service-ca\") pod \"oauth-openshift-5fbdbc85dc-8qzqj\" (UID: \"e34d5825-a10e-49cc-9a35-6a873a1a3b41\") " pod="openshift-authentication/oauth-openshift-5fbdbc85dc-8qzqj" Feb 27 07:52:24 crc kubenswrapper[4612]: I0227 07:52:24.901171 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/e34d5825-a10e-49cc-9a35-6a873a1a3b41-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-5fbdbc85dc-8qzqj\" (UID: \"e34d5825-a10e-49cc-9a35-6a873a1a3b41\") " pod="openshift-authentication/oauth-openshift-5fbdbc85dc-8qzqj" Feb 27 07:52:24 crc kubenswrapper[4612]: I0227 07:52:24.901659 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/e34d5825-a10e-49cc-9a35-6a873a1a3b41-v4-0-config-system-service-ca\") pod \"oauth-openshift-5fbdbc85dc-8qzqj\" (UID: \"e34d5825-a10e-49cc-9a35-6a873a1a3b41\") " pod="openshift-authentication/oauth-openshift-5fbdbc85dc-8qzqj" Feb 27 07:52:24 crc kubenswrapper[4612]: I0227 07:52:24.901732 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/e34d5825-a10e-49cc-9a35-6a873a1a3b41-v4-0-config-system-session\") pod \"oauth-openshift-5fbdbc85dc-8qzqj\" (UID: \"e34d5825-a10e-49cc-9a35-6a873a1a3b41\") " pod="openshift-authentication/oauth-openshift-5fbdbc85dc-8qzqj" Feb 27 07:52:24 crc kubenswrapper[4612]: I0227 07:52:24.901801 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/e34d5825-a10e-49cc-9a35-6a873a1a3b41-audit-dir\") pod \"oauth-openshift-5fbdbc85dc-8qzqj\" (UID: \"e34d5825-a10e-49cc-9a35-6a873a1a3b41\") " pod="openshift-authentication/oauth-openshift-5fbdbc85dc-8qzqj" Feb 27 07:52:24 crc kubenswrapper[4612]: I0227 07:52:24.901890 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/e34d5825-a10e-49cc-9a35-6a873a1a3b41-audit-dir\") pod \"oauth-openshift-5fbdbc85dc-8qzqj\" (UID: \"e34d5825-a10e-49cc-9a35-6a873a1a3b41\") " pod="openshift-authentication/oauth-openshift-5fbdbc85dc-8qzqj" Feb 27 07:52:24 crc kubenswrapper[4612]: I0227 07:52:24.904242 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/e34d5825-a10e-49cc-9a35-6a873a1a3b41-v4-0-config-user-template-error\") pod \"oauth-openshift-5fbdbc85dc-8qzqj\" (UID: \"e34d5825-a10e-49cc-9a35-6a873a1a3b41\") " pod="openshift-authentication/oauth-openshift-5fbdbc85dc-8qzqj" Feb 27 07:52:24 crc kubenswrapper[4612]: I0227 07:52:24.904466 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/e34d5825-a10e-49cc-9a35-6a873a1a3b41-v4-0-config-system-router-certs\") pod \"oauth-openshift-5fbdbc85dc-8qzqj\" (UID: \"e34d5825-a10e-49cc-9a35-6a873a1a3b41\") " pod="openshift-authentication/oauth-openshift-5fbdbc85dc-8qzqj" Feb 27 07:52:24 crc kubenswrapper[4612]: I0227 07:52:24.904934 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/e34d5825-a10e-49cc-9a35-6a873a1a3b41-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-5fbdbc85dc-8qzqj\" (UID: \"e34d5825-a10e-49cc-9a35-6a873a1a3b41\") " pod="openshift-authentication/oauth-openshift-5fbdbc85dc-8qzqj" Feb 27 07:52:24 crc kubenswrapper[4612]: I0227 07:52:24.905404 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/e34d5825-a10e-49cc-9a35-6a873a1a3b41-v4-0-config-system-session\") pod \"oauth-openshift-5fbdbc85dc-8qzqj\" (UID: \"e34d5825-a10e-49cc-9a35-6a873a1a3b41\") " pod="openshift-authentication/oauth-openshift-5fbdbc85dc-8qzqj" Feb 27 07:52:24 crc kubenswrapper[4612]: I0227 07:52:24.906196 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/e34d5825-a10e-49cc-9a35-6a873a1a3b41-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-5fbdbc85dc-8qzqj\" (UID: \"e34d5825-a10e-49cc-9a35-6a873a1a3b41\") " pod="openshift-authentication/oauth-openshift-5fbdbc85dc-8qzqj" Feb 27 07:52:24 crc kubenswrapper[4612]: I0227 07:52:24.913315 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/e34d5825-a10e-49cc-9a35-6a873a1a3b41-v4-0-config-user-template-login\") pod \"oauth-openshift-5fbdbc85dc-8qzqj\" (UID: \"e34d5825-a10e-49cc-9a35-6a873a1a3b41\") " pod="openshift-authentication/oauth-openshift-5fbdbc85dc-8qzqj" Feb 27 07:52:24 crc kubenswrapper[4612]: I0227 07:52:24.913503 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/e34d5825-a10e-49cc-9a35-6a873a1a3b41-v4-0-config-system-serving-cert\") pod \"oauth-openshift-5fbdbc85dc-8qzqj\" (UID: \"e34d5825-a10e-49cc-9a35-6a873a1a3b41\") " pod="openshift-authentication/oauth-openshift-5fbdbc85dc-8qzqj" Feb 27 07:52:24 crc kubenswrapper[4612]: I0227 07:52:24.914881 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/e34d5825-a10e-49cc-9a35-6a873a1a3b41-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-5fbdbc85dc-8qzqj\" (UID: \"e34d5825-a10e-49cc-9a35-6a873a1a3b41\") " pod="openshift-authentication/oauth-openshift-5fbdbc85dc-8qzqj" Feb 27 07:52:24 crc kubenswrapper[4612]: I0227 07:52:24.919626 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-76r9t\" (UniqueName: \"kubernetes.io/projected/e34d5825-a10e-49cc-9a35-6a873a1a3b41-kube-api-access-76r9t\") pod \"oauth-openshift-5fbdbc85dc-8qzqj\" (UID: \"e34d5825-a10e-49cc-9a35-6a873a1a3b41\") " pod="openshift-authentication/oauth-openshift-5fbdbc85dc-8qzqj" Feb 27 07:52:25 crc kubenswrapper[4612]: I0227 07:52:25.044520 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-5fbdbc85dc-8qzqj" Feb 27 07:52:25 crc kubenswrapper[4612]: I0227 07:52:25.156455 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-77df9fd6fb-6kkpl" event={"ID":"a7b8e4c7-3bf3-4ef0-8aca-96037f1eea24","Type":"ContainerStarted","Data":"0c781d8c7babab9dc74f6dbb6ccc001984e6d120e91846d2e7b2f8be375c7123"} Feb 27 07:52:25 crc kubenswrapper[4612]: I0227 07:52:25.156503 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-77df9fd6fb-6kkpl" event={"ID":"a7b8e4c7-3bf3-4ef0-8aca-96037f1eea24","Type":"ContainerStarted","Data":"b86493dc28f0a7bfaa381c94a396a422619e0e6ddf79946d3ffd692e9fc59b8a"} Feb 27 07:52:25 crc kubenswrapper[4612]: I0227 07:52:25.156954 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-77df9fd6fb-6kkpl" Feb 27 07:52:25 crc kubenswrapper[4612]: I0227 07:52:25.160007 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-f9b49f89c-mxx7s" event={"ID":"c243dedd-5585-4353-9303-6e42ebc82f69","Type":"ContainerStarted","Data":"87605c6e32084b6a4ded19b4a44467818afafd3718fec41aa9a794efbd0c3a3d"} Feb 27 07:52:25 crc kubenswrapper[4612]: I0227 07:52:25.160064 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-f9b49f89c-mxx7s" event={"ID":"c243dedd-5585-4353-9303-6e42ebc82f69","Type":"ContainerStarted","Data":"13de94f2eb37fcf0f1e4f97d7dee8d76bdff281fa6013dd9ecb3003e524d9c70"} Feb 27 07:52:25 crc kubenswrapper[4612]: I0227 07:52:25.160887 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-f9b49f89c-mxx7s" Feb 27 07:52:25 crc kubenswrapper[4612]: I0227 07:52:25.176618 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-f9b49f89c-mxx7s" Feb 27 07:52:25 crc kubenswrapper[4612]: I0227 07:52:25.194399 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-77df9fd6fb-6kkpl" podStartSLOduration=3.19437621 podStartE2EDuration="3.19437621s" podCreationTimestamp="2026-02-27 07:52:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 07:52:25.193262929 +0000 UTC m=+203.047192937" watchObservedRunningTime="2026-02-27 07:52:25.19437621 +0000 UTC m=+203.048306208" Feb 27 07:52:25 crc kubenswrapper[4612]: I0227 07:52:25.223262 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-f9b49f89c-mxx7s" podStartSLOduration=3.223235739 podStartE2EDuration="3.223235739s" podCreationTimestamp="2026-02-27 07:52:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 07:52:25.209758251 +0000 UTC m=+203.063688259" watchObservedRunningTime="2026-02-27 07:52:25.223235739 +0000 UTC m=+203.077165747" Feb 27 07:52:25 crc kubenswrapper[4612]: E0227 07:52:25.248733 4612 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod458167a3_25b3_4cd1_8628_865ad916d3a6.slice/crio-47ac8b300890eb2dbe6d5ee7578f5f522e8820a79526ada246c4cb452b44b1ad\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcc6862b7_a2ab_4051_9d83_17774bfc5ac7.slice\": RecentStats: unable to find data in memory cache]" Feb 27 07:52:25 crc kubenswrapper[4612]: I0227 07:52:25.330981 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-77df9fd6fb-6kkpl" Feb 27 07:52:25 crc kubenswrapper[4612]: I0227 07:52:25.503355 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-5fbdbc85dc-8qzqj"] Feb 27 07:52:26 crc kubenswrapper[4612]: I0227 07:52:26.166292 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-5fbdbc85dc-8qzqj" event={"ID":"e34d5825-a10e-49cc-9a35-6a873a1a3b41","Type":"ContainerStarted","Data":"e64fa8e624b4356a6bb093e08b35609656c9d43b22e056b3cd5a8d5b54ac022d"} Feb 27 07:52:26 crc kubenswrapper[4612]: I0227 07:52:26.166610 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-5fbdbc85dc-8qzqj" event={"ID":"e34d5825-a10e-49cc-9a35-6a873a1a3b41","Type":"ContainerStarted","Data":"9dd290807f6aec5cf28d06ee6c70e44f25ebce143a5797fb301f0dbb5b5c23b6"} Feb 27 07:52:26 crc kubenswrapper[4612]: I0227 07:52:26.168590 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-5fbdbc85dc-8qzqj" Feb 27 07:52:26 crc kubenswrapper[4612]: I0227 07:52:26.172363 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-5fbdbc85dc-8qzqj" Feb 27 07:52:26 crc kubenswrapper[4612]: I0227 07:52:26.189360 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-5fbdbc85dc-8qzqj" podStartSLOduration=35.189345237 podStartE2EDuration="35.189345237s" podCreationTimestamp="2026-02-27 07:51:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 07:52:26.187234588 +0000 UTC m=+204.041164586" watchObservedRunningTime="2026-02-27 07:52:26.189345237 +0000 UTC m=+204.043275235" Feb 27 07:52:31 crc kubenswrapper[4612]: I0227 07:52:31.722464 4612 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Feb 27 07:52:31 crc kubenswrapper[4612]: I0227 07:52:31.724133 4612 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" containerID="cri-o://4cad3a2314cde773106537076b97d74afb9103af33e0450a5d804a1026fe06b7" gracePeriod=15 Feb 27 07:52:31 crc kubenswrapper[4612]: I0227 07:52:31.724122 4612 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" containerID="cri-o://ff9c7f0695586759696c95557a43826bf1198542aa46d532b4725d7d4fbc11f0" gracePeriod=15 Feb 27 07:52:31 crc kubenswrapper[4612]: I0227 07:52:31.724286 4612 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" containerID="cri-o://cc3a5e13e8220c8fa63de382ecc288979c3fcd96a0ee16c496af5e7a272b7355" gracePeriod=15 Feb 27 07:52:31 crc kubenswrapper[4612]: I0227 07:52:31.724323 4612 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" containerID="cri-o://4fc98ba294403eadeb9ad6861f1a1932dc5298dfa790e50d15148773ad290e65" gracePeriod=15 Feb 27 07:52:31 crc kubenswrapper[4612]: I0227 07:52:31.724356 4612 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" containerID="cri-o://a86d4fd23cb02a252fef6ed720b0f3eee818373ba88fe9c5586ced6566913221" gracePeriod=15 Feb 27 07:52:31 crc kubenswrapper[4612]: I0227 07:52:31.724439 4612 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Feb 27 07:52:31 crc kubenswrapper[4612]: E0227 07:52:31.724679 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Feb 27 07:52:31 crc kubenswrapper[4612]: I0227 07:52:31.724709 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Feb 27 07:52:31 crc kubenswrapper[4612]: E0227 07:52:31.724722 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 27 07:52:31 crc kubenswrapper[4612]: I0227 07:52:31.724731 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 27 07:52:31 crc kubenswrapper[4612]: E0227 07:52:31.724741 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Feb 27 07:52:31 crc kubenswrapper[4612]: I0227 07:52:31.724748 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Feb 27 07:52:31 crc kubenswrapper[4612]: E0227 07:52:31.724759 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Feb 27 07:52:31 crc kubenswrapper[4612]: I0227 07:52:31.724766 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Feb 27 07:52:31 crc kubenswrapper[4612]: E0227 07:52:31.724777 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 27 07:52:31 crc kubenswrapper[4612]: I0227 07:52:31.724783 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 27 07:52:31 crc kubenswrapper[4612]: E0227 07:52:31.724794 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Feb 27 07:52:31 crc kubenswrapper[4612]: I0227 07:52:31.724801 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Feb 27 07:52:31 crc kubenswrapper[4612]: E0227 07:52:31.724811 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 27 07:52:31 crc kubenswrapper[4612]: I0227 07:52:31.724819 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 27 07:52:31 crc kubenswrapper[4612]: E0227 07:52:31.724829 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 27 07:52:31 crc kubenswrapper[4612]: I0227 07:52:31.724836 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 27 07:52:31 crc kubenswrapper[4612]: E0227 07:52:31.724846 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Feb 27 07:52:31 crc kubenswrapper[4612]: I0227 07:52:31.724853 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Feb 27 07:52:31 crc kubenswrapper[4612]: I0227 07:52:31.724959 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Feb 27 07:52:31 crc kubenswrapper[4612]: I0227 07:52:31.724991 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 27 07:52:31 crc kubenswrapper[4612]: I0227 07:52:31.725000 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Feb 27 07:52:31 crc kubenswrapper[4612]: I0227 07:52:31.725010 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Feb 27 07:52:31 crc kubenswrapper[4612]: I0227 07:52:31.725021 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 27 07:52:31 crc kubenswrapper[4612]: I0227 07:52:31.725028 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 27 07:52:31 crc kubenswrapper[4612]: I0227 07:52:31.725037 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Feb 27 07:52:31 crc kubenswrapper[4612]: E0227 07:52:31.725136 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 27 07:52:31 crc kubenswrapper[4612]: I0227 07:52:31.725145 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 27 07:52:31 crc kubenswrapper[4612]: I0227 07:52:31.725241 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 27 07:52:31 crc kubenswrapper[4612]: I0227 07:52:31.725453 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 27 07:52:31 crc kubenswrapper[4612]: I0227 07:52:31.726592 4612 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Feb 27 07:52:31 crc kubenswrapper[4612]: I0227 07:52:31.727402 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 27 07:52:31 crc kubenswrapper[4612]: I0227 07:52:31.739016 4612 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="f4b27818a5e8e43d0dc095d08835c792" podUID="71bb4a3aecc4ba5b26c4b7318770ce13" Feb 27 07:52:31 crc kubenswrapper[4612]: I0227 07:52:31.770112 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Feb 27 07:52:31 crc kubenswrapper[4612]: I0227 07:52:31.784018 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 27 07:52:31 crc kubenswrapper[4612]: I0227 07:52:31.784131 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 27 07:52:31 crc kubenswrapper[4612]: I0227 07:52:31.784160 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 27 07:52:31 crc kubenswrapper[4612]: I0227 07:52:31.784213 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 27 07:52:31 crc kubenswrapper[4612]: I0227 07:52:31.784250 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 27 07:52:31 crc kubenswrapper[4612]: I0227 07:52:31.784278 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 27 07:52:31 crc kubenswrapper[4612]: I0227 07:52:31.784302 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 27 07:52:31 crc kubenswrapper[4612]: I0227 07:52:31.784317 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 27 07:52:31 crc kubenswrapper[4612]: I0227 07:52:31.885226 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 27 07:52:31 crc kubenswrapper[4612]: I0227 07:52:31.885303 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 27 07:52:31 crc kubenswrapper[4612]: I0227 07:52:31.885314 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 27 07:52:31 crc kubenswrapper[4612]: I0227 07:52:31.885367 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 27 07:52:31 crc kubenswrapper[4612]: I0227 07:52:31.885378 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 27 07:52:31 crc kubenswrapper[4612]: I0227 07:52:31.885328 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 27 07:52:31 crc kubenswrapper[4612]: I0227 07:52:31.885450 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 27 07:52:31 crc kubenswrapper[4612]: I0227 07:52:31.885493 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 27 07:52:31 crc kubenswrapper[4612]: I0227 07:52:31.885540 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 27 07:52:31 crc kubenswrapper[4612]: I0227 07:52:31.885578 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 27 07:52:31 crc kubenswrapper[4612]: I0227 07:52:31.885611 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 27 07:52:31 crc kubenswrapper[4612]: I0227 07:52:31.885646 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 27 07:52:31 crc kubenswrapper[4612]: I0227 07:52:31.885658 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 27 07:52:31 crc kubenswrapper[4612]: I0227 07:52:31.885631 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 27 07:52:31 crc kubenswrapper[4612]: I0227 07:52:31.885758 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 27 07:52:31 crc kubenswrapper[4612]: I0227 07:52:31.885845 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 27 07:52:32 crc kubenswrapper[4612]: I0227 07:52:32.065527 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 27 07:52:32 crc kubenswrapper[4612]: W0227 07:52:32.100250 4612 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf85e55b1a89d02b0cb034b1ea31ed45a.slice/crio-839b1bdb5c5fb490808e001d94fc7c5f376f73765cec415cb20a3e3e402d77c0 WatchSource:0}: Error finding container 839b1bdb5c5fb490808e001d94fc7c5f376f73765cec415cb20a3e3e402d77c0: Status 404 returned error can't find the container with id 839b1bdb5c5fb490808e001d94fc7c5f376f73765cec415cb20a3e3e402d77c0 Feb 27 07:52:32 crc kubenswrapper[4612]: E0227 07:52:32.103092 4612 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.158:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.18980b2f858a3be9 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 07:52:32.102087657 +0000 UTC m=+209.956017645,LastTimestamp:2026-02-27 07:52:32.102087657 +0000 UTC m=+209.956017645,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 07:52:32 crc kubenswrapper[4612]: I0227 07:52:32.208727 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/3.log" Feb 27 07:52:32 crc kubenswrapper[4612]: I0227 07:52:32.210461 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Feb 27 07:52:32 crc kubenswrapper[4612]: I0227 07:52:32.211059 4612 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="a86d4fd23cb02a252fef6ed720b0f3eee818373ba88fe9c5586ced6566913221" exitCode=2 Feb 27 07:52:32 crc kubenswrapper[4612]: I0227 07:52:32.215012 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"839b1bdb5c5fb490808e001d94fc7c5f376f73765cec415cb20a3e3e402d77c0"} Feb 27 07:52:32 crc kubenswrapper[4612]: I0227 07:52:32.855946 4612 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.158:6443: connect: connection refused" Feb 27 07:52:33 crc kubenswrapper[4612]: E0227 07:52:33.043872 4612 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.158:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.18980b2f858a3be9 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 07:52:32.102087657 +0000 UTC m=+209.956017645,LastTimestamp:2026-02-27 07:52:32.102087657 +0000 UTC m=+209.956017645,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 07:52:33 crc kubenswrapper[4612]: E0227 07:52:33.183415 4612 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.158:6443: connect: connection refused" Feb 27 07:52:33 crc kubenswrapper[4612]: E0227 07:52:33.183791 4612 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.158:6443: connect: connection refused" Feb 27 07:52:33 crc kubenswrapper[4612]: E0227 07:52:33.184094 4612 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.158:6443: connect: connection refused" Feb 27 07:52:33 crc kubenswrapper[4612]: E0227 07:52:33.184507 4612 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.158:6443: connect: connection refused" Feb 27 07:52:33 crc kubenswrapper[4612]: E0227 07:52:33.185369 4612 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.158:6443: connect: connection refused" Feb 27 07:52:33 crc kubenswrapper[4612]: I0227 07:52:33.185426 4612 controller.go:115] "failed to update lease using latest lease, fallback to ensure lease" err="failed 5 attempts to update lease" Feb 27 07:52:33 crc kubenswrapper[4612]: E0227 07:52:33.185761 4612 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.158:6443: connect: connection refused" interval="200ms" Feb 27 07:52:33 crc kubenswrapper[4612]: I0227 07:52:33.221615 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"cf5178b3d4f12046a14455df679931fd7394d9e3c06206130a4bf955e6d1525b"} Feb 27 07:52:33 crc kubenswrapper[4612]: I0227 07:52:33.222334 4612 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.158:6443: connect: connection refused" Feb 27 07:52:33 crc kubenswrapper[4612]: I0227 07:52:33.223849 4612 generic.go:334] "Generic (PLEG): container finished" podID="c13ed974-66aa-467d-9d11-4b375e53b8bc" containerID="5e4fd3279c97258bdd7509bed769c9e8ec75b544f388b59f8ed8b6f42e19ad71" exitCode=0 Feb 27 07:52:33 crc kubenswrapper[4612]: I0227 07:52:33.223928 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"c13ed974-66aa-467d-9d11-4b375e53b8bc","Type":"ContainerDied","Data":"5e4fd3279c97258bdd7509bed769c9e8ec75b544f388b59f8ed8b6f42e19ad71"} Feb 27 07:52:33 crc kubenswrapper[4612]: I0227 07:52:33.224464 4612 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.158:6443: connect: connection refused" Feb 27 07:52:33 crc kubenswrapper[4612]: I0227 07:52:33.224813 4612 status_manager.go:851] "Failed to get status for pod" podUID="c13ed974-66aa-467d-9d11-4b375e53b8bc" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.158:6443: connect: connection refused" Feb 27 07:52:33 crc kubenswrapper[4612]: I0227 07:52:33.226589 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/3.log" Feb 27 07:52:33 crc kubenswrapper[4612]: I0227 07:52:33.227815 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Feb 27 07:52:33 crc kubenswrapper[4612]: I0227 07:52:33.228508 4612 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="4cad3a2314cde773106537076b97d74afb9103af33e0450a5d804a1026fe06b7" exitCode=0 Feb 27 07:52:33 crc kubenswrapper[4612]: I0227 07:52:33.228543 4612 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="4fc98ba294403eadeb9ad6861f1a1932dc5298dfa790e50d15148773ad290e65" exitCode=0 Feb 27 07:52:33 crc kubenswrapper[4612]: I0227 07:52:33.228560 4612 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="cc3a5e13e8220c8fa63de382ecc288979c3fcd96a0ee16c496af5e7a272b7355" exitCode=0 Feb 27 07:52:33 crc kubenswrapper[4612]: I0227 07:52:33.228546 4612 scope.go:117] "RemoveContainer" containerID="6fdb61a61c41ceec6aa497297d1853ae6fec429a9dc998e0f49e8f41f41b7e9c" Feb 27 07:52:33 crc kubenswrapper[4612]: E0227 07:52:33.387113 4612 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.158:6443: connect: connection refused" interval="400ms" Feb 27 07:52:33 crc kubenswrapper[4612]: E0227 07:52:33.788187 4612 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.158:6443: connect: connection refused" interval="800ms" Feb 27 07:52:34 crc kubenswrapper[4612]: I0227 07:52:34.236284 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Feb 27 07:52:34 crc kubenswrapper[4612]: I0227 07:52:34.466671 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Feb 27 07:52:34 crc kubenswrapper[4612]: I0227 07:52:34.467972 4612 status_manager.go:851] "Failed to get status for pod" podUID="c13ed974-66aa-467d-9d11-4b375e53b8bc" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.158:6443: connect: connection refused" Feb 27 07:52:34 crc kubenswrapper[4612]: I0227 07:52:34.468549 4612 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.158:6443: connect: connection refused" Feb 27 07:52:34 crc kubenswrapper[4612]: I0227 07:52:34.519496 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c13ed974-66aa-467d-9d11-4b375e53b8bc-kube-api-access\") pod \"c13ed974-66aa-467d-9d11-4b375e53b8bc\" (UID: \"c13ed974-66aa-467d-9d11-4b375e53b8bc\") " Feb 27 07:52:34 crc kubenswrapper[4612]: I0227 07:52:34.519856 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/c13ed974-66aa-467d-9d11-4b375e53b8bc-var-lock\") pod \"c13ed974-66aa-467d-9d11-4b375e53b8bc\" (UID: \"c13ed974-66aa-467d-9d11-4b375e53b8bc\") " Feb 27 07:52:34 crc kubenswrapper[4612]: I0227 07:52:34.519898 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/c13ed974-66aa-467d-9d11-4b375e53b8bc-kubelet-dir\") pod \"c13ed974-66aa-467d-9d11-4b375e53b8bc\" (UID: \"c13ed974-66aa-467d-9d11-4b375e53b8bc\") " Feb 27 07:52:34 crc kubenswrapper[4612]: I0227 07:52:34.519977 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c13ed974-66aa-467d-9d11-4b375e53b8bc-var-lock" (OuterVolumeSpecName: "var-lock") pod "c13ed974-66aa-467d-9d11-4b375e53b8bc" (UID: "c13ed974-66aa-467d-9d11-4b375e53b8bc"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 27 07:52:34 crc kubenswrapper[4612]: I0227 07:52:34.520087 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c13ed974-66aa-467d-9d11-4b375e53b8bc-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "c13ed974-66aa-467d-9d11-4b375e53b8bc" (UID: "c13ed974-66aa-467d-9d11-4b375e53b8bc"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 27 07:52:34 crc kubenswrapper[4612]: I0227 07:52:34.520207 4612 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/c13ed974-66aa-467d-9d11-4b375e53b8bc-var-lock\") on node \"crc\" DevicePath \"\"" Feb 27 07:52:34 crc kubenswrapper[4612]: I0227 07:52:34.520225 4612 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/c13ed974-66aa-467d-9d11-4b375e53b8bc-kubelet-dir\") on node \"crc\" DevicePath \"\"" Feb 27 07:52:34 crc kubenswrapper[4612]: I0227 07:52:34.527076 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c13ed974-66aa-467d-9d11-4b375e53b8bc-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "c13ed974-66aa-467d-9d11-4b375e53b8bc" (UID: "c13ed974-66aa-467d-9d11-4b375e53b8bc"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 07:52:34 crc kubenswrapper[4612]: E0227 07:52:34.589326 4612 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.158:6443: connect: connection refused" interval="1.6s" Feb 27 07:52:34 crc kubenswrapper[4612]: I0227 07:52:34.621607 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c13ed974-66aa-467d-9d11-4b375e53b8bc-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 27 07:52:35 crc kubenswrapper[4612]: I0227 07:52:35.014532 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Feb 27 07:52:35 crc kubenswrapper[4612]: I0227 07:52:35.015929 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 27 07:52:35 crc kubenswrapper[4612]: I0227 07:52:35.016329 4612 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.158:6443: connect: connection refused" Feb 27 07:52:35 crc kubenswrapper[4612]: I0227 07:52:35.016575 4612 status_manager.go:851] "Failed to get status for pod" podUID="c13ed974-66aa-467d-9d11-4b375e53b8bc" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.158:6443: connect: connection refused" Feb 27 07:52:35 crc kubenswrapper[4612]: I0227 07:52:35.017113 4612 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.158:6443: connect: connection refused" Feb 27 07:52:35 crc kubenswrapper[4612]: I0227 07:52:35.128796 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Feb 27 07:52:35 crc kubenswrapper[4612]: I0227 07:52:35.129060 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 27 07:52:35 crc kubenswrapper[4612]: I0227 07:52:35.129228 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Feb 27 07:52:35 crc kubenswrapper[4612]: I0227 07:52:35.129382 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Feb 27 07:52:35 crc kubenswrapper[4612]: I0227 07:52:35.129237 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir" (OuterVolumeSpecName: "cert-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "cert-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 27 07:52:35 crc kubenswrapper[4612]: I0227 07:52:35.129550 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 27 07:52:35 crc kubenswrapper[4612]: I0227 07:52:35.130030 4612 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") on node \"crc\" DevicePath \"\"" Feb 27 07:52:35 crc kubenswrapper[4612]: I0227 07:52:35.130183 4612 reconciler_common.go:293] "Volume detached for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") on node \"crc\" DevicePath \"\"" Feb 27 07:52:35 crc kubenswrapper[4612]: I0227 07:52:35.130281 4612 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") on node \"crc\" DevicePath \"\"" Feb 27 07:52:35 crc kubenswrapper[4612]: I0227 07:52:35.253175 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Feb 27 07:52:35 crc kubenswrapper[4612]: I0227 07:52:35.254908 4612 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="ff9c7f0695586759696c95557a43826bf1198542aa46d532b4725d7d4fbc11f0" exitCode=0 Feb 27 07:52:35 crc kubenswrapper[4612]: I0227 07:52:35.254997 4612 scope.go:117] "RemoveContainer" containerID="4cad3a2314cde773106537076b97d74afb9103af33e0450a5d804a1026fe06b7" Feb 27 07:52:35 crc kubenswrapper[4612]: I0227 07:52:35.254998 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 27 07:52:35 crc kubenswrapper[4612]: I0227 07:52:35.256896 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"c13ed974-66aa-467d-9d11-4b375e53b8bc","Type":"ContainerDied","Data":"17264508b2043f15cff030f0af3aec3ce4093f5cb6025efd2e53ec84bc7ad125"} Feb 27 07:52:35 crc kubenswrapper[4612]: I0227 07:52:35.256930 4612 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="17264508b2043f15cff030f0af3aec3ce4093f5cb6025efd2e53ec84bc7ad125" Feb 27 07:52:35 crc kubenswrapper[4612]: I0227 07:52:35.256988 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Feb 27 07:52:35 crc kubenswrapper[4612]: I0227 07:52:35.260507 4612 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.158:6443: connect: connection refused" Feb 27 07:52:35 crc kubenswrapper[4612]: I0227 07:52:35.265542 4612 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.158:6443: connect: connection refused" Feb 27 07:52:35 crc kubenswrapper[4612]: I0227 07:52:35.266019 4612 status_manager.go:851] "Failed to get status for pod" podUID="c13ed974-66aa-467d-9d11-4b375e53b8bc" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.158:6443: connect: connection refused" Feb 27 07:52:35 crc kubenswrapper[4612]: I0227 07:52:35.270640 4612 scope.go:117] "RemoveContainer" containerID="4fc98ba294403eadeb9ad6861f1a1932dc5298dfa790e50d15148773ad290e65" Feb 27 07:52:35 crc kubenswrapper[4612]: I0227 07:52:35.273974 4612 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.158:6443: connect: connection refused" Feb 27 07:52:35 crc kubenswrapper[4612]: I0227 07:52:35.274228 4612 status_manager.go:851] "Failed to get status for pod" podUID="c13ed974-66aa-467d-9d11-4b375e53b8bc" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.158:6443: connect: connection refused" Feb 27 07:52:35 crc kubenswrapper[4612]: I0227 07:52:35.274803 4612 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.158:6443: connect: connection refused" Feb 27 07:52:35 crc kubenswrapper[4612]: I0227 07:52:35.290248 4612 scope.go:117] "RemoveContainer" containerID="cc3a5e13e8220c8fa63de382ecc288979c3fcd96a0ee16c496af5e7a272b7355" Feb 27 07:52:35 crc kubenswrapper[4612]: I0227 07:52:35.309634 4612 scope.go:117] "RemoveContainer" containerID="a86d4fd23cb02a252fef6ed720b0f3eee818373ba88fe9c5586ced6566913221" Feb 27 07:52:35 crc kubenswrapper[4612]: I0227 07:52:35.322895 4612 scope.go:117] "RemoveContainer" containerID="ff9c7f0695586759696c95557a43826bf1198542aa46d532b4725d7d4fbc11f0" Feb 27 07:52:35 crc kubenswrapper[4612]: I0227 07:52:35.339409 4612 scope.go:117] "RemoveContainer" containerID="61e0e770b36396d456078614494ef06c46bd43529d791281df3a621652328e21" Feb 27 07:52:35 crc kubenswrapper[4612]: I0227 07:52:35.377453 4612 scope.go:117] "RemoveContainer" containerID="4cad3a2314cde773106537076b97d74afb9103af33e0450a5d804a1026fe06b7" Feb 27 07:52:35 crc kubenswrapper[4612]: E0227 07:52:35.377973 4612 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4cad3a2314cde773106537076b97d74afb9103af33e0450a5d804a1026fe06b7\": container with ID starting with 4cad3a2314cde773106537076b97d74afb9103af33e0450a5d804a1026fe06b7 not found: ID does not exist" containerID="4cad3a2314cde773106537076b97d74afb9103af33e0450a5d804a1026fe06b7" Feb 27 07:52:35 crc kubenswrapper[4612]: I0227 07:52:35.378204 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4cad3a2314cde773106537076b97d74afb9103af33e0450a5d804a1026fe06b7"} err="failed to get container status \"4cad3a2314cde773106537076b97d74afb9103af33e0450a5d804a1026fe06b7\": rpc error: code = NotFound desc = could not find container \"4cad3a2314cde773106537076b97d74afb9103af33e0450a5d804a1026fe06b7\": container with ID starting with 4cad3a2314cde773106537076b97d74afb9103af33e0450a5d804a1026fe06b7 not found: ID does not exist" Feb 27 07:52:35 crc kubenswrapper[4612]: I0227 07:52:35.378240 4612 scope.go:117] "RemoveContainer" containerID="4fc98ba294403eadeb9ad6861f1a1932dc5298dfa790e50d15148773ad290e65" Feb 27 07:52:35 crc kubenswrapper[4612]: E0227 07:52:35.378931 4612 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4fc98ba294403eadeb9ad6861f1a1932dc5298dfa790e50d15148773ad290e65\": container with ID starting with 4fc98ba294403eadeb9ad6861f1a1932dc5298dfa790e50d15148773ad290e65 not found: ID does not exist" containerID="4fc98ba294403eadeb9ad6861f1a1932dc5298dfa790e50d15148773ad290e65" Feb 27 07:52:35 crc kubenswrapper[4612]: I0227 07:52:35.379014 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4fc98ba294403eadeb9ad6861f1a1932dc5298dfa790e50d15148773ad290e65"} err="failed to get container status \"4fc98ba294403eadeb9ad6861f1a1932dc5298dfa790e50d15148773ad290e65\": rpc error: code = NotFound desc = could not find container \"4fc98ba294403eadeb9ad6861f1a1932dc5298dfa790e50d15148773ad290e65\": container with ID starting with 4fc98ba294403eadeb9ad6861f1a1932dc5298dfa790e50d15148773ad290e65 not found: ID does not exist" Feb 27 07:52:35 crc kubenswrapper[4612]: I0227 07:52:35.379190 4612 scope.go:117] "RemoveContainer" containerID="cc3a5e13e8220c8fa63de382ecc288979c3fcd96a0ee16c496af5e7a272b7355" Feb 27 07:52:35 crc kubenswrapper[4612]: E0227 07:52:35.379561 4612 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cc3a5e13e8220c8fa63de382ecc288979c3fcd96a0ee16c496af5e7a272b7355\": container with ID starting with cc3a5e13e8220c8fa63de382ecc288979c3fcd96a0ee16c496af5e7a272b7355 not found: ID does not exist" containerID="cc3a5e13e8220c8fa63de382ecc288979c3fcd96a0ee16c496af5e7a272b7355" Feb 27 07:52:35 crc kubenswrapper[4612]: I0227 07:52:35.379634 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cc3a5e13e8220c8fa63de382ecc288979c3fcd96a0ee16c496af5e7a272b7355"} err="failed to get container status \"cc3a5e13e8220c8fa63de382ecc288979c3fcd96a0ee16c496af5e7a272b7355\": rpc error: code = NotFound desc = could not find container \"cc3a5e13e8220c8fa63de382ecc288979c3fcd96a0ee16c496af5e7a272b7355\": container with ID starting with cc3a5e13e8220c8fa63de382ecc288979c3fcd96a0ee16c496af5e7a272b7355 not found: ID does not exist" Feb 27 07:52:35 crc kubenswrapper[4612]: I0227 07:52:35.379663 4612 scope.go:117] "RemoveContainer" containerID="a86d4fd23cb02a252fef6ed720b0f3eee818373ba88fe9c5586ced6566913221" Feb 27 07:52:35 crc kubenswrapper[4612]: E0227 07:52:35.380408 4612 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a86d4fd23cb02a252fef6ed720b0f3eee818373ba88fe9c5586ced6566913221\": container with ID starting with a86d4fd23cb02a252fef6ed720b0f3eee818373ba88fe9c5586ced6566913221 not found: ID does not exist" containerID="a86d4fd23cb02a252fef6ed720b0f3eee818373ba88fe9c5586ced6566913221" Feb 27 07:52:35 crc kubenswrapper[4612]: I0227 07:52:35.380600 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a86d4fd23cb02a252fef6ed720b0f3eee818373ba88fe9c5586ced6566913221"} err="failed to get container status \"a86d4fd23cb02a252fef6ed720b0f3eee818373ba88fe9c5586ced6566913221\": rpc error: code = NotFound desc = could not find container \"a86d4fd23cb02a252fef6ed720b0f3eee818373ba88fe9c5586ced6566913221\": container with ID starting with a86d4fd23cb02a252fef6ed720b0f3eee818373ba88fe9c5586ced6566913221 not found: ID does not exist" Feb 27 07:52:35 crc kubenswrapper[4612]: I0227 07:52:35.380755 4612 scope.go:117] "RemoveContainer" containerID="ff9c7f0695586759696c95557a43826bf1198542aa46d532b4725d7d4fbc11f0" Feb 27 07:52:35 crc kubenswrapper[4612]: E0227 07:52:35.381222 4612 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ff9c7f0695586759696c95557a43826bf1198542aa46d532b4725d7d4fbc11f0\": container with ID starting with ff9c7f0695586759696c95557a43826bf1198542aa46d532b4725d7d4fbc11f0 not found: ID does not exist" containerID="ff9c7f0695586759696c95557a43826bf1198542aa46d532b4725d7d4fbc11f0" Feb 27 07:52:35 crc kubenswrapper[4612]: I0227 07:52:35.381304 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ff9c7f0695586759696c95557a43826bf1198542aa46d532b4725d7d4fbc11f0"} err="failed to get container status \"ff9c7f0695586759696c95557a43826bf1198542aa46d532b4725d7d4fbc11f0\": rpc error: code = NotFound desc = could not find container \"ff9c7f0695586759696c95557a43826bf1198542aa46d532b4725d7d4fbc11f0\": container with ID starting with ff9c7f0695586759696c95557a43826bf1198542aa46d532b4725d7d4fbc11f0 not found: ID does not exist" Feb 27 07:52:35 crc kubenswrapper[4612]: I0227 07:52:35.381369 4612 scope.go:117] "RemoveContainer" containerID="61e0e770b36396d456078614494ef06c46bd43529d791281df3a621652328e21" Feb 27 07:52:35 crc kubenswrapper[4612]: E0227 07:52:35.381891 4612 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"61e0e770b36396d456078614494ef06c46bd43529d791281df3a621652328e21\": container with ID starting with 61e0e770b36396d456078614494ef06c46bd43529d791281df3a621652328e21 not found: ID does not exist" containerID="61e0e770b36396d456078614494ef06c46bd43529d791281df3a621652328e21" Feb 27 07:52:35 crc kubenswrapper[4612]: I0227 07:52:35.381948 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"61e0e770b36396d456078614494ef06c46bd43529d791281df3a621652328e21"} err="failed to get container status \"61e0e770b36396d456078614494ef06c46bd43529d791281df3a621652328e21\": rpc error: code = NotFound desc = could not find container \"61e0e770b36396d456078614494ef06c46bd43529d791281df3a621652328e21\": container with ID starting with 61e0e770b36396d456078614494ef06c46bd43529d791281df3a621652328e21 not found: ID does not exist" Feb 27 07:52:35 crc kubenswrapper[4612]: E0227 07:52:35.401274 4612 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod458167a3_25b3_4cd1_8628_865ad916d3a6.slice/crio-47ac8b300890eb2dbe6d5ee7578f5f522e8820a79526ada246c4cb452b44b1ad\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcc6862b7_a2ab_4051_9d83_17774bfc5ac7.slice\": RecentStats: unable to find data in memory cache]" Feb 27 07:52:36 crc kubenswrapper[4612]: E0227 07:52:36.190867 4612 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.158:6443: connect: connection refused" interval="3.2s" Feb 27 07:52:36 crc kubenswrapper[4612]: I0227 07:52:36.860853 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4b27818a5e8e43d0dc095d08835c792" path="/var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/volumes" Feb 27 07:52:39 crc kubenswrapper[4612]: E0227 07:52:39.392757 4612 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.158:6443: connect: connection refused" interval="6.4s" Feb 27 07:52:42 crc kubenswrapper[4612]: I0227 07:52:42.859312 4612 status_manager.go:851] "Failed to get status for pod" podUID="c13ed974-66aa-467d-9d11-4b375e53b8bc" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.158:6443: connect: connection refused" Feb 27 07:52:42 crc kubenswrapper[4612]: I0227 07:52:42.860128 4612 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.158:6443: connect: connection refused" Feb 27 07:52:43 crc kubenswrapper[4612]: E0227 07:52:43.045425 4612 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.158:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.18980b2f858a3be9 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-27 07:52:32.102087657 +0000 UTC m=+209.956017645,LastTimestamp:2026-02-27 07:52:32.102087657 +0000 UTC m=+209.956017645,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 27 07:52:44 crc kubenswrapper[4612]: I0227 07:52:44.852472 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 27 07:52:44 crc kubenswrapper[4612]: I0227 07:52:44.853956 4612 status_manager.go:851] "Failed to get status for pod" podUID="c13ed974-66aa-467d-9d11-4b375e53b8bc" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.158:6443: connect: connection refused" Feb 27 07:52:44 crc kubenswrapper[4612]: I0227 07:52:44.854669 4612 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.158:6443: connect: connection refused" Feb 27 07:52:44 crc kubenswrapper[4612]: I0227 07:52:44.875551 4612 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="3190f547-944a-4358-b5f7-eb2194721535" Feb 27 07:52:44 crc kubenswrapper[4612]: I0227 07:52:44.875594 4612 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="3190f547-944a-4358-b5f7-eb2194721535" Feb 27 07:52:44 crc kubenswrapper[4612]: E0227 07:52:44.876155 4612 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.158:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 27 07:52:44 crc kubenswrapper[4612]: I0227 07:52:44.876987 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 27 07:52:44 crc kubenswrapper[4612]: W0227 07:52:44.912355 4612 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod71bb4a3aecc4ba5b26c4b7318770ce13.slice/crio-899d671f054e035083b283322baad791c966214fb3eed423096b59eba123eae5 WatchSource:0}: Error finding container 899d671f054e035083b283322baad791c966214fb3eed423096b59eba123eae5: Status 404 returned error can't find the container with id 899d671f054e035083b283322baad791c966214fb3eed423096b59eba123eae5 Feb 27 07:52:45 crc kubenswrapper[4612]: I0227 07:52:45.342314 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"b1cc589c7adf32670c073ab150f99a492403b6fad4d35e35c336e3ba0cc3e115"} Feb 27 07:52:45 crc kubenswrapper[4612]: I0227 07:52:45.342754 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"899d671f054e035083b283322baad791c966214fb3eed423096b59eba123eae5"} Feb 27 07:52:45 crc kubenswrapper[4612]: I0227 07:52:45.348177 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/cluster-policy-controller/0.log" Feb 27 07:52:45 crc kubenswrapper[4612]: I0227 07:52:45.349292 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Feb 27 07:52:45 crc kubenswrapper[4612]: I0227 07:52:45.349335 4612 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="81b42fcb5b5a6dd8e9bbf8836e9a20cf3f603ac4913e90d4907df9338209f1be" exitCode=1 Feb 27 07:52:45 crc kubenswrapper[4612]: I0227 07:52:45.349364 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"81b42fcb5b5a6dd8e9bbf8836e9a20cf3f603ac4913e90d4907df9338209f1be"} Feb 27 07:52:45 crc kubenswrapper[4612]: I0227 07:52:45.349890 4612 scope.go:117] "RemoveContainer" containerID="81b42fcb5b5a6dd8e9bbf8836e9a20cf3f603ac4913e90d4907df9338209f1be" Feb 27 07:52:45 crc kubenswrapper[4612]: I0227 07:52:45.350440 4612 status_manager.go:851] "Failed to get status for pod" podUID="c13ed974-66aa-467d-9d11-4b375e53b8bc" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.158:6443: connect: connection refused" Feb 27 07:52:45 crc kubenswrapper[4612]: I0227 07:52:45.351462 4612 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.158:6443: connect: connection refused" Feb 27 07:52:45 crc kubenswrapper[4612]: I0227 07:52:45.351926 4612 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.158:6443: connect: connection refused" Feb 27 07:52:45 crc kubenswrapper[4612]: E0227 07:52:45.568048 4612 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod458167a3_25b3_4cd1_8628_865ad916d3a6.slice/crio-47ac8b300890eb2dbe6d5ee7578f5f522e8820a79526ada246c4cb452b44b1ad\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcc6862b7_a2ab_4051_9d83_17774bfc5ac7.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod71bb4a3aecc4ba5b26c4b7318770ce13.slice/crio-conmon-b1cc589c7adf32670c073ab150f99a492403b6fad4d35e35c336e3ba0cc3e115.scope\": RecentStats: unable to find data in memory cache]" Feb 27 07:52:45 crc kubenswrapper[4612]: E0227 07:52:45.793795 4612 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.158:6443: connect: connection refused" interval="7s" Feb 27 07:52:46 crc kubenswrapper[4612]: I0227 07:52:46.027149 4612 patch_prober.go:28] interesting pod/machine-config-daemon-924vb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 27 07:52:46 crc kubenswrapper[4612]: I0227 07:52:46.027256 4612 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 27 07:52:46 crc kubenswrapper[4612]: I0227 07:52:46.072487 4612 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 27 07:52:46 crc kubenswrapper[4612]: I0227 07:52:46.361893 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/cluster-policy-controller/0.log" Feb 27 07:52:46 crc kubenswrapper[4612]: I0227 07:52:46.362665 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Feb 27 07:52:46 crc kubenswrapper[4612]: I0227 07:52:46.362853 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"21bfc48e85464106700ed7f1c25238b60b3aa710e85ed865d86afe2318d73440"} Feb 27 07:52:46 crc kubenswrapper[4612]: I0227 07:52:46.364668 4612 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.158:6443: connect: connection refused" Feb 27 07:52:46 crc kubenswrapper[4612]: I0227 07:52:46.365145 4612 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.158:6443: connect: connection refused" Feb 27 07:52:46 crc kubenswrapper[4612]: I0227 07:52:46.365555 4612 status_manager.go:851] "Failed to get status for pod" podUID="c13ed974-66aa-467d-9d11-4b375e53b8bc" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.158:6443: connect: connection refused" Feb 27 07:52:46 crc kubenswrapper[4612]: I0227 07:52:46.367987 4612 generic.go:334] "Generic (PLEG): container finished" podID="71bb4a3aecc4ba5b26c4b7318770ce13" containerID="b1cc589c7adf32670c073ab150f99a492403b6fad4d35e35c336e3ba0cc3e115" exitCode=0 Feb 27 07:52:46 crc kubenswrapper[4612]: I0227 07:52:46.368047 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerDied","Data":"b1cc589c7adf32670c073ab150f99a492403b6fad4d35e35c336e3ba0cc3e115"} Feb 27 07:52:46 crc kubenswrapper[4612]: I0227 07:52:46.368475 4612 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="3190f547-944a-4358-b5f7-eb2194721535" Feb 27 07:52:46 crc kubenswrapper[4612]: I0227 07:52:46.368516 4612 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="3190f547-944a-4358-b5f7-eb2194721535" Feb 27 07:52:46 crc kubenswrapper[4612]: E0227 07:52:46.369009 4612 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.158:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 27 07:52:46 crc kubenswrapper[4612]: I0227 07:52:46.369278 4612 status_manager.go:851] "Failed to get status for pod" podUID="c13ed974-66aa-467d-9d11-4b375e53b8bc" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.158:6443: connect: connection refused" Feb 27 07:52:46 crc kubenswrapper[4612]: I0227 07:52:46.369831 4612 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.158:6443: connect: connection refused" Feb 27 07:52:46 crc kubenswrapper[4612]: I0227 07:52:46.370328 4612 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.158:6443: connect: connection refused" Feb 27 07:52:47 crc kubenswrapper[4612]: I0227 07:52:47.380667 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"c0b3cfe56b27b892908fc38ec0f583966148914af3b31d30775af2b86e99c717"} Feb 27 07:52:47 crc kubenswrapper[4612]: I0227 07:52:47.380931 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"c56cf1f71d8765e42259ee110e1ae060b5c20c0157aca9ad803bec2adbc9c5ee"} Feb 27 07:52:47 crc kubenswrapper[4612]: I0227 07:52:47.380941 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"154c7e914fdc5267a75f33910902537892071da65501f24373a31aa0aae033d7"} Feb 27 07:52:47 crc kubenswrapper[4612]: I0227 07:52:47.380949 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"4d331c7a809ec95d99f7ccc170aa5c9ef55a1086da85226fd97de5ba0683e773"} Feb 27 07:52:48 crc kubenswrapper[4612]: I0227 07:52:48.387391 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"df2e3ba4965cd7574784d838f8c308f898683656603634630f5f0c4ba34c896a"} Feb 27 07:52:48 crc kubenswrapper[4612]: I0227 07:52:48.387556 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 27 07:52:48 crc kubenswrapper[4612]: I0227 07:52:48.387634 4612 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="3190f547-944a-4358-b5f7-eb2194721535" Feb 27 07:52:48 crc kubenswrapper[4612]: I0227 07:52:48.387657 4612 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="3190f547-944a-4358-b5f7-eb2194721535" Feb 27 07:52:49 crc kubenswrapper[4612]: I0227 07:52:49.877655 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 27 07:52:49 crc kubenswrapper[4612]: I0227 07:52:49.878450 4612 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 27 07:52:49 crc kubenswrapper[4612]: I0227 07:52:49.884108 4612 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 27 07:52:51 crc kubenswrapper[4612]: I0227 07:52:51.773861 4612 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 27 07:52:51 crc kubenswrapper[4612]: I0227 07:52:51.778382 4612 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 27 07:52:52 crc kubenswrapper[4612]: I0227 07:52:52.420431 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 27 07:52:53 crc kubenswrapper[4612]: I0227 07:52:53.396039 4612 kubelet.go:1914] "Deleted mirror pod because it is outdated" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 27 07:52:53 crc kubenswrapper[4612]: I0227 07:52:53.424684 4612 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="3190f547-944a-4358-b5f7-eb2194721535" Feb 27 07:52:53 crc kubenswrapper[4612]: I0227 07:52:53.424740 4612 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="3190f547-944a-4358-b5f7-eb2194721535" Feb 27 07:52:53 crc kubenswrapper[4612]: I0227 07:52:53.428625 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 27 07:52:53 crc kubenswrapper[4612]: I0227 07:52:53.471686 4612 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="cce00ca4-d431-417a-8da0-b8fd9fa90536" Feb 27 07:52:54 crc kubenswrapper[4612]: I0227 07:52:54.428654 4612 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="3190f547-944a-4358-b5f7-eb2194721535" Feb 27 07:52:54 crc kubenswrapper[4612]: I0227 07:52:54.428728 4612 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="3190f547-944a-4358-b5f7-eb2194721535" Feb 27 07:52:54 crc kubenswrapper[4612]: I0227 07:52:54.433343 4612 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="cce00ca4-d431-417a-8da0-b8fd9fa90536" Feb 27 07:52:55 crc kubenswrapper[4612]: E0227 07:52:55.706310 4612 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod458167a3_25b3_4cd1_8628_865ad916d3a6.slice/crio-47ac8b300890eb2dbe6d5ee7578f5f522e8820a79526ada246c4cb452b44b1ad\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcc6862b7_a2ab_4051_9d83_17774bfc5ac7.slice\": RecentStats: unable to find data in memory cache]" Feb 27 07:53:03 crc kubenswrapper[4612]: I0227 07:53:03.135276 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Feb 27 07:53:03 crc kubenswrapper[4612]: I0227 07:53:03.485757 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Feb 27 07:53:03 crc kubenswrapper[4612]: I0227 07:53:03.704720 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 27 07:53:04 crc kubenswrapper[4612]: I0227 07:53:04.182358 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Feb 27 07:53:04 crc kubenswrapper[4612]: I0227 07:53:04.213445 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Feb 27 07:53:04 crc kubenswrapper[4612]: I0227 07:53:04.225360 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Feb 27 07:53:04 crc kubenswrapper[4612]: I0227 07:53:04.416876 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Feb 27 07:53:04 crc kubenswrapper[4612]: I0227 07:53:04.442392 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Feb 27 07:53:04 crc kubenswrapper[4612]: I0227 07:53:04.672179 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Feb 27 07:53:04 crc kubenswrapper[4612]: I0227 07:53:04.709291 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Feb 27 07:53:04 crc kubenswrapper[4612]: I0227 07:53:04.870387 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Feb 27 07:53:04 crc kubenswrapper[4612]: I0227 07:53:04.900892 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Feb 27 07:53:04 crc kubenswrapper[4612]: I0227 07:53:04.943487 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Feb 27 07:53:04 crc kubenswrapper[4612]: I0227 07:53:04.953684 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Feb 27 07:53:04 crc kubenswrapper[4612]: I0227 07:53:04.959397 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Feb 27 07:53:05 crc kubenswrapper[4612]: I0227 07:53:05.085369 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Feb 27 07:53:05 crc kubenswrapper[4612]: I0227 07:53:05.114912 4612 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Feb 27 07:53:05 crc kubenswrapper[4612]: I0227 07:53:05.217823 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Feb 27 07:53:05 crc kubenswrapper[4612]: I0227 07:53:05.320310 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Feb 27 07:53:05 crc kubenswrapper[4612]: I0227 07:53:05.390539 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Feb 27 07:53:05 crc kubenswrapper[4612]: I0227 07:53:05.412622 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Feb 27 07:53:05 crc kubenswrapper[4612]: I0227 07:53:05.495633 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Feb 27 07:53:05 crc kubenswrapper[4612]: I0227 07:53:05.509272 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Feb 27 07:53:05 crc kubenswrapper[4612]: I0227 07:53:05.604610 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Feb 27 07:53:05 crc kubenswrapper[4612]: I0227 07:53:05.635577 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Feb 27 07:53:05 crc kubenswrapper[4612]: I0227 07:53:05.670372 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Feb 27 07:53:05 crc kubenswrapper[4612]: I0227 07:53:05.712535 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Feb 27 07:53:05 crc kubenswrapper[4612]: I0227 07:53:05.791110 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Feb 27 07:53:05 crc kubenswrapper[4612]: I0227 07:53:05.909447 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Feb 27 07:53:05 crc kubenswrapper[4612]: I0227 07:53:05.946886 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Feb 27 07:53:05 crc kubenswrapper[4612]: I0227 07:53:05.959814 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Feb 27 07:53:05 crc kubenswrapper[4612]: I0227 07:53:05.992541 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Feb 27 07:53:06 crc kubenswrapper[4612]: I0227 07:53:06.002669 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Feb 27 07:53:06 crc kubenswrapper[4612]: I0227 07:53:06.023383 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Feb 27 07:53:06 crc kubenswrapper[4612]: I0227 07:53:06.035437 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Feb 27 07:53:06 crc kubenswrapper[4612]: I0227 07:53:06.097491 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Feb 27 07:53:06 crc kubenswrapper[4612]: I0227 07:53:06.118429 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Feb 27 07:53:06 crc kubenswrapper[4612]: I0227 07:53:06.162511 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Feb 27 07:53:06 crc kubenswrapper[4612]: I0227 07:53:06.715505 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Feb 27 07:53:06 crc kubenswrapper[4612]: I0227 07:53:06.720918 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Feb 27 07:53:06 crc kubenswrapper[4612]: I0227 07:53:06.734836 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Feb 27 07:53:06 crc kubenswrapper[4612]: I0227 07:53:06.735210 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Feb 27 07:53:06 crc kubenswrapper[4612]: I0227 07:53:06.889109 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Feb 27 07:53:07 crc kubenswrapper[4612]: I0227 07:53:07.068569 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Feb 27 07:53:07 crc kubenswrapper[4612]: I0227 07:53:07.183019 4612 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Feb 27 07:53:07 crc kubenswrapper[4612]: I0227 07:53:07.192470 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Feb 27 07:53:07 crc kubenswrapper[4612]: I0227 07:53:07.340771 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Feb 27 07:53:07 crc kubenswrapper[4612]: I0227 07:53:07.481373 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Feb 27 07:53:07 crc kubenswrapper[4612]: I0227 07:53:07.503360 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Feb 27 07:53:07 crc kubenswrapper[4612]: I0227 07:53:07.507341 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Feb 27 07:53:07 crc kubenswrapper[4612]: I0227 07:53:07.525931 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Feb 27 07:53:07 crc kubenswrapper[4612]: I0227 07:53:07.607133 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Feb 27 07:53:07 crc kubenswrapper[4612]: I0227 07:53:07.610342 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Feb 27 07:53:07 crc kubenswrapper[4612]: I0227 07:53:07.617870 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Feb 27 07:53:07 crc kubenswrapper[4612]: I0227 07:53:07.619610 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Feb 27 07:53:07 crc kubenswrapper[4612]: I0227 07:53:07.664458 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Feb 27 07:53:07 crc kubenswrapper[4612]: I0227 07:53:07.759001 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Feb 27 07:53:07 crc kubenswrapper[4612]: I0227 07:53:07.772503 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Feb 27 07:53:07 crc kubenswrapper[4612]: I0227 07:53:07.878984 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Feb 27 07:53:07 crc kubenswrapper[4612]: I0227 07:53:07.885920 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Feb 27 07:53:07 crc kubenswrapper[4612]: I0227 07:53:07.898958 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Feb 27 07:53:07 crc kubenswrapper[4612]: I0227 07:53:07.919046 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Feb 27 07:53:07 crc kubenswrapper[4612]: I0227 07:53:07.960515 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Feb 27 07:53:08 crc kubenswrapper[4612]: I0227 07:53:08.009820 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Feb 27 07:53:08 crc kubenswrapper[4612]: I0227 07:53:08.077544 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Feb 27 07:53:08 crc kubenswrapper[4612]: I0227 07:53:08.163371 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Feb 27 07:53:08 crc kubenswrapper[4612]: I0227 07:53:08.176434 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Feb 27 07:53:08 crc kubenswrapper[4612]: I0227 07:53:08.178970 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Feb 27 07:53:08 crc kubenswrapper[4612]: I0227 07:53:08.190910 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Feb 27 07:53:08 crc kubenswrapper[4612]: I0227 07:53:08.223536 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Feb 27 07:53:08 crc kubenswrapper[4612]: I0227 07:53:08.278829 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Feb 27 07:53:08 crc kubenswrapper[4612]: I0227 07:53:08.315327 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Feb 27 07:53:08 crc kubenswrapper[4612]: I0227 07:53:08.374953 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Feb 27 07:53:08 crc kubenswrapper[4612]: I0227 07:53:08.557100 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Feb 27 07:53:08 crc kubenswrapper[4612]: I0227 07:53:08.606800 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Feb 27 07:53:08 crc kubenswrapper[4612]: I0227 07:53:08.680292 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Feb 27 07:53:08 crc kubenswrapper[4612]: I0227 07:53:08.862996 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Feb 27 07:53:08 crc kubenswrapper[4612]: I0227 07:53:08.966831 4612 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Feb 27 07:53:08 crc kubenswrapper[4612]: I0227 07:53:08.967992 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podStartSLOduration=37.967976949 podStartE2EDuration="37.967976949s" podCreationTimestamp="2026-02-27 07:52:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 07:52:53.408118301 +0000 UTC m=+231.262048299" watchObservedRunningTime="2026-02-27 07:53:08.967976949 +0000 UTC m=+246.821906937" Feb 27 07:53:08 crc kubenswrapper[4612]: I0227 07:53:08.971161 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Feb 27 07:53:08 crc kubenswrapper[4612]: I0227 07:53:08.971214 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Feb 27 07:53:08 crc kubenswrapper[4612]: I0227 07:53:08.977403 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 27 07:53:09 crc kubenswrapper[4612]: I0227 07:53:09.008655 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=16.008636398 podStartE2EDuration="16.008636398s" podCreationTimestamp="2026-02-27 07:52:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 07:53:08.991167178 +0000 UTC m=+246.845097186" watchObservedRunningTime="2026-02-27 07:53:09.008636398 +0000 UTC m=+246.862566396" Feb 27 07:53:09 crc kubenswrapper[4612]: I0227 07:53:09.013359 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Feb 27 07:53:09 crc kubenswrapper[4612]: I0227 07:53:09.081071 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Feb 27 07:53:09 crc kubenswrapper[4612]: I0227 07:53:09.102603 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Feb 27 07:53:09 crc kubenswrapper[4612]: I0227 07:53:09.102629 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Feb 27 07:53:09 crc kubenswrapper[4612]: I0227 07:53:09.111983 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Feb 27 07:53:09 crc kubenswrapper[4612]: I0227 07:53:09.261533 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Feb 27 07:53:09 crc kubenswrapper[4612]: I0227 07:53:09.278930 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Feb 27 07:53:09 crc kubenswrapper[4612]: I0227 07:53:09.301228 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Feb 27 07:53:09 crc kubenswrapper[4612]: I0227 07:53:09.320011 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Feb 27 07:53:09 crc kubenswrapper[4612]: I0227 07:53:09.351582 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Feb 27 07:53:09 crc kubenswrapper[4612]: I0227 07:53:09.442060 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Feb 27 07:53:09 crc kubenswrapper[4612]: I0227 07:53:09.496201 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Feb 27 07:53:09 crc kubenswrapper[4612]: I0227 07:53:09.518341 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Feb 27 07:53:09 crc kubenswrapper[4612]: I0227 07:53:09.539472 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Feb 27 07:53:09 crc kubenswrapper[4612]: I0227 07:53:09.559209 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Feb 27 07:53:09 crc kubenswrapper[4612]: I0227 07:53:09.620606 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Feb 27 07:53:09 crc kubenswrapper[4612]: I0227 07:53:09.716126 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Feb 27 07:53:09 crc kubenswrapper[4612]: I0227 07:53:09.781525 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Feb 27 07:53:09 crc kubenswrapper[4612]: I0227 07:53:09.808358 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Feb 27 07:53:09 crc kubenswrapper[4612]: I0227 07:53:09.825800 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Feb 27 07:53:09 crc kubenswrapper[4612]: I0227 07:53:09.834793 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Feb 27 07:53:09 crc kubenswrapper[4612]: I0227 07:53:09.845533 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Feb 27 07:53:09 crc kubenswrapper[4612]: I0227 07:53:09.983428 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Feb 27 07:53:10 crc kubenswrapper[4612]: I0227 07:53:10.008165 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Feb 27 07:53:10 crc kubenswrapper[4612]: I0227 07:53:10.008312 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Feb 27 07:53:10 crc kubenswrapper[4612]: I0227 07:53:10.270168 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Feb 27 07:53:10 crc kubenswrapper[4612]: I0227 07:53:10.284229 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Feb 27 07:53:10 crc kubenswrapper[4612]: I0227 07:53:10.477163 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Feb 27 07:53:10 crc kubenswrapper[4612]: I0227 07:53:10.568641 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Feb 27 07:53:10 crc kubenswrapper[4612]: I0227 07:53:10.623221 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Feb 27 07:53:10 crc kubenswrapper[4612]: I0227 07:53:10.796952 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Feb 27 07:53:10 crc kubenswrapper[4612]: I0227 07:53:10.808725 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Feb 27 07:53:10 crc kubenswrapper[4612]: I0227 07:53:10.844577 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Feb 27 07:53:10 crc kubenswrapper[4612]: I0227 07:53:10.914046 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Feb 27 07:53:10 crc kubenswrapper[4612]: I0227 07:53:10.943533 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Feb 27 07:53:10 crc kubenswrapper[4612]: I0227 07:53:10.943770 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Feb 27 07:53:10 crc kubenswrapper[4612]: I0227 07:53:10.967285 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Feb 27 07:53:10 crc kubenswrapper[4612]: I0227 07:53:10.972936 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Feb 27 07:53:10 crc kubenswrapper[4612]: I0227 07:53:10.996349 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Feb 27 07:53:11 crc kubenswrapper[4612]: I0227 07:53:11.042438 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Feb 27 07:53:11 crc kubenswrapper[4612]: I0227 07:53:11.083517 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Feb 27 07:53:11 crc kubenswrapper[4612]: I0227 07:53:11.165229 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Feb 27 07:53:11 crc kubenswrapper[4612]: I0227 07:53:11.214271 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Feb 27 07:53:11 crc kubenswrapper[4612]: I0227 07:53:11.268874 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Feb 27 07:53:11 crc kubenswrapper[4612]: I0227 07:53:11.342336 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Feb 27 07:53:11 crc kubenswrapper[4612]: I0227 07:53:11.392886 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Feb 27 07:53:11 crc kubenswrapper[4612]: I0227 07:53:11.702344 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Feb 27 07:53:11 crc kubenswrapper[4612]: I0227 07:53:11.723182 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Feb 27 07:53:11 crc kubenswrapper[4612]: I0227 07:53:11.748755 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Feb 27 07:53:11 crc kubenswrapper[4612]: I0227 07:53:11.841774 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Feb 27 07:53:11 crc kubenswrapper[4612]: I0227 07:53:11.900403 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Feb 27 07:53:12 crc kubenswrapper[4612]: I0227 07:53:12.011750 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Feb 27 07:53:12 crc kubenswrapper[4612]: I0227 07:53:12.020415 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Feb 27 07:53:12 crc kubenswrapper[4612]: I0227 07:53:12.057600 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Feb 27 07:53:12 crc kubenswrapper[4612]: I0227 07:53:12.132932 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Feb 27 07:53:12 crc kubenswrapper[4612]: I0227 07:53:12.147000 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Feb 27 07:53:12 crc kubenswrapper[4612]: I0227 07:53:12.211199 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Feb 27 07:53:12 crc kubenswrapper[4612]: I0227 07:53:12.295599 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Feb 27 07:53:12 crc kubenswrapper[4612]: I0227 07:53:12.312078 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Feb 27 07:53:12 crc kubenswrapper[4612]: I0227 07:53:12.322346 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Feb 27 07:53:12 crc kubenswrapper[4612]: I0227 07:53:12.327290 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Feb 27 07:53:12 crc kubenswrapper[4612]: I0227 07:53:12.356641 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Feb 27 07:53:12 crc kubenswrapper[4612]: I0227 07:53:12.387295 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Feb 27 07:53:12 crc kubenswrapper[4612]: I0227 07:53:12.634293 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Feb 27 07:53:12 crc kubenswrapper[4612]: I0227 07:53:12.674174 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Feb 27 07:53:12 crc kubenswrapper[4612]: I0227 07:53:12.719450 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Feb 27 07:53:12 crc kubenswrapper[4612]: I0227 07:53:12.762471 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Feb 27 07:53:12 crc kubenswrapper[4612]: I0227 07:53:12.788978 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Feb 27 07:53:12 crc kubenswrapper[4612]: I0227 07:53:12.797939 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Feb 27 07:53:12 crc kubenswrapper[4612]: I0227 07:53:12.807489 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Feb 27 07:53:12 crc kubenswrapper[4612]: I0227 07:53:12.956394 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Feb 27 07:53:12 crc kubenswrapper[4612]: I0227 07:53:12.970934 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Feb 27 07:53:13 crc kubenswrapper[4612]: I0227 07:53:13.035440 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Feb 27 07:53:13 crc kubenswrapper[4612]: I0227 07:53:13.121521 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Feb 27 07:53:13 crc kubenswrapper[4612]: I0227 07:53:13.168400 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Feb 27 07:53:13 crc kubenswrapper[4612]: I0227 07:53:13.249717 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Feb 27 07:53:13 crc kubenswrapper[4612]: I0227 07:53:13.296007 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Feb 27 07:53:13 crc kubenswrapper[4612]: I0227 07:53:13.332459 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Feb 27 07:53:13 crc kubenswrapper[4612]: I0227 07:53:13.350424 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Feb 27 07:53:13 crc kubenswrapper[4612]: I0227 07:53:13.429528 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Feb 27 07:53:13 crc kubenswrapper[4612]: I0227 07:53:13.459848 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Feb 27 07:53:13 crc kubenswrapper[4612]: I0227 07:53:13.485657 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Feb 27 07:53:13 crc kubenswrapper[4612]: I0227 07:53:13.535195 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Feb 27 07:53:13 crc kubenswrapper[4612]: I0227 07:53:13.825713 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Feb 27 07:53:13 crc kubenswrapper[4612]: I0227 07:53:13.873557 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Feb 27 07:53:13 crc kubenswrapper[4612]: I0227 07:53:13.889860 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Feb 27 07:53:14 crc kubenswrapper[4612]: I0227 07:53:14.009665 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Feb 27 07:53:14 crc kubenswrapper[4612]: I0227 07:53:14.077268 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Feb 27 07:53:14 crc kubenswrapper[4612]: I0227 07:53:14.134940 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Feb 27 07:53:14 crc kubenswrapper[4612]: I0227 07:53:14.166935 4612 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Feb 27 07:53:14 crc kubenswrapper[4612]: I0227 07:53:14.174264 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Feb 27 07:53:14 crc kubenswrapper[4612]: I0227 07:53:14.190865 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Feb 27 07:53:14 crc kubenswrapper[4612]: I0227 07:53:14.251789 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Feb 27 07:53:14 crc kubenswrapper[4612]: I0227 07:53:14.303416 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Feb 27 07:53:14 crc kubenswrapper[4612]: I0227 07:53:14.368762 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Feb 27 07:53:14 crc kubenswrapper[4612]: I0227 07:53:14.391934 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Feb 27 07:53:14 crc kubenswrapper[4612]: I0227 07:53:14.454474 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Feb 27 07:53:14 crc kubenswrapper[4612]: I0227 07:53:14.547760 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Feb 27 07:53:14 crc kubenswrapper[4612]: I0227 07:53:14.577360 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Feb 27 07:53:14 crc kubenswrapper[4612]: I0227 07:53:14.668454 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Feb 27 07:53:14 crc kubenswrapper[4612]: I0227 07:53:14.725452 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Feb 27 07:53:14 crc kubenswrapper[4612]: I0227 07:53:14.736413 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Feb 27 07:53:14 crc kubenswrapper[4612]: I0227 07:53:14.742535 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Feb 27 07:53:14 crc kubenswrapper[4612]: I0227 07:53:14.773188 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Feb 27 07:53:14 crc kubenswrapper[4612]: I0227 07:53:14.821848 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Feb 27 07:53:14 crc kubenswrapper[4612]: I0227 07:53:14.867352 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Feb 27 07:53:14 crc kubenswrapper[4612]: I0227 07:53:14.871277 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Feb 27 07:53:14 crc kubenswrapper[4612]: I0227 07:53:14.871418 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Feb 27 07:53:15 crc kubenswrapper[4612]: I0227 07:53:15.040368 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Feb 27 07:53:15 crc kubenswrapper[4612]: I0227 07:53:15.049248 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Feb 27 07:53:15 crc kubenswrapper[4612]: I0227 07:53:15.092802 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Feb 27 07:53:15 crc kubenswrapper[4612]: I0227 07:53:15.120376 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Feb 27 07:53:15 crc kubenswrapper[4612]: I0227 07:53:15.241464 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Feb 27 07:53:15 crc kubenswrapper[4612]: I0227 07:53:15.290164 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Feb 27 07:53:15 crc kubenswrapper[4612]: I0227 07:53:15.474773 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Feb 27 07:53:15 crc kubenswrapper[4612]: I0227 07:53:15.485940 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Feb 27 07:53:15 crc kubenswrapper[4612]: I0227 07:53:15.535973 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Feb 27 07:53:15 crc kubenswrapper[4612]: I0227 07:53:15.553017 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Feb 27 07:53:15 crc kubenswrapper[4612]: I0227 07:53:15.591516 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Feb 27 07:53:15 crc kubenswrapper[4612]: I0227 07:53:15.620152 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Feb 27 07:53:15 crc kubenswrapper[4612]: I0227 07:53:15.647435 4612 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Feb 27 07:53:15 crc kubenswrapper[4612]: I0227 07:53:15.663091 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Feb 27 07:53:15 crc kubenswrapper[4612]: I0227 07:53:15.692808 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Feb 27 07:53:15 crc kubenswrapper[4612]: I0227 07:53:15.730176 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Feb 27 07:53:15 crc kubenswrapper[4612]: I0227 07:53:15.735212 4612 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Feb 27 07:53:15 crc kubenswrapper[4612]: I0227 07:53:15.735440 4612 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" containerID="cri-o://cf5178b3d4f12046a14455df679931fd7394d9e3c06206130a4bf955e6d1525b" gracePeriod=5 Feb 27 07:53:15 crc kubenswrapper[4612]: I0227 07:53:15.784296 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Feb 27 07:53:15 crc kubenswrapper[4612]: I0227 07:53:15.784339 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Feb 27 07:53:15 crc kubenswrapper[4612]: I0227 07:53:15.844810 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Feb 27 07:53:15 crc kubenswrapper[4612]: I0227 07:53:15.850135 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Feb 27 07:53:15 crc kubenswrapper[4612]: I0227 07:53:15.863362 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Feb 27 07:53:15 crc kubenswrapper[4612]: I0227 07:53:15.886887 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Feb 27 07:53:15 crc kubenswrapper[4612]: I0227 07:53:15.892759 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Feb 27 07:53:15 crc kubenswrapper[4612]: I0227 07:53:15.923053 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Feb 27 07:53:16 crc kubenswrapper[4612]: I0227 07:53:16.026711 4612 patch_prober.go:28] interesting pod/machine-config-daemon-924vb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 27 07:53:16 crc kubenswrapper[4612]: I0227 07:53:16.026768 4612 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 27 07:53:16 crc kubenswrapper[4612]: I0227 07:53:16.049450 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Feb 27 07:53:16 crc kubenswrapper[4612]: I0227 07:53:16.139559 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Feb 27 07:53:16 crc kubenswrapper[4612]: I0227 07:53:16.228189 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Feb 27 07:53:16 crc kubenswrapper[4612]: I0227 07:53:16.346614 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Feb 27 07:53:16 crc kubenswrapper[4612]: I0227 07:53:16.640798 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Feb 27 07:53:16 crc kubenswrapper[4612]: I0227 07:53:16.662199 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Feb 27 07:53:16 crc kubenswrapper[4612]: I0227 07:53:16.662856 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Feb 27 07:53:16 crc kubenswrapper[4612]: I0227 07:53:16.754798 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Feb 27 07:53:16 crc kubenswrapper[4612]: I0227 07:53:16.806243 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Feb 27 07:53:16 crc kubenswrapper[4612]: I0227 07:53:16.889827 4612 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Feb 27 07:53:16 crc kubenswrapper[4612]: I0227 07:53:16.907719 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Feb 27 07:53:17 crc kubenswrapper[4612]: I0227 07:53:17.076323 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Feb 27 07:53:17 crc kubenswrapper[4612]: I0227 07:53:17.097548 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Feb 27 07:53:17 crc kubenswrapper[4612]: I0227 07:53:17.160685 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Feb 27 07:53:17 crc kubenswrapper[4612]: I0227 07:53:17.366215 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Feb 27 07:53:17 crc kubenswrapper[4612]: I0227 07:53:17.393766 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Feb 27 07:53:17 crc kubenswrapper[4612]: I0227 07:53:17.461317 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Feb 27 07:53:17 crc kubenswrapper[4612]: I0227 07:53:17.610500 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Feb 27 07:53:17 crc kubenswrapper[4612]: I0227 07:53:17.778566 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Feb 27 07:53:17 crc kubenswrapper[4612]: I0227 07:53:17.787989 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Feb 27 07:53:17 crc kubenswrapper[4612]: I0227 07:53:17.839450 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Feb 27 07:53:17 crc kubenswrapper[4612]: I0227 07:53:17.971544 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Feb 27 07:53:17 crc kubenswrapper[4612]: I0227 07:53:17.996996 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Feb 27 07:53:18 crc kubenswrapper[4612]: I0227 07:53:18.021039 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Feb 27 07:53:18 crc kubenswrapper[4612]: I0227 07:53:18.037997 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Feb 27 07:53:18 crc kubenswrapper[4612]: I0227 07:53:18.104303 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Feb 27 07:53:18 crc kubenswrapper[4612]: I0227 07:53:18.262243 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Feb 27 07:53:18 crc kubenswrapper[4612]: I0227 07:53:18.392712 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Feb 27 07:53:18 crc kubenswrapper[4612]: I0227 07:53:18.419283 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Feb 27 07:53:18 crc kubenswrapper[4612]: I0227 07:53:18.518989 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Feb 27 07:53:18 crc kubenswrapper[4612]: I0227 07:53:18.753643 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Feb 27 07:53:18 crc kubenswrapper[4612]: I0227 07:53:18.914204 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Feb 27 07:53:18 crc kubenswrapper[4612]: I0227 07:53:18.966614 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Feb 27 07:53:19 crc kubenswrapper[4612]: I0227 07:53:19.049329 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Feb 27 07:53:19 crc kubenswrapper[4612]: I0227 07:53:19.455320 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Feb 27 07:53:19 crc kubenswrapper[4612]: I0227 07:53:19.545892 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Feb 27 07:53:19 crc kubenswrapper[4612]: I0227 07:53:19.887562 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Feb 27 07:53:20 crc kubenswrapper[4612]: I0227 07:53:20.453455 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Feb 27 07:53:20 crc kubenswrapper[4612]: I0227 07:53:20.533763 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Feb 27 07:53:20 crc kubenswrapper[4612]: I0227 07:53:20.689229 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Feb 27 07:53:20 crc kubenswrapper[4612]: I0227 07:53:20.784826 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Feb 27 07:53:20 crc kubenswrapper[4612]: I0227 07:53:20.868248 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Feb 27 07:53:21 crc kubenswrapper[4612]: I0227 07:53:21.014308 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Feb 27 07:53:21 crc kubenswrapper[4612]: I0227 07:53:21.326521 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Feb 27 07:53:21 crc kubenswrapper[4612]: I0227 07:53:21.326589 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 27 07:53:21 crc kubenswrapper[4612]: I0227 07:53:21.505582 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Feb 27 07:53:21 crc kubenswrapper[4612]: I0227 07:53:21.505772 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Feb 27 07:53:21 crc kubenswrapper[4612]: I0227 07:53:21.505803 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Feb 27 07:53:21 crc kubenswrapper[4612]: I0227 07:53:21.505831 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Feb 27 07:53:21 crc kubenswrapper[4612]: I0227 07:53:21.505854 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Feb 27 07:53:21 crc kubenswrapper[4612]: I0227 07:53:21.505966 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log" (OuterVolumeSpecName: "var-log") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 27 07:53:21 crc kubenswrapper[4612]: I0227 07:53:21.505970 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests" (OuterVolumeSpecName: "manifests") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "manifests". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 27 07:53:21 crc kubenswrapper[4612]: I0227 07:53:21.506015 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock" (OuterVolumeSpecName: "var-lock") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 27 07:53:21 crc kubenswrapper[4612]: I0227 07:53:21.505993 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 27 07:53:21 crc kubenswrapper[4612]: I0227 07:53:21.506233 4612 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") on node \"crc\" DevicePath \"\"" Feb 27 07:53:21 crc kubenswrapper[4612]: I0227 07:53:21.506256 4612 reconciler_common.go:293] "Volume detached for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") on node \"crc\" DevicePath \"\"" Feb 27 07:53:21 crc kubenswrapper[4612]: I0227 07:53:21.506269 4612 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") on node \"crc\" DevicePath \"\"" Feb 27 07:53:21 crc kubenswrapper[4612]: I0227 07:53:21.506280 4612 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") on node \"crc\" DevicePath \"\"" Feb 27 07:53:21 crc kubenswrapper[4612]: I0227 07:53:21.516654 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir" (OuterVolumeSpecName: "pod-resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "pod-resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 27 07:53:21 crc kubenswrapper[4612]: I0227 07:53:21.608029 4612 reconciler_common.go:293] "Volume detached for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") on node \"crc\" DevicePath \"\"" Feb 27 07:53:21 crc kubenswrapper[4612]: I0227 07:53:21.609879 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Feb 27 07:53:21 crc kubenswrapper[4612]: I0227 07:53:21.609954 4612 generic.go:334] "Generic (PLEG): container finished" podID="f85e55b1a89d02b0cb034b1ea31ed45a" containerID="cf5178b3d4f12046a14455df679931fd7394d9e3c06206130a4bf955e6d1525b" exitCode=137 Feb 27 07:53:21 crc kubenswrapper[4612]: I0227 07:53:21.610025 4612 scope.go:117] "RemoveContainer" containerID="cf5178b3d4f12046a14455df679931fd7394d9e3c06206130a4bf955e6d1525b" Feb 27 07:53:21 crc kubenswrapper[4612]: I0227 07:53:21.610277 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 27 07:53:21 crc kubenswrapper[4612]: I0227 07:53:21.638878 4612 scope.go:117] "RemoveContainer" containerID="cf5178b3d4f12046a14455df679931fd7394d9e3c06206130a4bf955e6d1525b" Feb 27 07:53:21 crc kubenswrapper[4612]: E0227 07:53:21.639357 4612 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cf5178b3d4f12046a14455df679931fd7394d9e3c06206130a4bf955e6d1525b\": container with ID starting with cf5178b3d4f12046a14455df679931fd7394d9e3c06206130a4bf955e6d1525b not found: ID does not exist" containerID="cf5178b3d4f12046a14455df679931fd7394d9e3c06206130a4bf955e6d1525b" Feb 27 07:53:21 crc kubenswrapper[4612]: I0227 07:53:21.639462 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cf5178b3d4f12046a14455df679931fd7394d9e3c06206130a4bf955e6d1525b"} err="failed to get container status \"cf5178b3d4f12046a14455df679931fd7394d9e3c06206130a4bf955e6d1525b\": rpc error: code = NotFound desc = could not find container \"cf5178b3d4f12046a14455df679931fd7394d9e3c06206130a4bf955e6d1525b\": container with ID starting with cf5178b3d4f12046a14455df679931fd7394d9e3c06206130a4bf955e6d1525b not found: ID does not exist" Feb 27 07:53:22 crc kubenswrapper[4612]: I0227 07:53:22.179981 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-f9b49f89c-mxx7s"] Feb 27 07:53:22 crc kubenswrapper[4612]: I0227 07:53:22.180325 4612 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-f9b49f89c-mxx7s" podUID="c243dedd-5585-4353-9303-6e42ebc82f69" containerName="controller-manager" containerID="cri-o://87605c6e32084b6a4ded19b4a44467818afafd3718fec41aa9a794efbd0c3a3d" gracePeriod=30 Feb 27 07:53:22 crc kubenswrapper[4612]: I0227 07:53:22.190295 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-77df9fd6fb-6kkpl"] Feb 27 07:53:22 crc kubenswrapper[4612]: I0227 07:53:22.190568 4612 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-77df9fd6fb-6kkpl" podUID="a7b8e4c7-3bf3-4ef0-8aca-96037f1eea24" containerName="route-controller-manager" containerID="cri-o://0c781d8c7babab9dc74f6dbb6ccc001984e6d120e91846d2e7b2f8be375c7123" gracePeriod=30 Feb 27 07:53:22 crc kubenswrapper[4612]: I0227 07:53:22.506424 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-f9b49f89c-mxx7s" Feb 27 07:53:22 crc kubenswrapper[4612]: I0227 07:53:22.584756 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-77df9fd6fb-6kkpl" Feb 27 07:53:22 crc kubenswrapper[4612]: I0227 07:53:22.614764 4612 generic.go:334] "Generic (PLEG): container finished" podID="a7b8e4c7-3bf3-4ef0-8aca-96037f1eea24" containerID="0c781d8c7babab9dc74f6dbb6ccc001984e6d120e91846d2e7b2f8be375c7123" exitCode=0 Feb 27 07:53:22 crc kubenswrapper[4612]: I0227 07:53:22.614828 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-77df9fd6fb-6kkpl" event={"ID":"a7b8e4c7-3bf3-4ef0-8aca-96037f1eea24","Type":"ContainerDied","Data":"0c781d8c7babab9dc74f6dbb6ccc001984e6d120e91846d2e7b2f8be375c7123"} Feb 27 07:53:22 crc kubenswrapper[4612]: I0227 07:53:22.614853 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-77df9fd6fb-6kkpl" event={"ID":"a7b8e4c7-3bf3-4ef0-8aca-96037f1eea24","Type":"ContainerDied","Data":"b86493dc28f0a7bfaa381c94a396a422619e0e6ddf79946d3ffd692e9fc59b8a"} Feb 27 07:53:22 crc kubenswrapper[4612]: I0227 07:53:22.614869 4612 scope.go:117] "RemoveContainer" containerID="0c781d8c7babab9dc74f6dbb6ccc001984e6d120e91846d2e7b2f8be375c7123" Feb 27 07:53:22 crc kubenswrapper[4612]: I0227 07:53:22.614867 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-77df9fd6fb-6kkpl" Feb 27 07:53:22 crc kubenswrapper[4612]: I0227 07:53:22.617959 4612 generic.go:334] "Generic (PLEG): container finished" podID="c243dedd-5585-4353-9303-6e42ebc82f69" containerID="87605c6e32084b6a4ded19b4a44467818afafd3718fec41aa9a794efbd0c3a3d" exitCode=0 Feb 27 07:53:22 crc kubenswrapper[4612]: I0227 07:53:22.618013 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-f9b49f89c-mxx7s" event={"ID":"c243dedd-5585-4353-9303-6e42ebc82f69","Type":"ContainerDied","Data":"87605c6e32084b6a4ded19b4a44467818afafd3718fec41aa9a794efbd0c3a3d"} Feb 27 07:53:22 crc kubenswrapper[4612]: I0227 07:53:22.618034 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-f9b49f89c-mxx7s" event={"ID":"c243dedd-5585-4353-9303-6e42ebc82f69","Type":"ContainerDied","Data":"13de94f2eb37fcf0f1e4f97d7dee8d76bdff281fa6013dd9ecb3003e524d9c70"} Feb 27 07:53:22 crc kubenswrapper[4612]: I0227 07:53:22.618103 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-f9b49f89c-mxx7s" Feb 27 07:53:22 crc kubenswrapper[4612]: I0227 07:53:22.627836 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c243dedd-5585-4353-9303-6e42ebc82f69-serving-cert\") pod \"c243dedd-5585-4353-9303-6e42ebc82f69\" (UID: \"c243dedd-5585-4353-9303-6e42ebc82f69\") " Feb 27 07:53:22 crc kubenswrapper[4612]: I0227 07:53:22.627870 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c243dedd-5585-4353-9303-6e42ebc82f69-client-ca\") pod \"c243dedd-5585-4353-9303-6e42ebc82f69\" (UID: \"c243dedd-5585-4353-9303-6e42ebc82f69\") " Feb 27 07:53:22 crc kubenswrapper[4612]: I0227 07:53:22.627911 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c243dedd-5585-4353-9303-6e42ebc82f69-config\") pod \"c243dedd-5585-4353-9303-6e42ebc82f69\" (UID: \"c243dedd-5585-4353-9303-6e42ebc82f69\") " Feb 27 07:53:22 crc kubenswrapper[4612]: I0227 07:53:22.627982 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c243dedd-5585-4353-9303-6e42ebc82f69-proxy-ca-bundles\") pod \"c243dedd-5585-4353-9303-6e42ebc82f69\" (UID: \"c243dedd-5585-4353-9303-6e42ebc82f69\") " Feb 27 07:53:22 crc kubenswrapper[4612]: I0227 07:53:22.628010 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ql99m\" (UniqueName: \"kubernetes.io/projected/c243dedd-5585-4353-9303-6e42ebc82f69-kube-api-access-ql99m\") pod \"c243dedd-5585-4353-9303-6e42ebc82f69\" (UID: \"c243dedd-5585-4353-9303-6e42ebc82f69\") " Feb 27 07:53:22 crc kubenswrapper[4612]: I0227 07:53:22.628230 4612 scope.go:117] "RemoveContainer" containerID="0c781d8c7babab9dc74f6dbb6ccc001984e6d120e91846d2e7b2f8be375c7123" Feb 27 07:53:22 crc kubenswrapper[4612]: E0227 07:53:22.628565 4612 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0c781d8c7babab9dc74f6dbb6ccc001984e6d120e91846d2e7b2f8be375c7123\": container with ID starting with 0c781d8c7babab9dc74f6dbb6ccc001984e6d120e91846d2e7b2f8be375c7123 not found: ID does not exist" containerID="0c781d8c7babab9dc74f6dbb6ccc001984e6d120e91846d2e7b2f8be375c7123" Feb 27 07:53:22 crc kubenswrapper[4612]: I0227 07:53:22.628602 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0c781d8c7babab9dc74f6dbb6ccc001984e6d120e91846d2e7b2f8be375c7123"} err="failed to get container status \"0c781d8c7babab9dc74f6dbb6ccc001984e6d120e91846d2e7b2f8be375c7123\": rpc error: code = NotFound desc = could not find container \"0c781d8c7babab9dc74f6dbb6ccc001984e6d120e91846d2e7b2f8be375c7123\": container with ID starting with 0c781d8c7babab9dc74f6dbb6ccc001984e6d120e91846d2e7b2f8be375c7123 not found: ID does not exist" Feb 27 07:53:22 crc kubenswrapper[4612]: I0227 07:53:22.628627 4612 scope.go:117] "RemoveContainer" containerID="87605c6e32084b6a4ded19b4a44467818afafd3718fec41aa9a794efbd0c3a3d" Feb 27 07:53:22 crc kubenswrapper[4612]: I0227 07:53:22.628966 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c243dedd-5585-4353-9303-6e42ebc82f69-config" (OuterVolumeSpecName: "config") pod "c243dedd-5585-4353-9303-6e42ebc82f69" (UID: "c243dedd-5585-4353-9303-6e42ebc82f69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 07:53:22 crc kubenswrapper[4612]: I0227 07:53:22.629082 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c243dedd-5585-4353-9303-6e42ebc82f69-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "c243dedd-5585-4353-9303-6e42ebc82f69" (UID: "c243dedd-5585-4353-9303-6e42ebc82f69"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 07:53:22 crc kubenswrapper[4612]: I0227 07:53:22.630258 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c243dedd-5585-4353-9303-6e42ebc82f69-client-ca" (OuterVolumeSpecName: "client-ca") pod "c243dedd-5585-4353-9303-6e42ebc82f69" (UID: "c243dedd-5585-4353-9303-6e42ebc82f69"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 07:53:22 crc kubenswrapper[4612]: I0227 07:53:22.632979 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c243dedd-5585-4353-9303-6e42ebc82f69-kube-api-access-ql99m" (OuterVolumeSpecName: "kube-api-access-ql99m") pod "c243dedd-5585-4353-9303-6e42ebc82f69" (UID: "c243dedd-5585-4353-9303-6e42ebc82f69"). InnerVolumeSpecName "kube-api-access-ql99m". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 07:53:22 crc kubenswrapper[4612]: I0227 07:53:22.633452 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c243dedd-5585-4353-9303-6e42ebc82f69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "c243dedd-5585-4353-9303-6e42ebc82f69" (UID: "c243dedd-5585-4353-9303-6e42ebc82f69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 07:53:22 crc kubenswrapper[4612]: I0227 07:53:22.641812 4612 scope.go:117] "RemoveContainer" containerID="87605c6e32084b6a4ded19b4a44467818afafd3718fec41aa9a794efbd0c3a3d" Feb 27 07:53:22 crc kubenswrapper[4612]: E0227 07:53:22.642223 4612 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"87605c6e32084b6a4ded19b4a44467818afafd3718fec41aa9a794efbd0c3a3d\": container with ID starting with 87605c6e32084b6a4ded19b4a44467818afafd3718fec41aa9a794efbd0c3a3d not found: ID does not exist" containerID="87605c6e32084b6a4ded19b4a44467818afafd3718fec41aa9a794efbd0c3a3d" Feb 27 07:53:22 crc kubenswrapper[4612]: I0227 07:53:22.642268 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"87605c6e32084b6a4ded19b4a44467818afafd3718fec41aa9a794efbd0c3a3d"} err="failed to get container status \"87605c6e32084b6a4ded19b4a44467818afafd3718fec41aa9a794efbd0c3a3d\": rpc error: code = NotFound desc = could not find container \"87605c6e32084b6a4ded19b4a44467818afafd3718fec41aa9a794efbd0c3a3d\": container with ID starting with 87605c6e32084b6a4ded19b4a44467818afafd3718fec41aa9a794efbd0c3a3d not found: ID does not exist" Feb 27 07:53:22 crc kubenswrapper[4612]: I0227 07:53:22.729061 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qpq69\" (UniqueName: \"kubernetes.io/projected/a7b8e4c7-3bf3-4ef0-8aca-96037f1eea24-kube-api-access-qpq69\") pod \"a7b8e4c7-3bf3-4ef0-8aca-96037f1eea24\" (UID: \"a7b8e4c7-3bf3-4ef0-8aca-96037f1eea24\") " Feb 27 07:53:22 crc kubenswrapper[4612]: I0227 07:53:22.729239 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a7b8e4c7-3bf3-4ef0-8aca-96037f1eea24-client-ca\") pod \"a7b8e4c7-3bf3-4ef0-8aca-96037f1eea24\" (UID: \"a7b8e4c7-3bf3-4ef0-8aca-96037f1eea24\") " Feb 27 07:53:22 crc kubenswrapper[4612]: I0227 07:53:22.729326 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a7b8e4c7-3bf3-4ef0-8aca-96037f1eea24-config\") pod \"a7b8e4c7-3bf3-4ef0-8aca-96037f1eea24\" (UID: \"a7b8e4c7-3bf3-4ef0-8aca-96037f1eea24\") " Feb 27 07:53:22 crc kubenswrapper[4612]: I0227 07:53:22.730317 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a7b8e4c7-3bf3-4ef0-8aca-96037f1eea24-config" (OuterVolumeSpecName: "config") pod "a7b8e4c7-3bf3-4ef0-8aca-96037f1eea24" (UID: "a7b8e4c7-3bf3-4ef0-8aca-96037f1eea24"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 07:53:22 crc kubenswrapper[4612]: I0227 07:53:22.730666 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a7b8e4c7-3bf3-4ef0-8aca-96037f1eea24-client-ca" (OuterVolumeSpecName: "client-ca") pod "a7b8e4c7-3bf3-4ef0-8aca-96037f1eea24" (UID: "a7b8e4c7-3bf3-4ef0-8aca-96037f1eea24"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 07:53:22 crc kubenswrapper[4612]: I0227 07:53:22.730748 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a7b8e4c7-3bf3-4ef0-8aca-96037f1eea24-serving-cert\") pod \"a7b8e4c7-3bf3-4ef0-8aca-96037f1eea24\" (UID: \"a7b8e4c7-3bf3-4ef0-8aca-96037f1eea24\") " Feb 27 07:53:22 crc kubenswrapper[4612]: I0227 07:53:22.731656 4612 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c243dedd-5585-4353-9303-6e42ebc82f69-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 27 07:53:22 crc kubenswrapper[4612]: I0227 07:53:22.731683 4612 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c243dedd-5585-4353-9303-6e42ebc82f69-client-ca\") on node \"crc\" DevicePath \"\"" Feb 27 07:53:22 crc kubenswrapper[4612]: I0227 07:53:22.731722 4612 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c243dedd-5585-4353-9303-6e42ebc82f69-config\") on node \"crc\" DevicePath \"\"" Feb 27 07:53:22 crc kubenswrapper[4612]: I0227 07:53:22.731733 4612 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a7b8e4c7-3bf3-4ef0-8aca-96037f1eea24-client-ca\") on node \"crc\" DevicePath \"\"" Feb 27 07:53:22 crc kubenswrapper[4612]: I0227 07:53:22.731746 4612 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c243dedd-5585-4353-9303-6e42ebc82f69-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Feb 27 07:53:22 crc kubenswrapper[4612]: I0227 07:53:22.731758 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ql99m\" (UniqueName: \"kubernetes.io/projected/c243dedd-5585-4353-9303-6e42ebc82f69-kube-api-access-ql99m\") on node \"crc\" DevicePath \"\"" Feb 27 07:53:22 crc kubenswrapper[4612]: I0227 07:53:22.731769 4612 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a7b8e4c7-3bf3-4ef0-8aca-96037f1eea24-config\") on node \"crc\" DevicePath \"\"" Feb 27 07:53:22 crc kubenswrapper[4612]: I0227 07:53:22.732752 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a7b8e4c7-3bf3-4ef0-8aca-96037f1eea24-kube-api-access-qpq69" (OuterVolumeSpecName: "kube-api-access-qpq69") pod "a7b8e4c7-3bf3-4ef0-8aca-96037f1eea24" (UID: "a7b8e4c7-3bf3-4ef0-8aca-96037f1eea24"). InnerVolumeSpecName "kube-api-access-qpq69". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 07:53:22 crc kubenswrapper[4612]: I0227 07:53:22.733735 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a7b8e4c7-3bf3-4ef0-8aca-96037f1eea24-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "a7b8e4c7-3bf3-4ef0-8aca-96037f1eea24" (UID: "a7b8e4c7-3bf3-4ef0-8aca-96037f1eea24"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 07:53:22 crc kubenswrapper[4612]: I0227 07:53:22.832621 4612 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a7b8e4c7-3bf3-4ef0-8aca-96037f1eea24-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 27 07:53:22 crc kubenswrapper[4612]: I0227 07:53:22.832965 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qpq69\" (UniqueName: \"kubernetes.io/projected/a7b8e4c7-3bf3-4ef0-8aca-96037f1eea24-kube-api-access-qpq69\") on node \"crc\" DevicePath \"\"" Feb 27 07:53:22 crc kubenswrapper[4612]: I0227 07:53:22.863984 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" path="/var/lib/kubelet/pods/f85e55b1a89d02b0cb034b1ea31ed45a/volumes" Feb 27 07:53:22 crc kubenswrapper[4612]: I0227 07:53:22.864899 4612 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="" Feb 27 07:53:22 crc kubenswrapper[4612]: I0227 07:53:22.887777 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Feb 27 07:53:22 crc kubenswrapper[4612]: I0227 07:53:22.887864 4612 kubelet.go:2649] "Unable to find pod for mirror pod, skipping" mirrorPod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" mirrorPodUID="ee3f42f5-5dad-44e4-b28f-d7bf6387ed12" Feb 27 07:53:22 crc kubenswrapper[4612]: I0227 07:53:22.897801 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Feb 27 07:53:22 crc kubenswrapper[4612]: I0227 07:53:22.898097 4612 kubelet.go:2673] "Unable to find pod for mirror pod, skipping" mirrorPod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" mirrorPodUID="ee3f42f5-5dad-44e4-b28f-d7bf6387ed12" Feb 27 07:53:22 crc kubenswrapper[4612]: I0227 07:53:22.946255 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-77df9fd6fb-6kkpl"] Feb 27 07:53:22 crc kubenswrapper[4612]: I0227 07:53:22.970541 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-77df9fd6fb-6kkpl"] Feb 27 07:53:22 crc kubenswrapper[4612]: I0227 07:53:22.970638 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-f9b49f89c-mxx7s"] Feb 27 07:53:22 crc kubenswrapper[4612]: I0227 07:53:22.972991 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-f9b49f89c-mxx7s"] Feb 27 07:53:23 crc kubenswrapper[4612]: I0227 07:53:23.810837 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-7c6f84b55d-wrdhn"] Feb 27 07:53:23 crc kubenswrapper[4612]: E0227 07:53:23.811221 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c243dedd-5585-4353-9303-6e42ebc82f69" containerName="controller-manager" Feb 27 07:53:23 crc kubenswrapper[4612]: I0227 07:53:23.811243 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="c243dedd-5585-4353-9303-6e42ebc82f69" containerName="controller-manager" Feb 27 07:53:23 crc kubenswrapper[4612]: E0227 07:53:23.811273 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a7b8e4c7-3bf3-4ef0-8aca-96037f1eea24" containerName="route-controller-manager" Feb 27 07:53:23 crc kubenswrapper[4612]: I0227 07:53:23.811287 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="a7b8e4c7-3bf3-4ef0-8aca-96037f1eea24" containerName="route-controller-manager" Feb 27 07:53:23 crc kubenswrapper[4612]: E0227 07:53:23.811315 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Feb 27 07:53:23 crc kubenswrapper[4612]: I0227 07:53:23.811328 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Feb 27 07:53:23 crc kubenswrapper[4612]: E0227 07:53:23.811344 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c13ed974-66aa-467d-9d11-4b375e53b8bc" containerName="installer" Feb 27 07:53:23 crc kubenswrapper[4612]: I0227 07:53:23.811357 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="c13ed974-66aa-467d-9d11-4b375e53b8bc" containerName="installer" Feb 27 07:53:23 crc kubenswrapper[4612]: I0227 07:53:23.811527 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="c13ed974-66aa-467d-9d11-4b375e53b8bc" containerName="installer" Feb 27 07:53:23 crc kubenswrapper[4612]: I0227 07:53:23.811553 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Feb 27 07:53:23 crc kubenswrapper[4612]: I0227 07:53:23.811571 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="c243dedd-5585-4353-9303-6e42ebc82f69" containerName="controller-manager" Feb 27 07:53:23 crc kubenswrapper[4612]: I0227 07:53:23.811596 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="a7b8e4c7-3bf3-4ef0-8aca-96037f1eea24" containerName="route-controller-manager" Feb 27 07:53:23 crc kubenswrapper[4612]: I0227 07:53:23.812212 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7c6f84b55d-wrdhn" Feb 27 07:53:23 crc kubenswrapper[4612]: I0227 07:53:23.815851 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5cffc6cf87-sblh2"] Feb 27 07:53:23 crc kubenswrapper[4612]: I0227 07:53:23.817124 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5cffc6cf87-sblh2" Feb 27 07:53:23 crc kubenswrapper[4612]: I0227 07:53:23.818161 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Feb 27 07:53:23 crc kubenswrapper[4612]: I0227 07:53:23.819931 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Feb 27 07:53:23 crc kubenswrapper[4612]: I0227 07:53:23.818582 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Feb 27 07:53:23 crc kubenswrapper[4612]: I0227 07:53:23.818993 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Feb 27 07:53:23 crc kubenswrapper[4612]: I0227 07:53:23.819816 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Feb 27 07:53:23 crc kubenswrapper[4612]: I0227 07:53:23.819888 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Feb 27 07:53:23 crc kubenswrapper[4612]: I0227 07:53:23.822499 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Feb 27 07:53:23 crc kubenswrapper[4612]: I0227 07:53:23.822992 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Feb 27 07:53:23 crc kubenswrapper[4612]: I0227 07:53:23.823251 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Feb 27 07:53:23 crc kubenswrapper[4612]: I0227 07:53:23.824330 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Feb 27 07:53:23 crc kubenswrapper[4612]: I0227 07:53:23.824586 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Feb 27 07:53:23 crc kubenswrapper[4612]: I0227 07:53:23.825031 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Feb 27 07:53:23 crc kubenswrapper[4612]: I0227 07:53:23.842295 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Feb 27 07:53:23 crc kubenswrapper[4612]: I0227 07:53:23.848036 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-7c6f84b55d-wrdhn"] Feb 27 07:53:23 crc kubenswrapper[4612]: I0227 07:53:23.870643 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5cffc6cf87-sblh2"] Feb 27 07:53:23 crc kubenswrapper[4612]: I0227 07:53:23.948147 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/99f24b2c-4969-454a-bc54-1b76e67b302d-serving-cert\") pod \"route-controller-manager-5cffc6cf87-sblh2\" (UID: \"99f24b2c-4969-454a-bc54-1b76e67b302d\") " pod="openshift-route-controller-manager/route-controller-manager-5cffc6cf87-sblh2" Feb 27 07:53:23 crc kubenswrapper[4612]: I0227 07:53:23.949261 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/99f24b2c-4969-454a-bc54-1b76e67b302d-client-ca\") pod \"route-controller-manager-5cffc6cf87-sblh2\" (UID: \"99f24b2c-4969-454a-bc54-1b76e67b302d\") " pod="openshift-route-controller-manager/route-controller-manager-5cffc6cf87-sblh2" Feb 27 07:53:23 crc kubenswrapper[4612]: I0227 07:53:23.949509 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/99f24b2c-4969-454a-bc54-1b76e67b302d-config\") pod \"route-controller-manager-5cffc6cf87-sblh2\" (UID: \"99f24b2c-4969-454a-bc54-1b76e67b302d\") " pod="openshift-route-controller-manager/route-controller-manager-5cffc6cf87-sblh2" Feb 27 07:53:23 crc kubenswrapper[4612]: I0227 07:53:23.949570 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4xrxl\" (UniqueName: \"kubernetes.io/projected/99f24b2c-4969-454a-bc54-1b76e67b302d-kube-api-access-4xrxl\") pod \"route-controller-manager-5cffc6cf87-sblh2\" (UID: \"99f24b2c-4969-454a-bc54-1b76e67b302d\") " pod="openshift-route-controller-manager/route-controller-manager-5cffc6cf87-sblh2" Feb 27 07:53:23 crc kubenswrapper[4612]: I0227 07:53:23.949606 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5924d34a-c67c-45e1-8085-dafed86b5265-serving-cert\") pod \"controller-manager-7c6f84b55d-wrdhn\" (UID: \"5924d34a-c67c-45e1-8085-dafed86b5265\") " pod="openshift-controller-manager/controller-manager-7c6f84b55d-wrdhn" Feb 27 07:53:23 crc kubenswrapper[4612]: I0227 07:53:23.949661 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5924d34a-c67c-45e1-8085-dafed86b5265-client-ca\") pod \"controller-manager-7c6f84b55d-wrdhn\" (UID: \"5924d34a-c67c-45e1-8085-dafed86b5265\") " pod="openshift-controller-manager/controller-manager-7c6f84b55d-wrdhn" Feb 27 07:53:23 crc kubenswrapper[4612]: I0227 07:53:23.949805 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nl264\" (UniqueName: \"kubernetes.io/projected/5924d34a-c67c-45e1-8085-dafed86b5265-kube-api-access-nl264\") pod \"controller-manager-7c6f84b55d-wrdhn\" (UID: \"5924d34a-c67c-45e1-8085-dafed86b5265\") " pod="openshift-controller-manager/controller-manager-7c6f84b55d-wrdhn" Feb 27 07:53:23 crc kubenswrapper[4612]: I0227 07:53:23.949886 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/5924d34a-c67c-45e1-8085-dafed86b5265-proxy-ca-bundles\") pod \"controller-manager-7c6f84b55d-wrdhn\" (UID: \"5924d34a-c67c-45e1-8085-dafed86b5265\") " pod="openshift-controller-manager/controller-manager-7c6f84b55d-wrdhn" Feb 27 07:53:23 crc kubenswrapper[4612]: I0227 07:53:23.950051 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5924d34a-c67c-45e1-8085-dafed86b5265-config\") pod \"controller-manager-7c6f84b55d-wrdhn\" (UID: \"5924d34a-c67c-45e1-8085-dafed86b5265\") " pod="openshift-controller-manager/controller-manager-7c6f84b55d-wrdhn" Feb 27 07:53:24 crc kubenswrapper[4612]: I0227 07:53:24.051326 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4xrxl\" (UniqueName: \"kubernetes.io/projected/99f24b2c-4969-454a-bc54-1b76e67b302d-kube-api-access-4xrxl\") pod \"route-controller-manager-5cffc6cf87-sblh2\" (UID: \"99f24b2c-4969-454a-bc54-1b76e67b302d\") " pod="openshift-route-controller-manager/route-controller-manager-5cffc6cf87-sblh2" Feb 27 07:53:24 crc kubenswrapper[4612]: I0227 07:53:24.051413 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5924d34a-c67c-45e1-8085-dafed86b5265-serving-cert\") pod \"controller-manager-7c6f84b55d-wrdhn\" (UID: \"5924d34a-c67c-45e1-8085-dafed86b5265\") " pod="openshift-controller-manager/controller-manager-7c6f84b55d-wrdhn" Feb 27 07:53:24 crc kubenswrapper[4612]: I0227 07:53:24.051467 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5924d34a-c67c-45e1-8085-dafed86b5265-client-ca\") pod \"controller-manager-7c6f84b55d-wrdhn\" (UID: \"5924d34a-c67c-45e1-8085-dafed86b5265\") " pod="openshift-controller-manager/controller-manager-7c6f84b55d-wrdhn" Feb 27 07:53:24 crc kubenswrapper[4612]: I0227 07:53:24.051551 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nl264\" (UniqueName: \"kubernetes.io/projected/5924d34a-c67c-45e1-8085-dafed86b5265-kube-api-access-nl264\") pod \"controller-manager-7c6f84b55d-wrdhn\" (UID: \"5924d34a-c67c-45e1-8085-dafed86b5265\") " pod="openshift-controller-manager/controller-manager-7c6f84b55d-wrdhn" Feb 27 07:53:24 crc kubenswrapper[4612]: I0227 07:53:24.051601 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/5924d34a-c67c-45e1-8085-dafed86b5265-proxy-ca-bundles\") pod \"controller-manager-7c6f84b55d-wrdhn\" (UID: \"5924d34a-c67c-45e1-8085-dafed86b5265\") " pod="openshift-controller-manager/controller-manager-7c6f84b55d-wrdhn" Feb 27 07:53:24 crc kubenswrapper[4612]: I0227 07:53:24.051637 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5924d34a-c67c-45e1-8085-dafed86b5265-config\") pod \"controller-manager-7c6f84b55d-wrdhn\" (UID: \"5924d34a-c67c-45e1-8085-dafed86b5265\") " pod="openshift-controller-manager/controller-manager-7c6f84b55d-wrdhn" Feb 27 07:53:24 crc kubenswrapper[4612]: I0227 07:53:24.051725 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/99f24b2c-4969-454a-bc54-1b76e67b302d-serving-cert\") pod \"route-controller-manager-5cffc6cf87-sblh2\" (UID: \"99f24b2c-4969-454a-bc54-1b76e67b302d\") " pod="openshift-route-controller-manager/route-controller-manager-5cffc6cf87-sblh2" Feb 27 07:53:24 crc kubenswrapper[4612]: I0227 07:53:24.051781 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/99f24b2c-4969-454a-bc54-1b76e67b302d-client-ca\") pod \"route-controller-manager-5cffc6cf87-sblh2\" (UID: \"99f24b2c-4969-454a-bc54-1b76e67b302d\") " pod="openshift-route-controller-manager/route-controller-manager-5cffc6cf87-sblh2" Feb 27 07:53:24 crc kubenswrapper[4612]: I0227 07:53:24.051812 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/99f24b2c-4969-454a-bc54-1b76e67b302d-config\") pod \"route-controller-manager-5cffc6cf87-sblh2\" (UID: \"99f24b2c-4969-454a-bc54-1b76e67b302d\") " pod="openshift-route-controller-manager/route-controller-manager-5cffc6cf87-sblh2" Feb 27 07:53:24 crc kubenswrapper[4612]: I0227 07:53:24.053776 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5924d34a-c67c-45e1-8085-dafed86b5265-client-ca\") pod \"controller-manager-7c6f84b55d-wrdhn\" (UID: \"5924d34a-c67c-45e1-8085-dafed86b5265\") " pod="openshift-controller-manager/controller-manager-7c6f84b55d-wrdhn" Feb 27 07:53:24 crc kubenswrapper[4612]: I0227 07:53:24.053938 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/99f24b2c-4969-454a-bc54-1b76e67b302d-config\") pod \"route-controller-manager-5cffc6cf87-sblh2\" (UID: \"99f24b2c-4969-454a-bc54-1b76e67b302d\") " pod="openshift-route-controller-manager/route-controller-manager-5cffc6cf87-sblh2" Feb 27 07:53:24 crc kubenswrapper[4612]: I0227 07:53:24.054012 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/5924d34a-c67c-45e1-8085-dafed86b5265-proxy-ca-bundles\") pod \"controller-manager-7c6f84b55d-wrdhn\" (UID: \"5924d34a-c67c-45e1-8085-dafed86b5265\") " pod="openshift-controller-manager/controller-manager-7c6f84b55d-wrdhn" Feb 27 07:53:24 crc kubenswrapper[4612]: I0227 07:53:24.055797 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5924d34a-c67c-45e1-8085-dafed86b5265-config\") pod \"controller-manager-7c6f84b55d-wrdhn\" (UID: \"5924d34a-c67c-45e1-8085-dafed86b5265\") " pod="openshift-controller-manager/controller-manager-7c6f84b55d-wrdhn" Feb 27 07:53:24 crc kubenswrapper[4612]: I0227 07:53:24.055891 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/99f24b2c-4969-454a-bc54-1b76e67b302d-client-ca\") pod \"route-controller-manager-5cffc6cf87-sblh2\" (UID: \"99f24b2c-4969-454a-bc54-1b76e67b302d\") " pod="openshift-route-controller-manager/route-controller-manager-5cffc6cf87-sblh2" Feb 27 07:53:24 crc kubenswrapper[4612]: I0227 07:53:24.059417 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5924d34a-c67c-45e1-8085-dafed86b5265-serving-cert\") pod \"controller-manager-7c6f84b55d-wrdhn\" (UID: \"5924d34a-c67c-45e1-8085-dafed86b5265\") " pod="openshift-controller-manager/controller-manager-7c6f84b55d-wrdhn" Feb 27 07:53:24 crc kubenswrapper[4612]: I0227 07:53:24.062671 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/99f24b2c-4969-454a-bc54-1b76e67b302d-serving-cert\") pod \"route-controller-manager-5cffc6cf87-sblh2\" (UID: \"99f24b2c-4969-454a-bc54-1b76e67b302d\") " pod="openshift-route-controller-manager/route-controller-manager-5cffc6cf87-sblh2" Feb 27 07:53:24 crc kubenswrapper[4612]: I0227 07:53:24.084233 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nl264\" (UniqueName: \"kubernetes.io/projected/5924d34a-c67c-45e1-8085-dafed86b5265-kube-api-access-nl264\") pod \"controller-manager-7c6f84b55d-wrdhn\" (UID: \"5924d34a-c67c-45e1-8085-dafed86b5265\") " pod="openshift-controller-manager/controller-manager-7c6f84b55d-wrdhn" Feb 27 07:53:24 crc kubenswrapper[4612]: I0227 07:53:24.094227 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4xrxl\" (UniqueName: \"kubernetes.io/projected/99f24b2c-4969-454a-bc54-1b76e67b302d-kube-api-access-4xrxl\") pod \"route-controller-manager-5cffc6cf87-sblh2\" (UID: \"99f24b2c-4969-454a-bc54-1b76e67b302d\") " pod="openshift-route-controller-manager/route-controller-manager-5cffc6cf87-sblh2" Feb 27 07:53:24 crc kubenswrapper[4612]: I0227 07:53:24.141063 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7c6f84b55d-wrdhn" Feb 27 07:53:24 crc kubenswrapper[4612]: I0227 07:53:24.164355 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5cffc6cf87-sblh2" Feb 27 07:53:24 crc kubenswrapper[4612]: I0227 07:53:24.346960 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-7c6f84b55d-wrdhn"] Feb 27 07:53:24 crc kubenswrapper[4612]: I0227 07:53:24.394377 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5cffc6cf87-sblh2"] Feb 27 07:53:24 crc kubenswrapper[4612]: W0227 07:53:24.408799 4612 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod99f24b2c_4969_454a_bc54_1b76e67b302d.slice/crio-dd8ab33551958eab2f60268ae0030637cbc43952149529072ccfdb6ae773c509 WatchSource:0}: Error finding container dd8ab33551958eab2f60268ae0030637cbc43952149529072ccfdb6ae773c509: Status 404 returned error can't find the container with id dd8ab33551958eab2f60268ae0030637cbc43952149529072ccfdb6ae773c509 Feb 27 07:53:24 crc kubenswrapper[4612]: I0227 07:53:24.634080 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5cffc6cf87-sblh2" event={"ID":"99f24b2c-4969-454a-bc54-1b76e67b302d","Type":"ContainerStarted","Data":"74c155de9887d2fdea347e0c1509aa6d9a60150314b14aa13e5b9229a9343e46"} Feb 27 07:53:24 crc kubenswrapper[4612]: I0227 07:53:24.634164 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-5cffc6cf87-sblh2" Feb 27 07:53:24 crc kubenswrapper[4612]: I0227 07:53:24.634180 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5cffc6cf87-sblh2" event={"ID":"99f24b2c-4969-454a-bc54-1b76e67b302d","Type":"ContainerStarted","Data":"dd8ab33551958eab2f60268ae0030637cbc43952149529072ccfdb6ae773c509"} Feb 27 07:53:24 crc kubenswrapper[4612]: I0227 07:53:24.635996 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7c6f84b55d-wrdhn" event={"ID":"5924d34a-c67c-45e1-8085-dafed86b5265","Type":"ContainerStarted","Data":"987bab8e3e3b48d6feae67fbb11aa18ec8389febb0ce0196770c717c507ff227"} Feb 27 07:53:24 crc kubenswrapper[4612]: I0227 07:53:24.636036 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7c6f84b55d-wrdhn" event={"ID":"5924d34a-c67c-45e1-8085-dafed86b5265","Type":"ContainerStarted","Data":"c2ab1622b7eef1af793d34f1b84430a95a0a1c2fdc9325a2b1817b66463b6629"} Feb 27 07:53:24 crc kubenswrapper[4612]: I0227 07:53:24.636225 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-7c6f84b55d-wrdhn" Feb 27 07:53:24 crc kubenswrapper[4612]: I0227 07:53:24.640276 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-7c6f84b55d-wrdhn" Feb 27 07:53:24 crc kubenswrapper[4612]: I0227 07:53:24.650327 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-5cffc6cf87-sblh2" podStartSLOduration=2.650314433 podStartE2EDuration="2.650314433s" podCreationTimestamp="2026-02-27 07:53:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 07:53:24.649957082 +0000 UTC m=+262.503887080" watchObservedRunningTime="2026-02-27 07:53:24.650314433 +0000 UTC m=+262.504244431" Feb 27 07:53:24 crc kubenswrapper[4612]: I0227 07:53:24.825345 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-5cffc6cf87-sblh2" Feb 27 07:53:24 crc kubenswrapper[4612]: I0227 07:53:24.848222 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-7c6f84b55d-wrdhn" podStartSLOduration=2.848187448 podStartE2EDuration="2.848187448s" podCreationTimestamp="2026-02-27 07:53:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 07:53:24.674137275 +0000 UTC m=+262.528067293" watchObservedRunningTime="2026-02-27 07:53:24.848187448 +0000 UTC m=+262.702117496" Feb 27 07:53:24 crc kubenswrapper[4612]: I0227 07:53:24.860371 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a7b8e4c7-3bf3-4ef0-8aca-96037f1eea24" path="/var/lib/kubelet/pods/a7b8e4c7-3bf3-4ef0-8aca-96037f1eea24/volumes" Feb 27 07:53:24 crc kubenswrapper[4612]: I0227 07:53:24.860935 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c243dedd-5585-4353-9303-6e42ebc82f69" path="/var/lib/kubelet/pods/c243dedd-5585-4353-9303-6e42ebc82f69/volumes" Feb 27 07:53:34 crc kubenswrapper[4612]: I0227 07:53:34.697985 4612 generic.go:334] "Generic (PLEG): container finished" podID="6d061f42-06de-43e5-b4d0-dbca09e48630" containerID="872f6d6ba2cff41e30dff82c7bfa288c746d738f5f9784ceca408e98c3730656" exitCode=0 Feb 27 07:53:34 crc kubenswrapper[4612]: I0227 07:53:34.698109 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-hp72h" event={"ID":"6d061f42-06de-43e5-b4d0-dbca09e48630","Type":"ContainerDied","Data":"872f6d6ba2cff41e30dff82c7bfa288c746d738f5f9784ceca408e98c3730656"} Feb 27 07:53:34 crc kubenswrapper[4612]: I0227 07:53:34.699195 4612 scope.go:117] "RemoveContainer" containerID="872f6d6ba2cff41e30dff82c7bfa288c746d738f5f9784ceca408e98c3730656" Feb 27 07:53:35 crc kubenswrapper[4612]: I0227 07:53:35.704232 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-hp72h" event={"ID":"6d061f42-06de-43e5-b4d0-dbca09e48630","Type":"ContainerStarted","Data":"33022ff99e1ecbfeccfd65a15ffbd81939f7629ee36cb5c0959129bde7ef1c56"} Feb 27 07:53:35 crc kubenswrapper[4612]: I0227 07:53:35.705642 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-hp72h" Feb 27 07:53:35 crc kubenswrapper[4612]: I0227 07:53:35.709948 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-hp72h" Feb 27 07:53:42 crc kubenswrapper[4612]: I0227 07:53:42.142856 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-7c6f84b55d-wrdhn"] Feb 27 07:53:42 crc kubenswrapper[4612]: I0227 07:53:42.143497 4612 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-7c6f84b55d-wrdhn" podUID="5924d34a-c67c-45e1-8085-dafed86b5265" containerName="controller-manager" containerID="cri-o://987bab8e3e3b48d6feae67fbb11aa18ec8389febb0ce0196770c717c507ff227" gracePeriod=30 Feb 27 07:53:42 crc kubenswrapper[4612]: I0227 07:53:42.171740 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5cffc6cf87-sblh2"] Feb 27 07:53:42 crc kubenswrapper[4612]: I0227 07:53:42.172003 4612 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-5cffc6cf87-sblh2" podUID="99f24b2c-4969-454a-bc54-1b76e67b302d" containerName="route-controller-manager" containerID="cri-o://74c155de9887d2fdea347e0c1509aa6d9a60150314b14aa13e5b9229a9343e46" gracePeriod=30 Feb 27 07:53:42 crc kubenswrapper[4612]: I0227 07:53:42.520575 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7c6f84b55d-wrdhn" Feb 27 07:53:42 crc kubenswrapper[4612]: I0227 07:53:42.524245 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5cffc6cf87-sblh2" Feb 27 07:53:42 crc kubenswrapper[4612]: I0227 07:53:42.690645 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nl264\" (UniqueName: \"kubernetes.io/projected/5924d34a-c67c-45e1-8085-dafed86b5265-kube-api-access-nl264\") pod \"5924d34a-c67c-45e1-8085-dafed86b5265\" (UID: \"5924d34a-c67c-45e1-8085-dafed86b5265\") " Feb 27 07:53:42 crc kubenswrapper[4612]: I0227 07:53:42.690705 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/99f24b2c-4969-454a-bc54-1b76e67b302d-client-ca\") pod \"99f24b2c-4969-454a-bc54-1b76e67b302d\" (UID: \"99f24b2c-4969-454a-bc54-1b76e67b302d\") " Feb 27 07:53:42 crc kubenswrapper[4612]: I0227 07:53:42.690811 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/99f24b2c-4969-454a-bc54-1b76e67b302d-serving-cert\") pod \"99f24b2c-4969-454a-bc54-1b76e67b302d\" (UID: \"99f24b2c-4969-454a-bc54-1b76e67b302d\") " Feb 27 07:53:42 crc kubenswrapper[4612]: I0227 07:53:42.690856 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5924d34a-c67c-45e1-8085-dafed86b5265-serving-cert\") pod \"5924d34a-c67c-45e1-8085-dafed86b5265\" (UID: \"5924d34a-c67c-45e1-8085-dafed86b5265\") " Feb 27 07:53:42 crc kubenswrapper[4612]: I0227 07:53:42.690878 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5924d34a-c67c-45e1-8085-dafed86b5265-client-ca\") pod \"5924d34a-c67c-45e1-8085-dafed86b5265\" (UID: \"5924d34a-c67c-45e1-8085-dafed86b5265\") " Feb 27 07:53:42 crc kubenswrapper[4612]: I0227 07:53:42.690908 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/99f24b2c-4969-454a-bc54-1b76e67b302d-config\") pod \"99f24b2c-4969-454a-bc54-1b76e67b302d\" (UID: \"99f24b2c-4969-454a-bc54-1b76e67b302d\") " Feb 27 07:53:42 crc kubenswrapper[4612]: I0227 07:53:42.690938 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4xrxl\" (UniqueName: \"kubernetes.io/projected/99f24b2c-4969-454a-bc54-1b76e67b302d-kube-api-access-4xrxl\") pod \"99f24b2c-4969-454a-bc54-1b76e67b302d\" (UID: \"99f24b2c-4969-454a-bc54-1b76e67b302d\") " Feb 27 07:53:42 crc kubenswrapper[4612]: I0227 07:53:42.690961 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/5924d34a-c67c-45e1-8085-dafed86b5265-proxy-ca-bundles\") pod \"5924d34a-c67c-45e1-8085-dafed86b5265\" (UID: \"5924d34a-c67c-45e1-8085-dafed86b5265\") " Feb 27 07:53:42 crc kubenswrapper[4612]: I0227 07:53:42.690986 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5924d34a-c67c-45e1-8085-dafed86b5265-config\") pod \"5924d34a-c67c-45e1-8085-dafed86b5265\" (UID: \"5924d34a-c67c-45e1-8085-dafed86b5265\") " Feb 27 07:53:42 crc kubenswrapper[4612]: I0227 07:53:42.692086 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5924d34a-c67c-45e1-8085-dafed86b5265-config" (OuterVolumeSpecName: "config") pod "5924d34a-c67c-45e1-8085-dafed86b5265" (UID: "5924d34a-c67c-45e1-8085-dafed86b5265"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 07:53:42 crc kubenswrapper[4612]: I0227 07:53:42.692185 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/99f24b2c-4969-454a-bc54-1b76e67b302d-client-ca" (OuterVolumeSpecName: "client-ca") pod "99f24b2c-4969-454a-bc54-1b76e67b302d" (UID: "99f24b2c-4969-454a-bc54-1b76e67b302d"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 07:53:42 crc kubenswrapper[4612]: I0227 07:53:42.692227 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/99f24b2c-4969-454a-bc54-1b76e67b302d-config" (OuterVolumeSpecName: "config") pod "99f24b2c-4969-454a-bc54-1b76e67b302d" (UID: "99f24b2c-4969-454a-bc54-1b76e67b302d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 07:53:42 crc kubenswrapper[4612]: I0227 07:53:42.692554 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5924d34a-c67c-45e1-8085-dafed86b5265-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "5924d34a-c67c-45e1-8085-dafed86b5265" (UID: "5924d34a-c67c-45e1-8085-dafed86b5265"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 07:53:42 crc kubenswrapper[4612]: I0227 07:53:42.693190 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5924d34a-c67c-45e1-8085-dafed86b5265-client-ca" (OuterVolumeSpecName: "client-ca") pod "5924d34a-c67c-45e1-8085-dafed86b5265" (UID: "5924d34a-c67c-45e1-8085-dafed86b5265"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 07:53:42 crc kubenswrapper[4612]: I0227 07:53:42.695751 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/99f24b2c-4969-454a-bc54-1b76e67b302d-kube-api-access-4xrxl" (OuterVolumeSpecName: "kube-api-access-4xrxl") pod "99f24b2c-4969-454a-bc54-1b76e67b302d" (UID: "99f24b2c-4969-454a-bc54-1b76e67b302d"). InnerVolumeSpecName "kube-api-access-4xrxl". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 07:53:42 crc kubenswrapper[4612]: I0227 07:53:42.695928 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5924d34a-c67c-45e1-8085-dafed86b5265-kube-api-access-nl264" (OuterVolumeSpecName: "kube-api-access-nl264") pod "5924d34a-c67c-45e1-8085-dafed86b5265" (UID: "5924d34a-c67c-45e1-8085-dafed86b5265"). InnerVolumeSpecName "kube-api-access-nl264". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 07:53:42 crc kubenswrapper[4612]: I0227 07:53:42.695959 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5924d34a-c67c-45e1-8085-dafed86b5265-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5924d34a-c67c-45e1-8085-dafed86b5265" (UID: "5924d34a-c67c-45e1-8085-dafed86b5265"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 07:53:42 crc kubenswrapper[4612]: I0227 07:53:42.697761 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/99f24b2c-4969-454a-bc54-1b76e67b302d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "99f24b2c-4969-454a-bc54-1b76e67b302d" (UID: "99f24b2c-4969-454a-bc54-1b76e67b302d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 07:53:42 crc kubenswrapper[4612]: I0227 07:53:42.756267 4612 generic.go:334] "Generic (PLEG): container finished" podID="5924d34a-c67c-45e1-8085-dafed86b5265" containerID="987bab8e3e3b48d6feae67fbb11aa18ec8389febb0ce0196770c717c507ff227" exitCode=0 Feb 27 07:53:42 crc kubenswrapper[4612]: I0227 07:53:42.756313 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7c6f84b55d-wrdhn" event={"ID":"5924d34a-c67c-45e1-8085-dafed86b5265","Type":"ContainerDied","Data":"987bab8e3e3b48d6feae67fbb11aa18ec8389febb0ce0196770c717c507ff227"} Feb 27 07:53:42 crc kubenswrapper[4612]: I0227 07:53:42.756363 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7c6f84b55d-wrdhn" event={"ID":"5924d34a-c67c-45e1-8085-dafed86b5265","Type":"ContainerDied","Data":"c2ab1622b7eef1af793d34f1b84430a95a0a1c2fdc9325a2b1817b66463b6629"} Feb 27 07:53:42 crc kubenswrapper[4612]: I0227 07:53:42.756381 4612 scope.go:117] "RemoveContainer" containerID="987bab8e3e3b48d6feae67fbb11aa18ec8389febb0ce0196770c717c507ff227" Feb 27 07:53:42 crc kubenswrapper[4612]: I0227 07:53:42.756671 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7c6f84b55d-wrdhn" Feb 27 07:53:42 crc kubenswrapper[4612]: I0227 07:53:42.757588 4612 generic.go:334] "Generic (PLEG): container finished" podID="99f24b2c-4969-454a-bc54-1b76e67b302d" containerID="74c155de9887d2fdea347e0c1509aa6d9a60150314b14aa13e5b9229a9343e46" exitCode=0 Feb 27 07:53:42 crc kubenswrapper[4612]: I0227 07:53:42.757617 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5cffc6cf87-sblh2" event={"ID":"99f24b2c-4969-454a-bc54-1b76e67b302d","Type":"ContainerDied","Data":"74c155de9887d2fdea347e0c1509aa6d9a60150314b14aa13e5b9229a9343e46"} Feb 27 07:53:42 crc kubenswrapper[4612]: I0227 07:53:42.757632 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5cffc6cf87-sblh2" event={"ID":"99f24b2c-4969-454a-bc54-1b76e67b302d","Type":"ContainerDied","Data":"dd8ab33551958eab2f60268ae0030637cbc43952149529072ccfdb6ae773c509"} Feb 27 07:53:42 crc kubenswrapper[4612]: I0227 07:53:42.757660 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5cffc6cf87-sblh2" Feb 27 07:53:42 crc kubenswrapper[4612]: I0227 07:53:42.773639 4612 scope.go:117] "RemoveContainer" containerID="987bab8e3e3b48d6feae67fbb11aa18ec8389febb0ce0196770c717c507ff227" Feb 27 07:53:42 crc kubenswrapper[4612]: E0227 07:53:42.774035 4612 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"987bab8e3e3b48d6feae67fbb11aa18ec8389febb0ce0196770c717c507ff227\": container with ID starting with 987bab8e3e3b48d6feae67fbb11aa18ec8389febb0ce0196770c717c507ff227 not found: ID does not exist" containerID="987bab8e3e3b48d6feae67fbb11aa18ec8389febb0ce0196770c717c507ff227" Feb 27 07:53:42 crc kubenswrapper[4612]: I0227 07:53:42.774066 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"987bab8e3e3b48d6feae67fbb11aa18ec8389febb0ce0196770c717c507ff227"} err="failed to get container status \"987bab8e3e3b48d6feae67fbb11aa18ec8389febb0ce0196770c717c507ff227\": rpc error: code = NotFound desc = could not find container \"987bab8e3e3b48d6feae67fbb11aa18ec8389febb0ce0196770c717c507ff227\": container with ID starting with 987bab8e3e3b48d6feae67fbb11aa18ec8389febb0ce0196770c717c507ff227 not found: ID does not exist" Feb 27 07:53:42 crc kubenswrapper[4612]: I0227 07:53:42.774083 4612 scope.go:117] "RemoveContainer" containerID="74c155de9887d2fdea347e0c1509aa6d9a60150314b14aa13e5b9229a9343e46" Feb 27 07:53:42 crc kubenswrapper[4612]: I0227 07:53:42.791330 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5cffc6cf87-sblh2"] Feb 27 07:53:42 crc kubenswrapper[4612]: I0227 07:53:42.791786 4612 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/99f24b2c-4969-454a-bc54-1b76e67b302d-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 27 07:53:42 crc kubenswrapper[4612]: I0227 07:53:42.791810 4612 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5924d34a-c67c-45e1-8085-dafed86b5265-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 27 07:53:42 crc kubenswrapper[4612]: I0227 07:53:42.791820 4612 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5924d34a-c67c-45e1-8085-dafed86b5265-client-ca\") on node \"crc\" DevicePath \"\"" Feb 27 07:53:42 crc kubenswrapper[4612]: I0227 07:53:42.791831 4612 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/99f24b2c-4969-454a-bc54-1b76e67b302d-config\") on node \"crc\" DevicePath \"\"" Feb 27 07:53:42 crc kubenswrapper[4612]: I0227 07:53:42.791840 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4xrxl\" (UniqueName: \"kubernetes.io/projected/99f24b2c-4969-454a-bc54-1b76e67b302d-kube-api-access-4xrxl\") on node \"crc\" DevicePath \"\"" Feb 27 07:53:42 crc kubenswrapper[4612]: I0227 07:53:42.791850 4612 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/5924d34a-c67c-45e1-8085-dafed86b5265-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Feb 27 07:53:42 crc kubenswrapper[4612]: I0227 07:53:42.791858 4612 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5924d34a-c67c-45e1-8085-dafed86b5265-config\") on node \"crc\" DevicePath \"\"" Feb 27 07:53:42 crc kubenswrapper[4612]: I0227 07:53:42.791866 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nl264\" (UniqueName: \"kubernetes.io/projected/5924d34a-c67c-45e1-8085-dafed86b5265-kube-api-access-nl264\") on node \"crc\" DevicePath \"\"" Feb 27 07:53:42 crc kubenswrapper[4612]: I0227 07:53:42.791875 4612 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/99f24b2c-4969-454a-bc54-1b76e67b302d-client-ca\") on node \"crc\" DevicePath \"\"" Feb 27 07:53:42 crc kubenswrapper[4612]: I0227 07:53:42.792999 4612 scope.go:117] "RemoveContainer" containerID="74c155de9887d2fdea347e0c1509aa6d9a60150314b14aa13e5b9229a9343e46" Feb 27 07:53:42 crc kubenswrapper[4612]: E0227 07:53:42.793443 4612 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"74c155de9887d2fdea347e0c1509aa6d9a60150314b14aa13e5b9229a9343e46\": container with ID starting with 74c155de9887d2fdea347e0c1509aa6d9a60150314b14aa13e5b9229a9343e46 not found: ID does not exist" containerID="74c155de9887d2fdea347e0c1509aa6d9a60150314b14aa13e5b9229a9343e46" Feb 27 07:53:42 crc kubenswrapper[4612]: I0227 07:53:42.793485 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"74c155de9887d2fdea347e0c1509aa6d9a60150314b14aa13e5b9229a9343e46"} err="failed to get container status \"74c155de9887d2fdea347e0c1509aa6d9a60150314b14aa13e5b9229a9343e46\": rpc error: code = NotFound desc = could not find container \"74c155de9887d2fdea347e0c1509aa6d9a60150314b14aa13e5b9229a9343e46\": container with ID starting with 74c155de9887d2fdea347e0c1509aa6d9a60150314b14aa13e5b9229a9343e46 not found: ID does not exist" Feb 27 07:53:42 crc kubenswrapper[4612]: I0227 07:53:42.794873 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5cffc6cf87-sblh2"] Feb 27 07:53:42 crc kubenswrapper[4612]: I0227 07:53:42.811577 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-7c6f84b55d-wrdhn"] Feb 27 07:53:42 crc kubenswrapper[4612]: I0227 07:53:42.817861 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-7c6f84b55d-wrdhn"] Feb 27 07:53:42 crc kubenswrapper[4612]: I0227 07:53:42.858651 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5924d34a-c67c-45e1-8085-dafed86b5265" path="/var/lib/kubelet/pods/5924d34a-c67c-45e1-8085-dafed86b5265/volumes" Feb 27 07:53:42 crc kubenswrapper[4612]: I0227 07:53:42.859166 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="99f24b2c-4969-454a-bc54-1b76e67b302d" path="/var/lib/kubelet/pods/99f24b2c-4969-454a-bc54-1b76e67b302d/volumes" Feb 27 07:53:43 crc kubenswrapper[4612]: I0227 07:53:43.816513 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-55bb989677-vtr6v"] Feb 27 07:53:43 crc kubenswrapper[4612]: E0227 07:53:43.816908 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="99f24b2c-4969-454a-bc54-1b76e67b302d" containerName="route-controller-manager" Feb 27 07:53:43 crc kubenswrapper[4612]: I0227 07:53:43.816936 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="99f24b2c-4969-454a-bc54-1b76e67b302d" containerName="route-controller-manager" Feb 27 07:53:43 crc kubenswrapper[4612]: E0227 07:53:43.816955 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5924d34a-c67c-45e1-8085-dafed86b5265" containerName="controller-manager" Feb 27 07:53:43 crc kubenswrapper[4612]: I0227 07:53:43.816966 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="5924d34a-c67c-45e1-8085-dafed86b5265" containerName="controller-manager" Feb 27 07:53:43 crc kubenswrapper[4612]: I0227 07:53:43.817092 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="5924d34a-c67c-45e1-8085-dafed86b5265" containerName="controller-manager" Feb 27 07:53:43 crc kubenswrapper[4612]: I0227 07:53:43.817111 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="99f24b2c-4969-454a-bc54-1b76e67b302d" containerName="route-controller-manager" Feb 27 07:53:43 crc kubenswrapper[4612]: I0227 07:53:43.817589 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-55bb989677-vtr6v" Feb 27 07:53:43 crc kubenswrapper[4612]: I0227 07:53:43.820437 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Feb 27 07:53:43 crc kubenswrapper[4612]: I0227 07:53:43.820489 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Feb 27 07:53:43 crc kubenswrapper[4612]: I0227 07:53:43.820753 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Feb 27 07:53:43 crc kubenswrapper[4612]: I0227 07:53:43.821889 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Feb 27 07:53:43 crc kubenswrapper[4612]: I0227 07:53:43.822586 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Feb 27 07:53:43 crc kubenswrapper[4612]: I0227 07:53:43.823681 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Feb 27 07:53:43 crc kubenswrapper[4612]: I0227 07:53:43.825436 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-7c589966c4-2pn6j"] Feb 27 07:53:43 crc kubenswrapper[4612]: I0227 07:53:43.826256 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7c589966c4-2pn6j" Feb 27 07:53:43 crc kubenswrapper[4612]: I0227 07:53:43.828118 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Feb 27 07:53:43 crc kubenswrapper[4612]: I0227 07:53:43.828129 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Feb 27 07:53:43 crc kubenswrapper[4612]: I0227 07:53:43.828730 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Feb 27 07:53:43 crc kubenswrapper[4612]: I0227 07:53:43.828802 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Feb 27 07:53:43 crc kubenswrapper[4612]: I0227 07:53:43.833872 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Feb 27 07:53:43 crc kubenswrapper[4612]: I0227 07:53:43.833983 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Feb 27 07:53:43 crc kubenswrapper[4612]: I0227 07:53:43.841808 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Feb 27 07:53:43 crc kubenswrapper[4612]: I0227 07:53:43.844003 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-7c589966c4-2pn6j"] Feb 27 07:53:43 crc kubenswrapper[4612]: I0227 07:53:43.863552 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-55bb989677-vtr6v"] Feb 27 07:53:43 crc kubenswrapper[4612]: I0227 07:53:43.906234 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9stsn\" (UniqueName: \"kubernetes.io/projected/675d6b43-fe8b-41b1-b52b-3f5389fffae1-kube-api-access-9stsn\") pod \"route-controller-manager-55bb989677-vtr6v\" (UID: \"675d6b43-fe8b-41b1-b52b-3f5389fffae1\") " pod="openshift-route-controller-manager/route-controller-manager-55bb989677-vtr6v" Feb 27 07:53:43 crc kubenswrapper[4612]: I0227 07:53:43.906322 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/675d6b43-fe8b-41b1-b52b-3f5389fffae1-serving-cert\") pod \"route-controller-manager-55bb989677-vtr6v\" (UID: \"675d6b43-fe8b-41b1-b52b-3f5389fffae1\") " pod="openshift-route-controller-manager/route-controller-manager-55bb989677-vtr6v" Feb 27 07:53:43 crc kubenswrapper[4612]: I0227 07:53:43.906381 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/3301aca5-f994-4520-b8bb-02fef2be45a2-client-ca\") pod \"controller-manager-7c589966c4-2pn6j\" (UID: \"3301aca5-f994-4520-b8bb-02fef2be45a2\") " pod="openshift-controller-manager/controller-manager-7c589966c4-2pn6j" Feb 27 07:53:43 crc kubenswrapper[4612]: I0227 07:53:43.906416 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/3301aca5-f994-4520-b8bb-02fef2be45a2-proxy-ca-bundles\") pod \"controller-manager-7c589966c4-2pn6j\" (UID: \"3301aca5-f994-4520-b8bb-02fef2be45a2\") " pod="openshift-controller-manager/controller-manager-7c589966c4-2pn6j" Feb 27 07:53:43 crc kubenswrapper[4612]: I0227 07:53:43.906449 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3301aca5-f994-4520-b8bb-02fef2be45a2-serving-cert\") pod \"controller-manager-7c589966c4-2pn6j\" (UID: \"3301aca5-f994-4520-b8bb-02fef2be45a2\") " pod="openshift-controller-manager/controller-manager-7c589966c4-2pn6j" Feb 27 07:53:43 crc kubenswrapper[4612]: I0227 07:53:43.906532 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3301aca5-f994-4520-b8bb-02fef2be45a2-config\") pod \"controller-manager-7c589966c4-2pn6j\" (UID: \"3301aca5-f994-4520-b8bb-02fef2be45a2\") " pod="openshift-controller-manager/controller-manager-7c589966c4-2pn6j" Feb 27 07:53:43 crc kubenswrapper[4612]: I0227 07:53:43.906559 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/675d6b43-fe8b-41b1-b52b-3f5389fffae1-client-ca\") pod \"route-controller-manager-55bb989677-vtr6v\" (UID: \"675d6b43-fe8b-41b1-b52b-3f5389fffae1\") " pod="openshift-route-controller-manager/route-controller-manager-55bb989677-vtr6v" Feb 27 07:53:43 crc kubenswrapper[4612]: I0227 07:53:43.906585 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qpfzh\" (UniqueName: \"kubernetes.io/projected/3301aca5-f994-4520-b8bb-02fef2be45a2-kube-api-access-qpfzh\") pod \"controller-manager-7c589966c4-2pn6j\" (UID: \"3301aca5-f994-4520-b8bb-02fef2be45a2\") " pod="openshift-controller-manager/controller-manager-7c589966c4-2pn6j" Feb 27 07:53:43 crc kubenswrapper[4612]: I0227 07:53:43.906607 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/675d6b43-fe8b-41b1-b52b-3f5389fffae1-config\") pod \"route-controller-manager-55bb989677-vtr6v\" (UID: \"675d6b43-fe8b-41b1-b52b-3f5389fffae1\") " pod="openshift-route-controller-manager/route-controller-manager-55bb989677-vtr6v" Feb 27 07:53:44 crc kubenswrapper[4612]: I0227 07:53:44.007387 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9stsn\" (UniqueName: \"kubernetes.io/projected/675d6b43-fe8b-41b1-b52b-3f5389fffae1-kube-api-access-9stsn\") pod \"route-controller-manager-55bb989677-vtr6v\" (UID: \"675d6b43-fe8b-41b1-b52b-3f5389fffae1\") " pod="openshift-route-controller-manager/route-controller-manager-55bb989677-vtr6v" Feb 27 07:53:44 crc kubenswrapper[4612]: I0227 07:53:44.007444 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/675d6b43-fe8b-41b1-b52b-3f5389fffae1-serving-cert\") pod \"route-controller-manager-55bb989677-vtr6v\" (UID: \"675d6b43-fe8b-41b1-b52b-3f5389fffae1\") " pod="openshift-route-controller-manager/route-controller-manager-55bb989677-vtr6v" Feb 27 07:53:44 crc kubenswrapper[4612]: I0227 07:53:44.007474 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/3301aca5-f994-4520-b8bb-02fef2be45a2-client-ca\") pod \"controller-manager-7c589966c4-2pn6j\" (UID: \"3301aca5-f994-4520-b8bb-02fef2be45a2\") " pod="openshift-controller-manager/controller-manager-7c589966c4-2pn6j" Feb 27 07:53:44 crc kubenswrapper[4612]: I0227 07:53:44.007497 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/3301aca5-f994-4520-b8bb-02fef2be45a2-proxy-ca-bundles\") pod \"controller-manager-7c589966c4-2pn6j\" (UID: \"3301aca5-f994-4520-b8bb-02fef2be45a2\") " pod="openshift-controller-manager/controller-manager-7c589966c4-2pn6j" Feb 27 07:53:44 crc kubenswrapper[4612]: I0227 07:53:44.007517 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3301aca5-f994-4520-b8bb-02fef2be45a2-serving-cert\") pod \"controller-manager-7c589966c4-2pn6j\" (UID: \"3301aca5-f994-4520-b8bb-02fef2be45a2\") " pod="openshift-controller-manager/controller-manager-7c589966c4-2pn6j" Feb 27 07:53:44 crc kubenswrapper[4612]: I0227 07:53:44.007556 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3301aca5-f994-4520-b8bb-02fef2be45a2-config\") pod \"controller-manager-7c589966c4-2pn6j\" (UID: \"3301aca5-f994-4520-b8bb-02fef2be45a2\") " pod="openshift-controller-manager/controller-manager-7c589966c4-2pn6j" Feb 27 07:53:44 crc kubenswrapper[4612]: I0227 07:53:44.007574 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/675d6b43-fe8b-41b1-b52b-3f5389fffae1-client-ca\") pod \"route-controller-manager-55bb989677-vtr6v\" (UID: \"675d6b43-fe8b-41b1-b52b-3f5389fffae1\") " pod="openshift-route-controller-manager/route-controller-manager-55bb989677-vtr6v" Feb 27 07:53:44 crc kubenswrapper[4612]: I0227 07:53:44.007593 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qpfzh\" (UniqueName: \"kubernetes.io/projected/3301aca5-f994-4520-b8bb-02fef2be45a2-kube-api-access-qpfzh\") pod \"controller-manager-7c589966c4-2pn6j\" (UID: \"3301aca5-f994-4520-b8bb-02fef2be45a2\") " pod="openshift-controller-manager/controller-manager-7c589966c4-2pn6j" Feb 27 07:53:44 crc kubenswrapper[4612]: I0227 07:53:44.007608 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/675d6b43-fe8b-41b1-b52b-3f5389fffae1-config\") pod \"route-controller-manager-55bb989677-vtr6v\" (UID: \"675d6b43-fe8b-41b1-b52b-3f5389fffae1\") " pod="openshift-route-controller-manager/route-controller-manager-55bb989677-vtr6v" Feb 27 07:53:44 crc kubenswrapper[4612]: I0227 07:53:44.008949 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/3301aca5-f994-4520-b8bb-02fef2be45a2-client-ca\") pod \"controller-manager-7c589966c4-2pn6j\" (UID: \"3301aca5-f994-4520-b8bb-02fef2be45a2\") " pod="openshift-controller-manager/controller-manager-7c589966c4-2pn6j" Feb 27 07:53:44 crc kubenswrapper[4612]: I0227 07:53:44.009229 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/3301aca5-f994-4520-b8bb-02fef2be45a2-proxy-ca-bundles\") pod \"controller-manager-7c589966c4-2pn6j\" (UID: \"3301aca5-f994-4520-b8bb-02fef2be45a2\") " pod="openshift-controller-manager/controller-manager-7c589966c4-2pn6j" Feb 27 07:53:44 crc kubenswrapper[4612]: I0227 07:53:44.009665 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3301aca5-f994-4520-b8bb-02fef2be45a2-config\") pod \"controller-manager-7c589966c4-2pn6j\" (UID: \"3301aca5-f994-4520-b8bb-02fef2be45a2\") " pod="openshift-controller-manager/controller-manager-7c589966c4-2pn6j" Feb 27 07:53:44 crc kubenswrapper[4612]: I0227 07:53:44.009896 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/675d6b43-fe8b-41b1-b52b-3f5389fffae1-client-ca\") pod \"route-controller-manager-55bb989677-vtr6v\" (UID: \"675d6b43-fe8b-41b1-b52b-3f5389fffae1\") " pod="openshift-route-controller-manager/route-controller-manager-55bb989677-vtr6v" Feb 27 07:53:44 crc kubenswrapper[4612]: I0227 07:53:44.010194 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/675d6b43-fe8b-41b1-b52b-3f5389fffae1-config\") pod \"route-controller-manager-55bb989677-vtr6v\" (UID: \"675d6b43-fe8b-41b1-b52b-3f5389fffae1\") " pod="openshift-route-controller-manager/route-controller-manager-55bb989677-vtr6v" Feb 27 07:53:44 crc kubenswrapper[4612]: I0227 07:53:44.013613 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3301aca5-f994-4520-b8bb-02fef2be45a2-serving-cert\") pod \"controller-manager-7c589966c4-2pn6j\" (UID: \"3301aca5-f994-4520-b8bb-02fef2be45a2\") " pod="openshift-controller-manager/controller-manager-7c589966c4-2pn6j" Feb 27 07:53:44 crc kubenswrapper[4612]: I0227 07:53:44.018846 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/675d6b43-fe8b-41b1-b52b-3f5389fffae1-serving-cert\") pod \"route-controller-manager-55bb989677-vtr6v\" (UID: \"675d6b43-fe8b-41b1-b52b-3f5389fffae1\") " pod="openshift-route-controller-manager/route-controller-manager-55bb989677-vtr6v" Feb 27 07:53:44 crc kubenswrapper[4612]: I0227 07:53:44.025135 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9stsn\" (UniqueName: \"kubernetes.io/projected/675d6b43-fe8b-41b1-b52b-3f5389fffae1-kube-api-access-9stsn\") pod \"route-controller-manager-55bb989677-vtr6v\" (UID: \"675d6b43-fe8b-41b1-b52b-3f5389fffae1\") " pod="openshift-route-controller-manager/route-controller-manager-55bb989677-vtr6v" Feb 27 07:53:44 crc kubenswrapper[4612]: I0227 07:53:44.025981 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qpfzh\" (UniqueName: \"kubernetes.io/projected/3301aca5-f994-4520-b8bb-02fef2be45a2-kube-api-access-qpfzh\") pod \"controller-manager-7c589966c4-2pn6j\" (UID: \"3301aca5-f994-4520-b8bb-02fef2be45a2\") " pod="openshift-controller-manager/controller-manager-7c589966c4-2pn6j" Feb 27 07:53:44 crc kubenswrapper[4612]: I0227 07:53:44.138109 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-55bb989677-vtr6v" Feb 27 07:53:44 crc kubenswrapper[4612]: I0227 07:53:44.144190 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7c589966c4-2pn6j" Feb 27 07:53:44 crc kubenswrapper[4612]: I0227 07:53:44.382082 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-7c589966c4-2pn6j"] Feb 27 07:53:44 crc kubenswrapper[4612]: I0227 07:53:44.653383 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-55bb989677-vtr6v"] Feb 27 07:53:44 crc kubenswrapper[4612]: W0227 07:53:44.660414 4612 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod675d6b43_fe8b_41b1_b52b_3f5389fffae1.slice/crio-ba9aeee41cbd096c29a1a627ab74765aeae706ea0a8d501030552509964be7c9 WatchSource:0}: Error finding container ba9aeee41cbd096c29a1a627ab74765aeae706ea0a8d501030552509964be7c9: Status 404 returned error can't find the container with id ba9aeee41cbd096c29a1a627ab74765aeae706ea0a8d501030552509964be7c9 Feb 27 07:53:44 crc kubenswrapper[4612]: I0227 07:53:44.772982 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7c589966c4-2pn6j" event={"ID":"3301aca5-f994-4520-b8bb-02fef2be45a2","Type":"ContainerStarted","Data":"165e292b2620c937ad8cbcf15871bc28f6a4528d8076a9c80f677ae6a58ac0fe"} Feb 27 07:53:44 crc kubenswrapper[4612]: I0227 07:53:44.773302 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7c589966c4-2pn6j" event={"ID":"3301aca5-f994-4520-b8bb-02fef2be45a2","Type":"ContainerStarted","Data":"c73a87882715e7750bf07df39d4484fc62a3a3717ad1f7aec349ce2e62c57028"} Feb 27 07:53:44 crc kubenswrapper[4612]: I0227 07:53:44.773321 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-7c589966c4-2pn6j" Feb 27 07:53:44 crc kubenswrapper[4612]: I0227 07:53:44.774567 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-55bb989677-vtr6v" event={"ID":"675d6b43-fe8b-41b1-b52b-3f5389fffae1","Type":"ContainerStarted","Data":"ba9aeee41cbd096c29a1a627ab74765aeae706ea0a8d501030552509964be7c9"} Feb 27 07:53:44 crc kubenswrapper[4612]: I0227 07:53:44.777519 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-7c589966c4-2pn6j" Feb 27 07:53:44 crc kubenswrapper[4612]: I0227 07:53:44.791845 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-7c589966c4-2pn6j" podStartSLOduration=2.791823735 podStartE2EDuration="2.791823735s" podCreationTimestamp="2026-02-27 07:53:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 07:53:44.789268417 +0000 UTC m=+282.643198425" watchObservedRunningTime="2026-02-27 07:53:44.791823735 +0000 UTC m=+282.645753733" Feb 27 07:53:45 crc kubenswrapper[4612]: I0227 07:53:45.780227 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-55bb989677-vtr6v" event={"ID":"675d6b43-fe8b-41b1-b52b-3f5389fffae1","Type":"ContainerStarted","Data":"da4b53138f278105fda17730ca664d0a4bafcc5097bc37964b38c175c6943e34"} Feb 27 07:53:45 crc kubenswrapper[4612]: I0227 07:53:45.801388 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-55bb989677-vtr6v" podStartSLOduration=3.801371854 podStartE2EDuration="3.801371854s" podCreationTimestamp="2026-02-27 07:53:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 07:53:45.795093813 +0000 UTC m=+283.649023811" watchObservedRunningTime="2026-02-27 07:53:45.801371854 +0000 UTC m=+283.655301852" Feb 27 07:53:46 crc kubenswrapper[4612]: I0227 07:53:46.026795 4612 patch_prober.go:28] interesting pod/machine-config-daemon-924vb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 27 07:53:46 crc kubenswrapper[4612]: I0227 07:53:46.027089 4612 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 27 07:53:46 crc kubenswrapper[4612]: I0227 07:53:46.027198 4612 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-924vb" Feb 27 07:53:46 crc kubenswrapper[4612]: I0227 07:53:46.027809 4612 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"02aa743ba66d91783716a3af076f70d769d77dce3c83913aa953859928a3c90f"} pod="openshift-machine-config-operator/machine-config-daemon-924vb" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 27 07:53:46 crc kubenswrapper[4612]: I0227 07:53:46.027939 4612 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" containerName="machine-config-daemon" containerID="cri-o://02aa743ba66d91783716a3af076f70d769d77dce3c83913aa953859928a3c90f" gracePeriod=600 Feb 27 07:53:46 crc kubenswrapper[4612]: I0227 07:53:46.792481 4612 generic.go:334] "Generic (PLEG): container finished" podID="24599373-7adc-4a1b-8bb4-797bf726f43d" containerID="02aa743ba66d91783716a3af076f70d769d77dce3c83913aa953859928a3c90f" exitCode=0 Feb 27 07:53:46 crc kubenswrapper[4612]: I0227 07:53:46.792587 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-924vb" event={"ID":"24599373-7adc-4a1b-8bb4-797bf726f43d","Type":"ContainerDied","Data":"02aa743ba66d91783716a3af076f70d769d77dce3c83913aa953859928a3c90f"} Feb 27 07:53:46 crc kubenswrapper[4612]: I0227 07:53:46.793922 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-924vb" event={"ID":"24599373-7adc-4a1b-8bb4-797bf726f43d","Type":"ContainerStarted","Data":"50feff0d2db2ed217d5a022d814ec77b6e80057f575cba0e92f90bcac1d31a76"} Feb 27 07:53:46 crc kubenswrapper[4612]: I0227 07:53:46.794441 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-55bb989677-vtr6v" Feb 27 07:53:46 crc kubenswrapper[4612]: I0227 07:53:46.804294 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-55bb989677-vtr6v" Feb 27 07:54:00 crc kubenswrapper[4612]: I0227 07:54:00.171863 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29536314-5bct5"] Feb 27 07:54:00 crc kubenswrapper[4612]: I0227 07:54:00.172948 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536314-5bct5" Feb 27 07:54:00 crc kubenswrapper[4612]: I0227 07:54:00.181173 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-5zwfx" Feb 27 07:54:00 crc kubenswrapper[4612]: I0227 07:54:00.181188 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 27 07:54:00 crc kubenswrapper[4612]: I0227 07:54:00.181592 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 27 07:54:00 crc kubenswrapper[4612]: I0227 07:54:00.186477 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536314-5bct5"] Feb 27 07:54:00 crc kubenswrapper[4612]: I0227 07:54:00.216656 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-858xs\" (UniqueName: \"kubernetes.io/projected/01363531-9d69-4f11-90f2-d4ee8a952b1a-kube-api-access-858xs\") pod \"auto-csr-approver-29536314-5bct5\" (UID: \"01363531-9d69-4f11-90f2-d4ee8a952b1a\") " pod="openshift-infra/auto-csr-approver-29536314-5bct5" Feb 27 07:54:00 crc kubenswrapper[4612]: I0227 07:54:00.317840 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-858xs\" (UniqueName: \"kubernetes.io/projected/01363531-9d69-4f11-90f2-d4ee8a952b1a-kube-api-access-858xs\") pod \"auto-csr-approver-29536314-5bct5\" (UID: \"01363531-9d69-4f11-90f2-d4ee8a952b1a\") " pod="openshift-infra/auto-csr-approver-29536314-5bct5" Feb 27 07:54:00 crc kubenswrapper[4612]: I0227 07:54:00.338423 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-858xs\" (UniqueName: \"kubernetes.io/projected/01363531-9d69-4f11-90f2-d4ee8a952b1a-kube-api-access-858xs\") pod \"auto-csr-approver-29536314-5bct5\" (UID: \"01363531-9d69-4f11-90f2-d4ee8a952b1a\") " pod="openshift-infra/auto-csr-approver-29536314-5bct5" Feb 27 07:54:00 crc kubenswrapper[4612]: I0227 07:54:00.495109 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536314-5bct5" Feb 27 07:54:00 crc kubenswrapper[4612]: I0227 07:54:00.512245 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-7s8n7"] Feb 27 07:54:00 crc kubenswrapper[4612]: I0227 07:54:00.513090 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-7s8n7" Feb 27 07:54:00 crc kubenswrapper[4612]: I0227 07:54:00.536655 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-7s8n7"] Feb 27 07:54:00 crc kubenswrapper[4612]: I0227 07:54:00.624434 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rv5nr\" (UniqueName: \"kubernetes.io/projected/4f93a18d-29a1-4225-b99a-c801b91ce8d0-kube-api-access-rv5nr\") pod \"image-registry-66df7c8f76-7s8n7\" (UID: \"4f93a18d-29a1-4225-b99a-c801b91ce8d0\") " pod="openshift-image-registry/image-registry-66df7c8f76-7s8n7" Feb 27 07:54:00 crc kubenswrapper[4612]: I0227 07:54:00.624490 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/4f93a18d-29a1-4225-b99a-c801b91ce8d0-installation-pull-secrets\") pod \"image-registry-66df7c8f76-7s8n7\" (UID: \"4f93a18d-29a1-4225-b99a-c801b91ce8d0\") " pod="openshift-image-registry/image-registry-66df7c8f76-7s8n7" Feb 27 07:54:00 crc kubenswrapper[4612]: I0227 07:54:00.624554 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/4f93a18d-29a1-4225-b99a-c801b91ce8d0-registry-tls\") pod \"image-registry-66df7c8f76-7s8n7\" (UID: \"4f93a18d-29a1-4225-b99a-c801b91ce8d0\") " pod="openshift-image-registry/image-registry-66df7c8f76-7s8n7" Feb 27 07:54:00 crc kubenswrapper[4612]: I0227 07:54:00.624586 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/4f93a18d-29a1-4225-b99a-c801b91ce8d0-ca-trust-extracted\") pod \"image-registry-66df7c8f76-7s8n7\" (UID: \"4f93a18d-29a1-4225-b99a-c801b91ce8d0\") " pod="openshift-image-registry/image-registry-66df7c8f76-7s8n7" Feb 27 07:54:00 crc kubenswrapper[4612]: I0227 07:54:00.624620 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-7s8n7\" (UID: \"4f93a18d-29a1-4225-b99a-c801b91ce8d0\") " pod="openshift-image-registry/image-registry-66df7c8f76-7s8n7" Feb 27 07:54:00 crc kubenswrapper[4612]: I0227 07:54:00.624648 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4f93a18d-29a1-4225-b99a-c801b91ce8d0-trusted-ca\") pod \"image-registry-66df7c8f76-7s8n7\" (UID: \"4f93a18d-29a1-4225-b99a-c801b91ce8d0\") " pod="openshift-image-registry/image-registry-66df7c8f76-7s8n7" Feb 27 07:54:00 crc kubenswrapper[4612]: I0227 07:54:00.624676 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/4f93a18d-29a1-4225-b99a-c801b91ce8d0-registry-certificates\") pod \"image-registry-66df7c8f76-7s8n7\" (UID: \"4f93a18d-29a1-4225-b99a-c801b91ce8d0\") " pod="openshift-image-registry/image-registry-66df7c8f76-7s8n7" Feb 27 07:54:00 crc kubenswrapper[4612]: I0227 07:54:00.624717 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/4f93a18d-29a1-4225-b99a-c801b91ce8d0-bound-sa-token\") pod \"image-registry-66df7c8f76-7s8n7\" (UID: \"4f93a18d-29a1-4225-b99a-c801b91ce8d0\") " pod="openshift-image-registry/image-registry-66df7c8f76-7s8n7" Feb 27 07:54:00 crc kubenswrapper[4612]: I0227 07:54:00.669065 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-7s8n7\" (UID: \"4f93a18d-29a1-4225-b99a-c801b91ce8d0\") " pod="openshift-image-registry/image-registry-66df7c8f76-7s8n7" Feb 27 07:54:00 crc kubenswrapper[4612]: I0227 07:54:00.726267 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/4f93a18d-29a1-4225-b99a-c801b91ce8d0-registry-tls\") pod \"image-registry-66df7c8f76-7s8n7\" (UID: \"4f93a18d-29a1-4225-b99a-c801b91ce8d0\") " pod="openshift-image-registry/image-registry-66df7c8f76-7s8n7" Feb 27 07:54:00 crc kubenswrapper[4612]: I0227 07:54:00.727148 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/4f93a18d-29a1-4225-b99a-c801b91ce8d0-ca-trust-extracted\") pod \"image-registry-66df7c8f76-7s8n7\" (UID: \"4f93a18d-29a1-4225-b99a-c801b91ce8d0\") " pod="openshift-image-registry/image-registry-66df7c8f76-7s8n7" Feb 27 07:54:00 crc kubenswrapper[4612]: I0227 07:54:00.727250 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4f93a18d-29a1-4225-b99a-c801b91ce8d0-trusted-ca\") pod \"image-registry-66df7c8f76-7s8n7\" (UID: \"4f93a18d-29a1-4225-b99a-c801b91ce8d0\") " pod="openshift-image-registry/image-registry-66df7c8f76-7s8n7" Feb 27 07:54:00 crc kubenswrapper[4612]: I0227 07:54:00.727338 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/4f93a18d-29a1-4225-b99a-c801b91ce8d0-registry-certificates\") pod \"image-registry-66df7c8f76-7s8n7\" (UID: \"4f93a18d-29a1-4225-b99a-c801b91ce8d0\") " pod="openshift-image-registry/image-registry-66df7c8f76-7s8n7" Feb 27 07:54:00 crc kubenswrapper[4612]: I0227 07:54:00.727414 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/4f93a18d-29a1-4225-b99a-c801b91ce8d0-bound-sa-token\") pod \"image-registry-66df7c8f76-7s8n7\" (UID: \"4f93a18d-29a1-4225-b99a-c801b91ce8d0\") " pod="openshift-image-registry/image-registry-66df7c8f76-7s8n7" Feb 27 07:54:00 crc kubenswrapper[4612]: I0227 07:54:00.727505 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rv5nr\" (UniqueName: \"kubernetes.io/projected/4f93a18d-29a1-4225-b99a-c801b91ce8d0-kube-api-access-rv5nr\") pod \"image-registry-66df7c8f76-7s8n7\" (UID: \"4f93a18d-29a1-4225-b99a-c801b91ce8d0\") " pod="openshift-image-registry/image-registry-66df7c8f76-7s8n7" Feb 27 07:54:00 crc kubenswrapper[4612]: I0227 07:54:00.727577 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/4f93a18d-29a1-4225-b99a-c801b91ce8d0-installation-pull-secrets\") pod \"image-registry-66df7c8f76-7s8n7\" (UID: \"4f93a18d-29a1-4225-b99a-c801b91ce8d0\") " pod="openshift-image-registry/image-registry-66df7c8f76-7s8n7" Feb 27 07:54:00 crc kubenswrapper[4612]: I0227 07:54:00.728741 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4f93a18d-29a1-4225-b99a-c801b91ce8d0-trusted-ca\") pod \"image-registry-66df7c8f76-7s8n7\" (UID: \"4f93a18d-29a1-4225-b99a-c801b91ce8d0\") " pod="openshift-image-registry/image-registry-66df7c8f76-7s8n7" Feb 27 07:54:00 crc kubenswrapper[4612]: I0227 07:54:00.728918 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/4f93a18d-29a1-4225-b99a-c801b91ce8d0-registry-certificates\") pod \"image-registry-66df7c8f76-7s8n7\" (UID: \"4f93a18d-29a1-4225-b99a-c801b91ce8d0\") " pod="openshift-image-registry/image-registry-66df7c8f76-7s8n7" Feb 27 07:54:00 crc kubenswrapper[4612]: I0227 07:54:00.729092 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/4f93a18d-29a1-4225-b99a-c801b91ce8d0-ca-trust-extracted\") pod \"image-registry-66df7c8f76-7s8n7\" (UID: \"4f93a18d-29a1-4225-b99a-c801b91ce8d0\") " pod="openshift-image-registry/image-registry-66df7c8f76-7s8n7" Feb 27 07:54:00 crc kubenswrapper[4612]: I0227 07:54:00.751300 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/4f93a18d-29a1-4225-b99a-c801b91ce8d0-installation-pull-secrets\") pod \"image-registry-66df7c8f76-7s8n7\" (UID: \"4f93a18d-29a1-4225-b99a-c801b91ce8d0\") " pod="openshift-image-registry/image-registry-66df7c8f76-7s8n7" Feb 27 07:54:00 crc kubenswrapper[4612]: I0227 07:54:00.752115 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/4f93a18d-29a1-4225-b99a-c801b91ce8d0-registry-tls\") pod \"image-registry-66df7c8f76-7s8n7\" (UID: \"4f93a18d-29a1-4225-b99a-c801b91ce8d0\") " pod="openshift-image-registry/image-registry-66df7c8f76-7s8n7" Feb 27 07:54:00 crc kubenswrapper[4612]: I0227 07:54:00.755154 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/4f93a18d-29a1-4225-b99a-c801b91ce8d0-bound-sa-token\") pod \"image-registry-66df7c8f76-7s8n7\" (UID: \"4f93a18d-29a1-4225-b99a-c801b91ce8d0\") " pod="openshift-image-registry/image-registry-66df7c8f76-7s8n7" Feb 27 07:54:00 crc kubenswrapper[4612]: I0227 07:54:00.755383 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rv5nr\" (UniqueName: \"kubernetes.io/projected/4f93a18d-29a1-4225-b99a-c801b91ce8d0-kube-api-access-rv5nr\") pod \"image-registry-66df7c8f76-7s8n7\" (UID: \"4f93a18d-29a1-4225-b99a-c801b91ce8d0\") " pod="openshift-image-registry/image-registry-66df7c8f76-7s8n7" Feb 27 07:54:00 crc kubenswrapper[4612]: I0227 07:54:00.833808 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536314-5bct5"] Feb 27 07:54:00 crc kubenswrapper[4612]: I0227 07:54:00.873189 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536314-5bct5" event={"ID":"01363531-9d69-4f11-90f2-d4ee8a952b1a","Type":"ContainerStarted","Data":"cd4a4d4736822d8ff6af0f92b2194309bd4abed31cea08531e8a1201e0b16fbe"} Feb 27 07:54:00 crc kubenswrapper[4612]: I0227 07:54:00.893818 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-7s8n7" Feb 27 07:54:01 crc kubenswrapper[4612]: I0227 07:54:01.056458 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-7s8n7"] Feb 27 07:54:01 crc kubenswrapper[4612]: I0227 07:54:01.882019 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-7s8n7" event={"ID":"4f93a18d-29a1-4225-b99a-c801b91ce8d0","Type":"ContainerStarted","Data":"44ae448c8c0cd62b8388e3efbd4fb24d4604867604bb6abc95abb3b38a04c007"} Feb 27 07:54:01 crc kubenswrapper[4612]: I0227 07:54:01.882300 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-7s8n7" Feb 27 07:54:01 crc kubenswrapper[4612]: I0227 07:54:01.882317 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-7s8n7" event={"ID":"4f93a18d-29a1-4225-b99a-c801b91ce8d0","Type":"ContainerStarted","Data":"01c96a15fee0e0155a49d5e09954b6b69fd2ca85a5342312e5ed3b47b8537b5a"} Feb 27 07:54:01 crc kubenswrapper[4612]: I0227 07:54:01.903766 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-7s8n7" podStartSLOduration=1.903738392 podStartE2EDuration="1.903738392s" podCreationTimestamp="2026-02-27 07:54:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 07:54:01.895380799 +0000 UTC m=+299.749310807" watchObservedRunningTime="2026-02-27 07:54:01.903738392 +0000 UTC m=+299.757668390" Feb 27 07:54:02 crc kubenswrapper[4612]: I0227 07:54:02.128238 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-55bb989677-vtr6v"] Feb 27 07:54:02 crc kubenswrapper[4612]: I0227 07:54:02.128486 4612 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-55bb989677-vtr6v" podUID="675d6b43-fe8b-41b1-b52b-3f5389fffae1" containerName="route-controller-manager" containerID="cri-o://da4b53138f278105fda17730ca664d0a4bafcc5097bc37964b38c175c6943e34" gracePeriod=30 Feb 27 07:54:02 crc kubenswrapper[4612]: I0227 07:54:02.458834 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-55bb989677-vtr6v" Feb 27 07:54:02 crc kubenswrapper[4612]: I0227 07:54:02.551267 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9stsn\" (UniqueName: \"kubernetes.io/projected/675d6b43-fe8b-41b1-b52b-3f5389fffae1-kube-api-access-9stsn\") pod \"675d6b43-fe8b-41b1-b52b-3f5389fffae1\" (UID: \"675d6b43-fe8b-41b1-b52b-3f5389fffae1\") " Feb 27 07:54:02 crc kubenswrapper[4612]: I0227 07:54:02.551584 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/675d6b43-fe8b-41b1-b52b-3f5389fffae1-serving-cert\") pod \"675d6b43-fe8b-41b1-b52b-3f5389fffae1\" (UID: \"675d6b43-fe8b-41b1-b52b-3f5389fffae1\") " Feb 27 07:54:02 crc kubenswrapper[4612]: I0227 07:54:02.552869 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/675d6b43-fe8b-41b1-b52b-3f5389fffae1-client-ca\") pod \"675d6b43-fe8b-41b1-b52b-3f5389fffae1\" (UID: \"675d6b43-fe8b-41b1-b52b-3f5389fffae1\") " Feb 27 07:54:02 crc kubenswrapper[4612]: I0227 07:54:02.552973 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/675d6b43-fe8b-41b1-b52b-3f5389fffae1-config\") pod \"675d6b43-fe8b-41b1-b52b-3f5389fffae1\" (UID: \"675d6b43-fe8b-41b1-b52b-3f5389fffae1\") " Feb 27 07:54:02 crc kubenswrapper[4612]: I0227 07:54:02.553345 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/675d6b43-fe8b-41b1-b52b-3f5389fffae1-client-ca" (OuterVolumeSpecName: "client-ca") pod "675d6b43-fe8b-41b1-b52b-3f5389fffae1" (UID: "675d6b43-fe8b-41b1-b52b-3f5389fffae1"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 07:54:02 crc kubenswrapper[4612]: I0227 07:54:02.553465 4612 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/675d6b43-fe8b-41b1-b52b-3f5389fffae1-client-ca\") on node \"crc\" DevicePath \"\"" Feb 27 07:54:02 crc kubenswrapper[4612]: I0227 07:54:02.553723 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/675d6b43-fe8b-41b1-b52b-3f5389fffae1-config" (OuterVolumeSpecName: "config") pod "675d6b43-fe8b-41b1-b52b-3f5389fffae1" (UID: "675d6b43-fe8b-41b1-b52b-3f5389fffae1"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 07:54:02 crc kubenswrapper[4612]: I0227 07:54:02.556667 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/675d6b43-fe8b-41b1-b52b-3f5389fffae1-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "675d6b43-fe8b-41b1-b52b-3f5389fffae1" (UID: "675d6b43-fe8b-41b1-b52b-3f5389fffae1"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 07:54:02 crc kubenswrapper[4612]: I0227 07:54:02.556842 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/675d6b43-fe8b-41b1-b52b-3f5389fffae1-kube-api-access-9stsn" (OuterVolumeSpecName: "kube-api-access-9stsn") pod "675d6b43-fe8b-41b1-b52b-3f5389fffae1" (UID: "675d6b43-fe8b-41b1-b52b-3f5389fffae1"). InnerVolumeSpecName "kube-api-access-9stsn". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 07:54:02 crc kubenswrapper[4612]: I0227 07:54:02.654748 4612 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/675d6b43-fe8b-41b1-b52b-3f5389fffae1-config\") on node \"crc\" DevicePath \"\"" Feb 27 07:54:02 crc kubenswrapper[4612]: I0227 07:54:02.655131 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9stsn\" (UniqueName: \"kubernetes.io/projected/675d6b43-fe8b-41b1-b52b-3f5389fffae1-kube-api-access-9stsn\") on node \"crc\" DevicePath \"\"" Feb 27 07:54:02 crc kubenswrapper[4612]: I0227 07:54:02.655267 4612 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/675d6b43-fe8b-41b1-b52b-3f5389fffae1-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 27 07:54:02 crc kubenswrapper[4612]: I0227 07:54:02.890775 4612 generic.go:334] "Generic (PLEG): container finished" podID="01363531-9d69-4f11-90f2-d4ee8a952b1a" containerID="be89e07ad393203e501bac58eb8537477d3eb47a0b38a39c6448750bdc8f791a" exitCode=0 Feb 27 07:54:02 crc kubenswrapper[4612]: I0227 07:54:02.890853 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536314-5bct5" event={"ID":"01363531-9d69-4f11-90f2-d4ee8a952b1a","Type":"ContainerDied","Data":"be89e07ad393203e501bac58eb8537477d3eb47a0b38a39c6448750bdc8f791a"} Feb 27 07:54:02 crc kubenswrapper[4612]: I0227 07:54:02.893862 4612 generic.go:334] "Generic (PLEG): container finished" podID="675d6b43-fe8b-41b1-b52b-3f5389fffae1" containerID="da4b53138f278105fda17730ca664d0a4bafcc5097bc37964b38c175c6943e34" exitCode=0 Feb 27 07:54:02 crc kubenswrapper[4612]: I0227 07:54:02.893899 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-55bb989677-vtr6v" Feb 27 07:54:02 crc kubenswrapper[4612]: I0227 07:54:02.893931 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-55bb989677-vtr6v" event={"ID":"675d6b43-fe8b-41b1-b52b-3f5389fffae1","Type":"ContainerDied","Data":"da4b53138f278105fda17730ca664d0a4bafcc5097bc37964b38c175c6943e34"} Feb 27 07:54:02 crc kubenswrapper[4612]: I0227 07:54:02.893947 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-55bb989677-vtr6v" event={"ID":"675d6b43-fe8b-41b1-b52b-3f5389fffae1","Type":"ContainerDied","Data":"ba9aeee41cbd096c29a1a627ab74765aeae706ea0a8d501030552509964be7c9"} Feb 27 07:54:02 crc kubenswrapper[4612]: I0227 07:54:02.893962 4612 scope.go:117] "RemoveContainer" containerID="da4b53138f278105fda17730ca664d0a4bafcc5097bc37964b38c175c6943e34" Feb 27 07:54:02 crc kubenswrapper[4612]: I0227 07:54:02.914978 4612 scope.go:117] "RemoveContainer" containerID="da4b53138f278105fda17730ca664d0a4bafcc5097bc37964b38c175c6943e34" Feb 27 07:54:02 crc kubenswrapper[4612]: E0227 07:54:02.915405 4612 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"da4b53138f278105fda17730ca664d0a4bafcc5097bc37964b38c175c6943e34\": container with ID starting with da4b53138f278105fda17730ca664d0a4bafcc5097bc37964b38c175c6943e34 not found: ID does not exist" containerID="da4b53138f278105fda17730ca664d0a4bafcc5097bc37964b38c175c6943e34" Feb 27 07:54:02 crc kubenswrapper[4612]: I0227 07:54:02.915485 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"da4b53138f278105fda17730ca664d0a4bafcc5097bc37964b38c175c6943e34"} err="failed to get container status \"da4b53138f278105fda17730ca664d0a4bafcc5097bc37964b38c175c6943e34\": rpc error: code = NotFound desc = could not find container \"da4b53138f278105fda17730ca664d0a4bafcc5097bc37964b38c175c6943e34\": container with ID starting with da4b53138f278105fda17730ca664d0a4bafcc5097bc37964b38c175c6943e34 not found: ID does not exist" Feb 27 07:54:02 crc kubenswrapper[4612]: I0227 07:54:02.926667 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-55bb989677-vtr6v"] Feb 27 07:54:02 crc kubenswrapper[4612]: I0227 07:54:02.933119 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-55bb989677-vtr6v"] Feb 27 07:54:03 crc kubenswrapper[4612]: I0227 07:54:03.835653 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5cffc6cf87-grxgq"] Feb 27 07:54:03 crc kubenswrapper[4612]: E0227 07:54:03.836376 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="675d6b43-fe8b-41b1-b52b-3f5389fffae1" containerName="route-controller-manager" Feb 27 07:54:03 crc kubenswrapper[4612]: I0227 07:54:03.836400 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="675d6b43-fe8b-41b1-b52b-3f5389fffae1" containerName="route-controller-manager" Feb 27 07:54:03 crc kubenswrapper[4612]: I0227 07:54:03.836572 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="675d6b43-fe8b-41b1-b52b-3f5389fffae1" containerName="route-controller-manager" Feb 27 07:54:03 crc kubenswrapper[4612]: I0227 07:54:03.837110 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5cffc6cf87-grxgq" Feb 27 07:54:03 crc kubenswrapper[4612]: I0227 07:54:03.841183 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Feb 27 07:54:03 crc kubenswrapper[4612]: I0227 07:54:03.842310 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Feb 27 07:54:03 crc kubenswrapper[4612]: I0227 07:54:03.842423 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Feb 27 07:54:03 crc kubenswrapper[4612]: I0227 07:54:03.843030 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Feb 27 07:54:03 crc kubenswrapper[4612]: I0227 07:54:03.846469 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Feb 27 07:54:03 crc kubenswrapper[4612]: I0227 07:54:03.856607 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5cffc6cf87-grxgq"] Feb 27 07:54:03 crc kubenswrapper[4612]: I0227 07:54:03.859737 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Feb 27 07:54:03 crc kubenswrapper[4612]: I0227 07:54:03.880171 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bd42dbdf-e0ae-4ce1-b042-247dc5e0ed06-config\") pod \"route-controller-manager-5cffc6cf87-grxgq\" (UID: \"bd42dbdf-e0ae-4ce1-b042-247dc5e0ed06\") " pod="openshift-route-controller-manager/route-controller-manager-5cffc6cf87-grxgq" Feb 27 07:54:03 crc kubenswrapper[4612]: I0227 07:54:03.880247 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bd42dbdf-e0ae-4ce1-b042-247dc5e0ed06-serving-cert\") pod \"route-controller-manager-5cffc6cf87-grxgq\" (UID: \"bd42dbdf-e0ae-4ce1-b042-247dc5e0ed06\") " pod="openshift-route-controller-manager/route-controller-manager-5cffc6cf87-grxgq" Feb 27 07:54:03 crc kubenswrapper[4612]: I0227 07:54:03.880729 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/bd42dbdf-e0ae-4ce1-b042-247dc5e0ed06-client-ca\") pod \"route-controller-manager-5cffc6cf87-grxgq\" (UID: \"bd42dbdf-e0ae-4ce1-b042-247dc5e0ed06\") " pod="openshift-route-controller-manager/route-controller-manager-5cffc6cf87-grxgq" Feb 27 07:54:03 crc kubenswrapper[4612]: I0227 07:54:03.880893 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7m22v\" (UniqueName: \"kubernetes.io/projected/bd42dbdf-e0ae-4ce1-b042-247dc5e0ed06-kube-api-access-7m22v\") pod \"route-controller-manager-5cffc6cf87-grxgq\" (UID: \"bd42dbdf-e0ae-4ce1-b042-247dc5e0ed06\") " pod="openshift-route-controller-manager/route-controller-manager-5cffc6cf87-grxgq" Feb 27 07:54:03 crc kubenswrapper[4612]: I0227 07:54:03.981616 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/bd42dbdf-e0ae-4ce1-b042-247dc5e0ed06-client-ca\") pod \"route-controller-manager-5cffc6cf87-grxgq\" (UID: \"bd42dbdf-e0ae-4ce1-b042-247dc5e0ed06\") " pod="openshift-route-controller-manager/route-controller-manager-5cffc6cf87-grxgq" Feb 27 07:54:03 crc kubenswrapper[4612]: I0227 07:54:03.982011 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7m22v\" (UniqueName: \"kubernetes.io/projected/bd42dbdf-e0ae-4ce1-b042-247dc5e0ed06-kube-api-access-7m22v\") pod \"route-controller-manager-5cffc6cf87-grxgq\" (UID: \"bd42dbdf-e0ae-4ce1-b042-247dc5e0ed06\") " pod="openshift-route-controller-manager/route-controller-manager-5cffc6cf87-grxgq" Feb 27 07:54:03 crc kubenswrapper[4612]: I0227 07:54:03.982046 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bd42dbdf-e0ae-4ce1-b042-247dc5e0ed06-config\") pod \"route-controller-manager-5cffc6cf87-grxgq\" (UID: \"bd42dbdf-e0ae-4ce1-b042-247dc5e0ed06\") " pod="openshift-route-controller-manager/route-controller-manager-5cffc6cf87-grxgq" Feb 27 07:54:03 crc kubenswrapper[4612]: I0227 07:54:03.982062 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bd42dbdf-e0ae-4ce1-b042-247dc5e0ed06-serving-cert\") pod \"route-controller-manager-5cffc6cf87-grxgq\" (UID: \"bd42dbdf-e0ae-4ce1-b042-247dc5e0ed06\") " pod="openshift-route-controller-manager/route-controller-manager-5cffc6cf87-grxgq" Feb 27 07:54:03 crc kubenswrapper[4612]: I0227 07:54:03.983012 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/bd42dbdf-e0ae-4ce1-b042-247dc5e0ed06-client-ca\") pod \"route-controller-manager-5cffc6cf87-grxgq\" (UID: \"bd42dbdf-e0ae-4ce1-b042-247dc5e0ed06\") " pod="openshift-route-controller-manager/route-controller-manager-5cffc6cf87-grxgq" Feb 27 07:54:03 crc kubenswrapper[4612]: I0227 07:54:03.984259 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bd42dbdf-e0ae-4ce1-b042-247dc5e0ed06-config\") pod \"route-controller-manager-5cffc6cf87-grxgq\" (UID: \"bd42dbdf-e0ae-4ce1-b042-247dc5e0ed06\") " pod="openshift-route-controller-manager/route-controller-manager-5cffc6cf87-grxgq" Feb 27 07:54:03 crc kubenswrapper[4612]: I0227 07:54:03.989912 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bd42dbdf-e0ae-4ce1-b042-247dc5e0ed06-serving-cert\") pod \"route-controller-manager-5cffc6cf87-grxgq\" (UID: \"bd42dbdf-e0ae-4ce1-b042-247dc5e0ed06\") " pod="openshift-route-controller-manager/route-controller-manager-5cffc6cf87-grxgq" Feb 27 07:54:04 crc kubenswrapper[4612]: I0227 07:54:04.006170 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7m22v\" (UniqueName: \"kubernetes.io/projected/bd42dbdf-e0ae-4ce1-b042-247dc5e0ed06-kube-api-access-7m22v\") pod \"route-controller-manager-5cffc6cf87-grxgq\" (UID: \"bd42dbdf-e0ae-4ce1-b042-247dc5e0ed06\") " pod="openshift-route-controller-manager/route-controller-manager-5cffc6cf87-grxgq" Feb 27 07:54:04 crc kubenswrapper[4612]: I0227 07:54:04.158367 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536314-5bct5" Feb 27 07:54:04 crc kubenswrapper[4612]: I0227 07:54:04.178111 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5cffc6cf87-grxgq" Feb 27 07:54:04 crc kubenswrapper[4612]: I0227 07:54:04.183788 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-858xs\" (UniqueName: \"kubernetes.io/projected/01363531-9d69-4f11-90f2-d4ee8a952b1a-kube-api-access-858xs\") pod \"01363531-9d69-4f11-90f2-d4ee8a952b1a\" (UID: \"01363531-9d69-4f11-90f2-d4ee8a952b1a\") " Feb 27 07:54:04 crc kubenswrapper[4612]: I0227 07:54:04.196938 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01363531-9d69-4f11-90f2-d4ee8a952b1a-kube-api-access-858xs" (OuterVolumeSpecName: "kube-api-access-858xs") pod "01363531-9d69-4f11-90f2-d4ee8a952b1a" (UID: "01363531-9d69-4f11-90f2-d4ee8a952b1a"). InnerVolumeSpecName "kube-api-access-858xs". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 07:54:04 crc kubenswrapper[4612]: I0227 07:54:04.285610 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-858xs\" (UniqueName: \"kubernetes.io/projected/01363531-9d69-4f11-90f2-d4ee8a952b1a-kube-api-access-858xs\") on node \"crc\" DevicePath \"\"" Feb 27 07:54:04 crc kubenswrapper[4612]: I0227 07:54:04.412639 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5cffc6cf87-grxgq"] Feb 27 07:54:04 crc kubenswrapper[4612]: I0227 07:54:04.858678 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="675d6b43-fe8b-41b1-b52b-3f5389fffae1" path="/var/lib/kubelet/pods/675d6b43-fe8b-41b1-b52b-3f5389fffae1/volumes" Feb 27 07:54:04 crc kubenswrapper[4612]: I0227 07:54:04.924640 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536314-5bct5" event={"ID":"01363531-9d69-4f11-90f2-d4ee8a952b1a","Type":"ContainerDied","Data":"cd4a4d4736822d8ff6af0f92b2194309bd4abed31cea08531e8a1201e0b16fbe"} Feb 27 07:54:04 crc kubenswrapper[4612]: I0227 07:54:04.924679 4612 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cd4a4d4736822d8ff6af0f92b2194309bd4abed31cea08531e8a1201e0b16fbe" Feb 27 07:54:04 crc kubenswrapper[4612]: I0227 07:54:04.924748 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536314-5bct5" Feb 27 07:54:04 crc kubenswrapper[4612]: I0227 07:54:04.927243 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5cffc6cf87-grxgq" event={"ID":"bd42dbdf-e0ae-4ce1-b042-247dc5e0ed06","Type":"ContainerStarted","Data":"92bc5fb8f6440c103b9e7bf9742d08371de0bca279febe51ab246726a0b5d87c"} Feb 27 07:54:04 crc kubenswrapper[4612]: I0227 07:54:04.927289 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5cffc6cf87-grxgq" event={"ID":"bd42dbdf-e0ae-4ce1-b042-247dc5e0ed06","Type":"ContainerStarted","Data":"2d45ac50f568a1970cd4ec729f52a62ff3d9f2c2ea0d4a5395a7387fdeac0ca5"} Feb 27 07:54:04 crc kubenswrapper[4612]: I0227 07:54:04.927619 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-5cffc6cf87-grxgq" Feb 27 07:54:04 crc kubenswrapper[4612]: I0227 07:54:04.951007 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-5cffc6cf87-grxgq" podStartSLOduration=2.950988132 podStartE2EDuration="2.950988132s" podCreationTimestamp="2026-02-27 07:54:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 07:54:04.94893628 +0000 UTC m=+302.802866278" watchObservedRunningTime="2026-02-27 07:54:04.950988132 +0000 UTC m=+302.804918130" Feb 27 07:54:05 crc kubenswrapper[4612]: I0227 07:54:05.137895 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-5cffc6cf87-grxgq" Feb 27 07:54:20 crc kubenswrapper[4612]: I0227 07:54:20.904103 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-7s8n7" Feb 27 07:54:21 crc kubenswrapper[4612]: I0227 07:54:21.000252 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-pdjlb"] Feb 27 07:54:42 crc kubenswrapper[4612]: I0227 07:54:42.151134 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-7c589966c4-2pn6j"] Feb 27 07:54:42 crc kubenswrapper[4612]: I0227 07:54:42.151847 4612 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-7c589966c4-2pn6j" podUID="3301aca5-f994-4520-b8bb-02fef2be45a2" containerName="controller-manager" containerID="cri-o://165e292b2620c937ad8cbcf15871bc28f6a4528d8076a9c80f677ae6a58ac0fe" gracePeriod=30 Feb 27 07:54:42 crc kubenswrapper[4612]: I0227 07:54:42.503121 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7c589966c4-2pn6j" Feb 27 07:54:42 crc kubenswrapper[4612]: I0227 07:54:42.679346 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/3301aca5-f994-4520-b8bb-02fef2be45a2-proxy-ca-bundles\") pod \"3301aca5-f994-4520-b8bb-02fef2be45a2\" (UID: \"3301aca5-f994-4520-b8bb-02fef2be45a2\") " Feb 27 07:54:42 crc kubenswrapper[4612]: I0227 07:54:42.679428 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qpfzh\" (UniqueName: \"kubernetes.io/projected/3301aca5-f994-4520-b8bb-02fef2be45a2-kube-api-access-qpfzh\") pod \"3301aca5-f994-4520-b8bb-02fef2be45a2\" (UID: \"3301aca5-f994-4520-b8bb-02fef2be45a2\") " Feb 27 07:54:42 crc kubenswrapper[4612]: I0227 07:54:42.679482 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/3301aca5-f994-4520-b8bb-02fef2be45a2-client-ca\") pod \"3301aca5-f994-4520-b8bb-02fef2be45a2\" (UID: \"3301aca5-f994-4520-b8bb-02fef2be45a2\") " Feb 27 07:54:42 crc kubenswrapper[4612]: I0227 07:54:42.679512 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3301aca5-f994-4520-b8bb-02fef2be45a2-serving-cert\") pod \"3301aca5-f994-4520-b8bb-02fef2be45a2\" (UID: \"3301aca5-f994-4520-b8bb-02fef2be45a2\") " Feb 27 07:54:42 crc kubenswrapper[4612]: I0227 07:54:42.679550 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3301aca5-f994-4520-b8bb-02fef2be45a2-config\") pod \"3301aca5-f994-4520-b8bb-02fef2be45a2\" (UID: \"3301aca5-f994-4520-b8bb-02fef2be45a2\") " Feb 27 07:54:42 crc kubenswrapper[4612]: I0227 07:54:42.680801 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3301aca5-f994-4520-b8bb-02fef2be45a2-config" (OuterVolumeSpecName: "config") pod "3301aca5-f994-4520-b8bb-02fef2be45a2" (UID: "3301aca5-f994-4520-b8bb-02fef2be45a2"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 07:54:42 crc kubenswrapper[4612]: I0227 07:54:42.681033 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3301aca5-f994-4520-b8bb-02fef2be45a2-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "3301aca5-f994-4520-b8bb-02fef2be45a2" (UID: "3301aca5-f994-4520-b8bb-02fef2be45a2"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 07:54:42 crc kubenswrapper[4612]: I0227 07:54:42.681742 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3301aca5-f994-4520-b8bb-02fef2be45a2-client-ca" (OuterVolumeSpecName: "client-ca") pod "3301aca5-f994-4520-b8bb-02fef2be45a2" (UID: "3301aca5-f994-4520-b8bb-02fef2be45a2"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 07:54:42 crc kubenswrapper[4612]: I0227 07:54:42.684970 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3301aca5-f994-4520-b8bb-02fef2be45a2-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "3301aca5-f994-4520-b8bb-02fef2be45a2" (UID: "3301aca5-f994-4520-b8bb-02fef2be45a2"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 07:54:42 crc kubenswrapper[4612]: I0227 07:54:42.690017 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3301aca5-f994-4520-b8bb-02fef2be45a2-kube-api-access-qpfzh" (OuterVolumeSpecName: "kube-api-access-qpfzh") pod "3301aca5-f994-4520-b8bb-02fef2be45a2" (UID: "3301aca5-f994-4520-b8bb-02fef2be45a2"). InnerVolumeSpecName "kube-api-access-qpfzh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 07:54:42 crc kubenswrapper[4612]: I0227 07:54:42.783401 4612 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3301aca5-f994-4520-b8bb-02fef2be45a2-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 27 07:54:42 crc kubenswrapper[4612]: I0227 07:54:42.783460 4612 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3301aca5-f994-4520-b8bb-02fef2be45a2-config\") on node \"crc\" DevicePath \"\"" Feb 27 07:54:42 crc kubenswrapper[4612]: I0227 07:54:42.783489 4612 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/3301aca5-f994-4520-b8bb-02fef2be45a2-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Feb 27 07:54:42 crc kubenswrapper[4612]: I0227 07:54:42.783510 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qpfzh\" (UniqueName: \"kubernetes.io/projected/3301aca5-f994-4520-b8bb-02fef2be45a2-kube-api-access-qpfzh\") on node \"crc\" DevicePath \"\"" Feb 27 07:54:42 crc kubenswrapper[4612]: I0227 07:54:42.783529 4612 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/3301aca5-f994-4520-b8bb-02fef2be45a2-client-ca\") on node \"crc\" DevicePath \"\"" Feb 27 07:54:43 crc kubenswrapper[4612]: I0227 07:54:43.180083 4612 generic.go:334] "Generic (PLEG): container finished" podID="3301aca5-f994-4520-b8bb-02fef2be45a2" containerID="165e292b2620c937ad8cbcf15871bc28f6a4528d8076a9c80f677ae6a58ac0fe" exitCode=0 Feb 27 07:54:43 crc kubenswrapper[4612]: I0227 07:54:43.180180 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7c589966c4-2pn6j" event={"ID":"3301aca5-f994-4520-b8bb-02fef2be45a2","Type":"ContainerDied","Data":"165e292b2620c937ad8cbcf15871bc28f6a4528d8076a9c80f677ae6a58ac0fe"} Feb 27 07:54:43 crc kubenswrapper[4612]: I0227 07:54:43.180246 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7c589966c4-2pn6j" event={"ID":"3301aca5-f994-4520-b8bb-02fef2be45a2","Type":"ContainerDied","Data":"c73a87882715e7750bf07df39d4484fc62a3a3717ad1f7aec349ce2e62c57028"} Feb 27 07:54:43 crc kubenswrapper[4612]: I0227 07:54:43.180250 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7c589966c4-2pn6j" Feb 27 07:54:43 crc kubenswrapper[4612]: I0227 07:54:43.180317 4612 scope.go:117] "RemoveContainer" containerID="165e292b2620c937ad8cbcf15871bc28f6a4528d8076a9c80f677ae6a58ac0fe" Feb 27 07:54:43 crc kubenswrapper[4612]: I0227 07:54:43.209488 4612 scope.go:117] "RemoveContainer" containerID="165e292b2620c937ad8cbcf15871bc28f6a4528d8076a9c80f677ae6a58ac0fe" Feb 27 07:54:43 crc kubenswrapper[4612]: E0227 07:54:43.210090 4612 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"165e292b2620c937ad8cbcf15871bc28f6a4528d8076a9c80f677ae6a58ac0fe\": container with ID starting with 165e292b2620c937ad8cbcf15871bc28f6a4528d8076a9c80f677ae6a58ac0fe not found: ID does not exist" containerID="165e292b2620c937ad8cbcf15871bc28f6a4528d8076a9c80f677ae6a58ac0fe" Feb 27 07:54:43 crc kubenswrapper[4612]: I0227 07:54:43.210115 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"165e292b2620c937ad8cbcf15871bc28f6a4528d8076a9c80f677ae6a58ac0fe"} err="failed to get container status \"165e292b2620c937ad8cbcf15871bc28f6a4528d8076a9c80f677ae6a58ac0fe\": rpc error: code = NotFound desc = could not find container \"165e292b2620c937ad8cbcf15871bc28f6a4528d8076a9c80f677ae6a58ac0fe\": container with ID starting with 165e292b2620c937ad8cbcf15871bc28f6a4528d8076a9c80f677ae6a58ac0fe not found: ID does not exist" Feb 27 07:54:43 crc kubenswrapper[4612]: I0227 07:54:43.216686 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-7c589966c4-2pn6j"] Feb 27 07:54:43 crc kubenswrapper[4612]: I0227 07:54:43.221324 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-7c589966c4-2pn6j"] Feb 27 07:54:43 crc kubenswrapper[4612]: I0227 07:54:43.867973 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-7c6f84b55d-bjrz5"] Feb 27 07:54:43 crc kubenswrapper[4612]: E0227 07:54:43.868360 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3301aca5-f994-4520-b8bb-02fef2be45a2" containerName="controller-manager" Feb 27 07:54:43 crc kubenswrapper[4612]: I0227 07:54:43.868380 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="3301aca5-f994-4520-b8bb-02fef2be45a2" containerName="controller-manager" Feb 27 07:54:43 crc kubenswrapper[4612]: E0227 07:54:43.868400 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="01363531-9d69-4f11-90f2-d4ee8a952b1a" containerName="oc" Feb 27 07:54:43 crc kubenswrapper[4612]: I0227 07:54:43.868413 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="01363531-9d69-4f11-90f2-d4ee8a952b1a" containerName="oc" Feb 27 07:54:43 crc kubenswrapper[4612]: I0227 07:54:43.868580 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="3301aca5-f994-4520-b8bb-02fef2be45a2" containerName="controller-manager" Feb 27 07:54:43 crc kubenswrapper[4612]: I0227 07:54:43.868599 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="01363531-9d69-4f11-90f2-d4ee8a952b1a" containerName="oc" Feb 27 07:54:43 crc kubenswrapper[4612]: I0227 07:54:43.869351 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7c6f84b55d-bjrz5" Feb 27 07:54:43 crc kubenswrapper[4612]: I0227 07:54:43.873839 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Feb 27 07:54:43 crc kubenswrapper[4612]: I0227 07:54:43.876325 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Feb 27 07:54:43 crc kubenswrapper[4612]: I0227 07:54:43.876927 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Feb 27 07:54:43 crc kubenswrapper[4612]: I0227 07:54:43.878078 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Feb 27 07:54:43 crc kubenswrapper[4612]: I0227 07:54:43.879465 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Feb 27 07:54:43 crc kubenswrapper[4612]: I0227 07:54:43.880085 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Feb 27 07:54:43 crc kubenswrapper[4612]: I0227 07:54:43.880566 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-7c6f84b55d-bjrz5"] Feb 27 07:54:43 crc kubenswrapper[4612]: I0227 07:54:43.888826 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Feb 27 07:54:44 crc kubenswrapper[4612]: I0227 07:54:44.001182 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6e1f0366-eeea-4ca6-9834-85ddc9cae0c1-serving-cert\") pod \"controller-manager-7c6f84b55d-bjrz5\" (UID: \"6e1f0366-eeea-4ca6-9834-85ddc9cae0c1\") " pod="openshift-controller-manager/controller-manager-7c6f84b55d-bjrz5" Feb 27 07:54:44 crc kubenswrapper[4612]: I0227 07:54:44.001570 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6e1f0366-eeea-4ca6-9834-85ddc9cae0c1-config\") pod \"controller-manager-7c6f84b55d-bjrz5\" (UID: \"6e1f0366-eeea-4ca6-9834-85ddc9cae0c1\") " pod="openshift-controller-manager/controller-manager-7c6f84b55d-bjrz5" Feb 27 07:54:44 crc kubenswrapper[4612]: I0227 07:54:44.001748 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k8xhc\" (UniqueName: \"kubernetes.io/projected/6e1f0366-eeea-4ca6-9834-85ddc9cae0c1-kube-api-access-k8xhc\") pod \"controller-manager-7c6f84b55d-bjrz5\" (UID: \"6e1f0366-eeea-4ca6-9834-85ddc9cae0c1\") " pod="openshift-controller-manager/controller-manager-7c6f84b55d-bjrz5" Feb 27 07:54:44 crc kubenswrapper[4612]: I0227 07:54:44.001887 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/6e1f0366-eeea-4ca6-9834-85ddc9cae0c1-proxy-ca-bundles\") pod \"controller-manager-7c6f84b55d-bjrz5\" (UID: \"6e1f0366-eeea-4ca6-9834-85ddc9cae0c1\") " pod="openshift-controller-manager/controller-manager-7c6f84b55d-bjrz5" Feb 27 07:54:44 crc kubenswrapper[4612]: I0227 07:54:44.002064 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6e1f0366-eeea-4ca6-9834-85ddc9cae0c1-client-ca\") pod \"controller-manager-7c6f84b55d-bjrz5\" (UID: \"6e1f0366-eeea-4ca6-9834-85ddc9cae0c1\") " pod="openshift-controller-manager/controller-manager-7c6f84b55d-bjrz5" Feb 27 07:54:44 crc kubenswrapper[4612]: I0227 07:54:44.103074 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6e1f0366-eeea-4ca6-9834-85ddc9cae0c1-client-ca\") pod \"controller-manager-7c6f84b55d-bjrz5\" (UID: \"6e1f0366-eeea-4ca6-9834-85ddc9cae0c1\") " pod="openshift-controller-manager/controller-manager-7c6f84b55d-bjrz5" Feb 27 07:54:44 crc kubenswrapper[4612]: I0227 07:54:44.103193 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6e1f0366-eeea-4ca6-9834-85ddc9cae0c1-serving-cert\") pod \"controller-manager-7c6f84b55d-bjrz5\" (UID: \"6e1f0366-eeea-4ca6-9834-85ddc9cae0c1\") " pod="openshift-controller-manager/controller-manager-7c6f84b55d-bjrz5" Feb 27 07:54:44 crc kubenswrapper[4612]: I0227 07:54:44.103230 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6e1f0366-eeea-4ca6-9834-85ddc9cae0c1-config\") pod \"controller-manager-7c6f84b55d-bjrz5\" (UID: \"6e1f0366-eeea-4ca6-9834-85ddc9cae0c1\") " pod="openshift-controller-manager/controller-manager-7c6f84b55d-bjrz5" Feb 27 07:54:44 crc kubenswrapper[4612]: I0227 07:54:44.103289 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k8xhc\" (UniqueName: \"kubernetes.io/projected/6e1f0366-eeea-4ca6-9834-85ddc9cae0c1-kube-api-access-k8xhc\") pod \"controller-manager-7c6f84b55d-bjrz5\" (UID: \"6e1f0366-eeea-4ca6-9834-85ddc9cae0c1\") " pod="openshift-controller-manager/controller-manager-7c6f84b55d-bjrz5" Feb 27 07:54:44 crc kubenswrapper[4612]: I0227 07:54:44.103340 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/6e1f0366-eeea-4ca6-9834-85ddc9cae0c1-proxy-ca-bundles\") pod \"controller-manager-7c6f84b55d-bjrz5\" (UID: \"6e1f0366-eeea-4ca6-9834-85ddc9cae0c1\") " pod="openshift-controller-manager/controller-manager-7c6f84b55d-bjrz5" Feb 27 07:54:44 crc kubenswrapper[4612]: I0227 07:54:44.104555 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6e1f0366-eeea-4ca6-9834-85ddc9cae0c1-client-ca\") pod \"controller-manager-7c6f84b55d-bjrz5\" (UID: \"6e1f0366-eeea-4ca6-9834-85ddc9cae0c1\") " pod="openshift-controller-manager/controller-manager-7c6f84b55d-bjrz5" Feb 27 07:54:44 crc kubenswrapper[4612]: I0227 07:54:44.105160 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/6e1f0366-eeea-4ca6-9834-85ddc9cae0c1-proxy-ca-bundles\") pod \"controller-manager-7c6f84b55d-bjrz5\" (UID: \"6e1f0366-eeea-4ca6-9834-85ddc9cae0c1\") " pod="openshift-controller-manager/controller-manager-7c6f84b55d-bjrz5" Feb 27 07:54:44 crc kubenswrapper[4612]: I0227 07:54:44.106450 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6e1f0366-eeea-4ca6-9834-85ddc9cae0c1-config\") pod \"controller-manager-7c6f84b55d-bjrz5\" (UID: \"6e1f0366-eeea-4ca6-9834-85ddc9cae0c1\") " pod="openshift-controller-manager/controller-manager-7c6f84b55d-bjrz5" Feb 27 07:54:44 crc kubenswrapper[4612]: I0227 07:54:44.108060 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6e1f0366-eeea-4ca6-9834-85ddc9cae0c1-serving-cert\") pod \"controller-manager-7c6f84b55d-bjrz5\" (UID: \"6e1f0366-eeea-4ca6-9834-85ddc9cae0c1\") " pod="openshift-controller-manager/controller-manager-7c6f84b55d-bjrz5" Feb 27 07:54:44 crc kubenswrapper[4612]: I0227 07:54:44.135309 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k8xhc\" (UniqueName: \"kubernetes.io/projected/6e1f0366-eeea-4ca6-9834-85ddc9cae0c1-kube-api-access-k8xhc\") pod \"controller-manager-7c6f84b55d-bjrz5\" (UID: \"6e1f0366-eeea-4ca6-9834-85ddc9cae0c1\") " pod="openshift-controller-manager/controller-manager-7c6f84b55d-bjrz5" Feb 27 07:54:44 crc kubenswrapper[4612]: I0227 07:54:44.210825 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7c6f84b55d-bjrz5" Feb 27 07:54:44 crc kubenswrapper[4612]: I0227 07:54:44.428068 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-7c6f84b55d-bjrz5"] Feb 27 07:54:44 crc kubenswrapper[4612]: I0227 07:54:44.862063 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3301aca5-f994-4520-b8bb-02fef2be45a2" path="/var/lib/kubelet/pods/3301aca5-f994-4520-b8bb-02fef2be45a2/volumes" Feb 27 07:54:45 crc kubenswrapper[4612]: I0227 07:54:45.195082 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7c6f84b55d-bjrz5" event={"ID":"6e1f0366-eeea-4ca6-9834-85ddc9cae0c1","Type":"ContainerStarted","Data":"cfbc59f217026af2c77b6bfed1a0405b58cac1a6ddb1499cb3436c0b18f2dc78"} Feb 27 07:54:45 crc kubenswrapper[4612]: I0227 07:54:45.195129 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7c6f84b55d-bjrz5" event={"ID":"6e1f0366-eeea-4ca6-9834-85ddc9cae0c1","Type":"ContainerStarted","Data":"05f91638cdefe377aa069eca1194a3ea9899a6114dc5847b48752b5381ec28d2"} Feb 27 07:54:45 crc kubenswrapper[4612]: I0227 07:54:45.196010 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-7c6f84b55d-bjrz5" Feb 27 07:54:45 crc kubenswrapper[4612]: I0227 07:54:45.216307 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-7c6f84b55d-bjrz5" Feb 27 07:54:45 crc kubenswrapper[4612]: I0227 07:54:45.240645 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-7c6f84b55d-bjrz5" podStartSLOduration=3.240616961 podStartE2EDuration="3.240616961s" podCreationTimestamp="2026-02-27 07:54:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 07:54:45.232380693 +0000 UTC m=+343.086310701" watchObservedRunningTime="2026-02-27 07:54:45.240616961 +0000 UTC m=+343.094546999" Feb 27 07:54:46 crc kubenswrapper[4612]: I0227 07:54:46.058479 4612 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-pdjlb" podUID="db4d6d91-61f5-4beb-809c-63ae9028a911" containerName="registry" containerID="cri-o://4cc002862ee09b83f37e5daf8763574294e4fa40534d98b44c07ee66a930c157" gracePeriod=30 Feb 27 07:54:46 crc kubenswrapper[4612]: I0227 07:54:46.206097 4612 generic.go:334] "Generic (PLEG): container finished" podID="db4d6d91-61f5-4beb-809c-63ae9028a911" containerID="4cc002862ee09b83f37e5daf8763574294e4fa40534d98b44c07ee66a930c157" exitCode=0 Feb 27 07:54:46 crc kubenswrapper[4612]: I0227 07:54:46.206197 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-pdjlb" event={"ID":"db4d6d91-61f5-4beb-809c-63ae9028a911","Type":"ContainerDied","Data":"4cc002862ee09b83f37e5daf8763574294e4fa40534d98b44c07ee66a930c157"} Feb 27 07:54:46 crc kubenswrapper[4612]: I0227 07:54:46.462062 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-pdjlb" Feb 27 07:54:46 crc kubenswrapper[4612]: I0227 07:54:46.636682 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/db4d6d91-61f5-4beb-809c-63ae9028a911-installation-pull-secrets\") pod \"db4d6d91-61f5-4beb-809c-63ae9028a911\" (UID: \"db4d6d91-61f5-4beb-809c-63ae9028a911\") " Feb 27 07:54:46 crc kubenswrapper[4612]: I0227 07:54:46.636761 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2bbj5\" (UniqueName: \"kubernetes.io/projected/db4d6d91-61f5-4beb-809c-63ae9028a911-kube-api-access-2bbj5\") pod \"db4d6d91-61f5-4beb-809c-63ae9028a911\" (UID: \"db4d6d91-61f5-4beb-809c-63ae9028a911\") " Feb 27 07:54:46 crc kubenswrapper[4612]: I0227 07:54:46.636839 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/db4d6d91-61f5-4beb-809c-63ae9028a911-registry-certificates\") pod \"db4d6d91-61f5-4beb-809c-63ae9028a911\" (UID: \"db4d6d91-61f5-4beb-809c-63ae9028a911\") " Feb 27 07:54:46 crc kubenswrapper[4612]: I0227 07:54:46.636900 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/db4d6d91-61f5-4beb-809c-63ae9028a911-registry-tls\") pod \"db4d6d91-61f5-4beb-809c-63ae9028a911\" (UID: \"db4d6d91-61f5-4beb-809c-63ae9028a911\") " Feb 27 07:54:46 crc kubenswrapper[4612]: I0227 07:54:46.637192 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"db4d6d91-61f5-4beb-809c-63ae9028a911\" (UID: \"db4d6d91-61f5-4beb-809c-63ae9028a911\") " Feb 27 07:54:46 crc kubenswrapper[4612]: I0227 07:54:46.637262 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/db4d6d91-61f5-4beb-809c-63ae9028a911-bound-sa-token\") pod \"db4d6d91-61f5-4beb-809c-63ae9028a911\" (UID: \"db4d6d91-61f5-4beb-809c-63ae9028a911\") " Feb 27 07:54:46 crc kubenswrapper[4612]: I0227 07:54:46.637304 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/db4d6d91-61f5-4beb-809c-63ae9028a911-ca-trust-extracted\") pod \"db4d6d91-61f5-4beb-809c-63ae9028a911\" (UID: \"db4d6d91-61f5-4beb-809c-63ae9028a911\") " Feb 27 07:54:46 crc kubenswrapper[4612]: I0227 07:54:46.637327 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/db4d6d91-61f5-4beb-809c-63ae9028a911-trusted-ca\") pod \"db4d6d91-61f5-4beb-809c-63ae9028a911\" (UID: \"db4d6d91-61f5-4beb-809c-63ae9028a911\") " Feb 27 07:54:46 crc kubenswrapper[4612]: I0227 07:54:46.638323 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/db4d6d91-61f5-4beb-809c-63ae9028a911-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "db4d6d91-61f5-4beb-809c-63ae9028a911" (UID: "db4d6d91-61f5-4beb-809c-63ae9028a911"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 07:54:46 crc kubenswrapper[4612]: I0227 07:54:46.638464 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/db4d6d91-61f5-4beb-809c-63ae9028a911-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "db4d6d91-61f5-4beb-809c-63ae9028a911" (UID: "db4d6d91-61f5-4beb-809c-63ae9028a911"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 07:54:46 crc kubenswrapper[4612]: I0227 07:54:46.649837 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/db4d6d91-61f5-4beb-809c-63ae9028a911-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "db4d6d91-61f5-4beb-809c-63ae9028a911" (UID: "db4d6d91-61f5-4beb-809c-63ae9028a911"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 07:54:46 crc kubenswrapper[4612]: I0227 07:54:46.651562 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/db4d6d91-61f5-4beb-809c-63ae9028a911-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "db4d6d91-61f5-4beb-809c-63ae9028a911" (UID: "db4d6d91-61f5-4beb-809c-63ae9028a911"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 07:54:46 crc kubenswrapper[4612]: I0227 07:54:46.651632 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/db4d6d91-61f5-4beb-809c-63ae9028a911-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "db4d6d91-61f5-4beb-809c-63ae9028a911" (UID: "db4d6d91-61f5-4beb-809c-63ae9028a911"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 07:54:46 crc kubenswrapper[4612]: I0227 07:54:46.651718 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/db4d6d91-61f5-4beb-809c-63ae9028a911-kube-api-access-2bbj5" (OuterVolumeSpecName: "kube-api-access-2bbj5") pod "db4d6d91-61f5-4beb-809c-63ae9028a911" (UID: "db4d6d91-61f5-4beb-809c-63ae9028a911"). InnerVolumeSpecName "kube-api-access-2bbj5". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 07:54:46 crc kubenswrapper[4612]: I0227 07:54:46.656102 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "db4d6d91-61f5-4beb-809c-63ae9028a911" (UID: "db4d6d91-61f5-4beb-809c-63ae9028a911"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Feb 27 07:54:46 crc kubenswrapper[4612]: I0227 07:54:46.679336 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/db4d6d91-61f5-4beb-809c-63ae9028a911-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "db4d6d91-61f5-4beb-809c-63ae9028a911" (UID: "db4d6d91-61f5-4beb-809c-63ae9028a911"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 07:54:46 crc kubenswrapper[4612]: I0227 07:54:46.738415 4612 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/db4d6d91-61f5-4beb-809c-63ae9028a911-bound-sa-token\") on node \"crc\" DevicePath \"\"" Feb 27 07:54:46 crc kubenswrapper[4612]: I0227 07:54:46.738437 4612 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/db4d6d91-61f5-4beb-809c-63ae9028a911-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Feb 27 07:54:46 crc kubenswrapper[4612]: I0227 07:54:46.738447 4612 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/db4d6d91-61f5-4beb-809c-63ae9028a911-trusted-ca\") on node \"crc\" DevicePath \"\"" Feb 27 07:54:46 crc kubenswrapper[4612]: I0227 07:54:46.738456 4612 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/db4d6d91-61f5-4beb-809c-63ae9028a911-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Feb 27 07:54:46 crc kubenswrapper[4612]: I0227 07:54:46.738467 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2bbj5\" (UniqueName: \"kubernetes.io/projected/db4d6d91-61f5-4beb-809c-63ae9028a911-kube-api-access-2bbj5\") on node \"crc\" DevicePath \"\"" Feb 27 07:54:46 crc kubenswrapper[4612]: I0227 07:54:46.738474 4612 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/db4d6d91-61f5-4beb-809c-63ae9028a911-registry-certificates\") on node \"crc\" DevicePath \"\"" Feb 27 07:54:46 crc kubenswrapper[4612]: I0227 07:54:46.738482 4612 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/db4d6d91-61f5-4beb-809c-63ae9028a911-registry-tls\") on node \"crc\" DevicePath \"\"" Feb 27 07:54:47 crc kubenswrapper[4612]: I0227 07:54:47.213586 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-pdjlb" event={"ID":"db4d6d91-61f5-4beb-809c-63ae9028a911","Type":"ContainerDied","Data":"064fd80a4871b89d6878288363b599fb3e86994c3833cddac443d4d4dc4c77cb"} Feb 27 07:54:47 crc kubenswrapper[4612]: I0227 07:54:47.213641 4612 scope.go:117] "RemoveContainer" containerID="4cc002862ee09b83f37e5daf8763574294e4fa40534d98b44c07ee66a930c157" Feb 27 07:54:47 crc kubenswrapper[4612]: I0227 07:54:47.213846 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-pdjlb" Feb 27 07:54:47 crc kubenswrapper[4612]: I0227 07:54:47.236524 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-pdjlb"] Feb 27 07:54:47 crc kubenswrapper[4612]: I0227 07:54:47.239720 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-pdjlb"] Feb 27 07:54:48 crc kubenswrapper[4612]: I0227 07:54:48.867781 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="db4d6d91-61f5-4beb-809c-63ae9028a911" path="/var/lib/kubelet/pods/db4d6d91-61f5-4beb-809c-63ae9028a911/volumes" Feb 27 07:54:52 crc kubenswrapper[4612]: I0227 07:54:52.187694 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-qlhr9"] Feb 27 07:54:52 crc kubenswrapper[4612]: I0227 07:54:52.190644 4612 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-qlhr9" podUID="7faf5006-23b1-4ef4-ad29-e0e676340a7c" containerName="registry-server" containerID="cri-o://bc3d2a71143833b070c3078c2c2694503dc694341f36a9dad42134afa1a0c8c9" gracePeriod=30 Feb 27 07:54:52 crc kubenswrapper[4612]: I0227 07:54:52.193859 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-p5fp8"] Feb 27 07:54:52 crc kubenswrapper[4612]: I0227 07:54:52.194206 4612 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-p5fp8" podUID="201de829-c0b7-4d50-9a3e-ef65d3c06916" containerName="registry-server" containerID="cri-o://a93c39714c2e5eda46c61ea58eae7a9e3f1ba38218d4c700a99e6cc7b1c55c31" gracePeriod=30 Feb 27 07:54:52 crc kubenswrapper[4612]: I0227 07:54:52.210400 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-hp72h"] Feb 27 07:54:52 crc kubenswrapper[4612]: I0227 07:54:52.210631 4612 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-hp72h" podUID="6d061f42-06de-43e5-b4d0-dbca09e48630" containerName="marketplace-operator" containerID="cri-o://33022ff99e1ecbfeccfd65a15ffbd81939f7629ee36cb5c0959129bde7ef1c56" gracePeriod=30 Feb 27 07:54:52 crc kubenswrapper[4612]: I0227 07:54:52.218431 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-w4brj"] Feb 27 07:54:52 crc kubenswrapper[4612]: I0227 07:54:52.218721 4612 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-w4brj" podUID="0bed0a85-ad8c-4670-8193-0e9a90e88d78" containerName="registry-server" containerID="cri-o://d9b35ac517f10a4d23b4a3238292b60f5e90ded7b9e5bd0bb0cecd7ba88e541b" gracePeriod=30 Feb 27 07:54:52 crc kubenswrapper[4612]: I0227 07:54:52.220870 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-fwkxc"] Feb 27 07:54:52 crc kubenswrapper[4612]: I0227 07:54:52.221278 4612 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-fwkxc" podUID="a321b7e1-dbb6-4839-b112-14d7674639c0" containerName="registry-server" containerID="cri-o://5957f89e071a4c4332d217cff26adc7774782198f5c9c6fb6298b97137125ca0" gracePeriod=30 Feb 27 07:54:52 crc kubenswrapper[4612]: I0227 07:54:52.231180 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-88rfh"] Feb 27 07:54:52 crc kubenswrapper[4612]: E0227 07:54:52.231402 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db4d6d91-61f5-4beb-809c-63ae9028a911" containerName="registry" Feb 27 07:54:52 crc kubenswrapper[4612]: I0227 07:54:52.231422 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="db4d6d91-61f5-4beb-809c-63ae9028a911" containerName="registry" Feb 27 07:54:52 crc kubenswrapper[4612]: I0227 07:54:52.231556 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="db4d6d91-61f5-4beb-809c-63ae9028a911" containerName="registry" Feb 27 07:54:52 crc kubenswrapper[4612]: I0227 07:54:52.232053 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-88rfh" Feb 27 07:54:52 crc kubenswrapper[4612]: I0227 07:54:52.253358 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-88rfh"] Feb 27 07:54:52 crc kubenswrapper[4612]: I0227 07:54:52.411431 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/6f0a0683-7e38-410b-8c31-137e9d3a841a-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-88rfh\" (UID: \"6f0a0683-7e38-410b-8c31-137e9d3a841a\") " pod="openshift-marketplace/marketplace-operator-79b997595-88rfh" Feb 27 07:54:52 crc kubenswrapper[4612]: I0227 07:54:52.411472 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/6f0a0683-7e38-410b-8c31-137e9d3a841a-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-88rfh\" (UID: \"6f0a0683-7e38-410b-8c31-137e9d3a841a\") " pod="openshift-marketplace/marketplace-operator-79b997595-88rfh" Feb 27 07:54:52 crc kubenswrapper[4612]: I0227 07:54:52.411498 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hjjzn\" (UniqueName: \"kubernetes.io/projected/6f0a0683-7e38-410b-8c31-137e9d3a841a-kube-api-access-hjjzn\") pod \"marketplace-operator-79b997595-88rfh\" (UID: \"6f0a0683-7e38-410b-8c31-137e9d3a841a\") " pod="openshift-marketplace/marketplace-operator-79b997595-88rfh" Feb 27 07:54:52 crc kubenswrapper[4612]: I0227 07:54:52.512763 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/6f0a0683-7e38-410b-8c31-137e9d3a841a-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-88rfh\" (UID: \"6f0a0683-7e38-410b-8c31-137e9d3a841a\") " pod="openshift-marketplace/marketplace-operator-79b997595-88rfh" Feb 27 07:54:52 crc kubenswrapper[4612]: I0227 07:54:52.512819 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/6f0a0683-7e38-410b-8c31-137e9d3a841a-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-88rfh\" (UID: \"6f0a0683-7e38-410b-8c31-137e9d3a841a\") " pod="openshift-marketplace/marketplace-operator-79b997595-88rfh" Feb 27 07:54:52 crc kubenswrapper[4612]: I0227 07:54:52.512848 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hjjzn\" (UniqueName: \"kubernetes.io/projected/6f0a0683-7e38-410b-8c31-137e9d3a841a-kube-api-access-hjjzn\") pod \"marketplace-operator-79b997595-88rfh\" (UID: \"6f0a0683-7e38-410b-8c31-137e9d3a841a\") " pod="openshift-marketplace/marketplace-operator-79b997595-88rfh" Feb 27 07:54:52 crc kubenswrapper[4612]: I0227 07:54:52.516246 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/6f0a0683-7e38-410b-8c31-137e9d3a841a-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-88rfh\" (UID: \"6f0a0683-7e38-410b-8c31-137e9d3a841a\") " pod="openshift-marketplace/marketplace-operator-79b997595-88rfh" Feb 27 07:54:52 crc kubenswrapper[4612]: I0227 07:54:52.525777 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/6f0a0683-7e38-410b-8c31-137e9d3a841a-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-88rfh\" (UID: \"6f0a0683-7e38-410b-8c31-137e9d3a841a\") " pod="openshift-marketplace/marketplace-operator-79b997595-88rfh" Feb 27 07:54:52 crc kubenswrapper[4612]: I0227 07:54:52.531764 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hjjzn\" (UniqueName: \"kubernetes.io/projected/6f0a0683-7e38-410b-8c31-137e9d3a841a-kube-api-access-hjjzn\") pod \"marketplace-operator-79b997595-88rfh\" (UID: \"6f0a0683-7e38-410b-8c31-137e9d3a841a\") " pod="openshift-marketplace/marketplace-operator-79b997595-88rfh" Feb 27 07:54:52 crc kubenswrapper[4612]: I0227 07:54:52.553607 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-88rfh" Feb 27 07:54:52 crc kubenswrapper[4612]: I0227 07:54:52.573772 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qlhr9" Feb 27 07:54:52 crc kubenswrapper[4612]: I0227 07:54:52.607146 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-p5fp8" Feb 27 07:54:52 crc kubenswrapper[4612]: I0227 07:54:52.717603 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-w4brj" Feb 27 07:54:52 crc kubenswrapper[4612]: I0227 07:54:52.721897 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/201de829-c0b7-4d50-9a3e-ef65d3c06916-catalog-content\") pod \"201de829-c0b7-4d50-9a3e-ef65d3c06916\" (UID: \"201de829-c0b7-4d50-9a3e-ef65d3c06916\") " Feb 27 07:54:52 crc kubenswrapper[4612]: I0227 07:54:52.722081 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/201de829-c0b7-4d50-9a3e-ef65d3c06916-utilities\") pod \"201de829-c0b7-4d50-9a3e-ef65d3c06916\" (UID: \"201de829-c0b7-4d50-9a3e-ef65d3c06916\") " Feb 27 07:54:52 crc kubenswrapper[4612]: I0227 07:54:52.725222 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/201de829-c0b7-4d50-9a3e-ef65d3c06916-utilities" (OuterVolumeSpecName: "utilities") pod "201de829-c0b7-4d50-9a3e-ef65d3c06916" (UID: "201de829-c0b7-4d50-9a3e-ef65d3c06916"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 07:54:52 crc kubenswrapper[4612]: I0227 07:54:52.725693 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qfjdb\" (UniqueName: \"kubernetes.io/projected/7faf5006-23b1-4ef4-ad29-e0e676340a7c-kube-api-access-qfjdb\") pod \"7faf5006-23b1-4ef4-ad29-e0e676340a7c\" (UID: \"7faf5006-23b1-4ef4-ad29-e0e676340a7c\") " Feb 27 07:54:52 crc kubenswrapper[4612]: I0227 07:54:52.725731 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7faf5006-23b1-4ef4-ad29-e0e676340a7c-catalog-content\") pod \"7faf5006-23b1-4ef4-ad29-e0e676340a7c\" (UID: \"7faf5006-23b1-4ef4-ad29-e0e676340a7c\") " Feb 27 07:54:52 crc kubenswrapper[4612]: I0227 07:54:52.725763 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7faf5006-23b1-4ef4-ad29-e0e676340a7c-utilities\") pod \"7faf5006-23b1-4ef4-ad29-e0e676340a7c\" (UID: \"7faf5006-23b1-4ef4-ad29-e0e676340a7c\") " Feb 27 07:54:52 crc kubenswrapper[4612]: I0227 07:54:52.725783 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m629f\" (UniqueName: \"kubernetes.io/projected/201de829-c0b7-4d50-9a3e-ef65d3c06916-kube-api-access-m629f\") pod \"201de829-c0b7-4d50-9a3e-ef65d3c06916\" (UID: \"201de829-c0b7-4d50-9a3e-ef65d3c06916\") " Feb 27 07:54:52 crc kubenswrapper[4612]: I0227 07:54:52.726387 4612 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/201de829-c0b7-4d50-9a3e-ef65d3c06916-utilities\") on node \"crc\" DevicePath \"\"" Feb 27 07:54:52 crc kubenswrapper[4612]: I0227 07:54:52.731989 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/201de829-c0b7-4d50-9a3e-ef65d3c06916-kube-api-access-m629f" (OuterVolumeSpecName: "kube-api-access-m629f") pod "201de829-c0b7-4d50-9a3e-ef65d3c06916" (UID: "201de829-c0b7-4d50-9a3e-ef65d3c06916"). InnerVolumeSpecName "kube-api-access-m629f". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 07:54:52 crc kubenswrapper[4612]: I0227 07:54:52.735844 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7faf5006-23b1-4ef4-ad29-e0e676340a7c-utilities" (OuterVolumeSpecName: "utilities") pod "7faf5006-23b1-4ef4-ad29-e0e676340a7c" (UID: "7faf5006-23b1-4ef4-ad29-e0e676340a7c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 07:54:52 crc kubenswrapper[4612]: I0227 07:54:52.737953 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7faf5006-23b1-4ef4-ad29-e0e676340a7c-kube-api-access-qfjdb" (OuterVolumeSpecName: "kube-api-access-qfjdb") pod "7faf5006-23b1-4ef4-ad29-e0e676340a7c" (UID: "7faf5006-23b1-4ef4-ad29-e0e676340a7c"). InnerVolumeSpecName "kube-api-access-qfjdb". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 07:54:52 crc kubenswrapper[4612]: I0227 07:54:52.740657 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-hp72h" Feb 27 07:54:52 crc kubenswrapper[4612]: I0227 07:54:52.746644 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-fwkxc" Feb 27 07:54:52 crc kubenswrapper[4612]: I0227 07:54:52.810423 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/201de829-c0b7-4d50-9a3e-ef65d3c06916-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "201de829-c0b7-4d50-9a3e-ef65d3c06916" (UID: "201de829-c0b7-4d50-9a3e-ef65d3c06916"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 07:54:52 crc kubenswrapper[4612]: I0227 07:54:52.828107 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cc6gh\" (UniqueName: \"kubernetes.io/projected/0bed0a85-ad8c-4670-8193-0e9a90e88d78-kube-api-access-cc6gh\") pod \"0bed0a85-ad8c-4670-8193-0e9a90e88d78\" (UID: \"0bed0a85-ad8c-4670-8193-0e9a90e88d78\") " Feb 27 07:54:52 crc kubenswrapper[4612]: I0227 07:54:52.828201 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0bed0a85-ad8c-4670-8193-0e9a90e88d78-utilities\") pod \"0bed0a85-ad8c-4670-8193-0e9a90e88d78\" (UID: \"0bed0a85-ad8c-4670-8193-0e9a90e88d78\") " Feb 27 07:54:52 crc kubenswrapper[4612]: I0227 07:54:52.828243 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0bed0a85-ad8c-4670-8193-0e9a90e88d78-catalog-content\") pod \"0bed0a85-ad8c-4670-8193-0e9a90e88d78\" (UID: \"0bed0a85-ad8c-4670-8193-0e9a90e88d78\") " Feb 27 07:54:52 crc kubenswrapper[4612]: I0227 07:54:52.828443 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qfjdb\" (UniqueName: \"kubernetes.io/projected/7faf5006-23b1-4ef4-ad29-e0e676340a7c-kube-api-access-qfjdb\") on node \"crc\" DevicePath \"\"" Feb 27 07:54:52 crc kubenswrapper[4612]: I0227 07:54:52.828461 4612 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7faf5006-23b1-4ef4-ad29-e0e676340a7c-utilities\") on node \"crc\" DevicePath \"\"" Feb 27 07:54:52 crc kubenswrapper[4612]: I0227 07:54:52.828471 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m629f\" (UniqueName: \"kubernetes.io/projected/201de829-c0b7-4d50-9a3e-ef65d3c06916-kube-api-access-m629f\") on node \"crc\" DevicePath \"\"" Feb 27 07:54:52 crc kubenswrapper[4612]: I0227 07:54:52.828481 4612 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/201de829-c0b7-4d50-9a3e-ef65d3c06916-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 27 07:54:52 crc kubenswrapper[4612]: I0227 07:54:52.831746 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0bed0a85-ad8c-4670-8193-0e9a90e88d78-utilities" (OuterVolumeSpecName: "utilities") pod "0bed0a85-ad8c-4670-8193-0e9a90e88d78" (UID: "0bed0a85-ad8c-4670-8193-0e9a90e88d78"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 07:54:52 crc kubenswrapper[4612]: I0227 07:54:52.834148 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0bed0a85-ad8c-4670-8193-0e9a90e88d78-kube-api-access-cc6gh" (OuterVolumeSpecName: "kube-api-access-cc6gh") pod "0bed0a85-ad8c-4670-8193-0e9a90e88d78" (UID: "0bed0a85-ad8c-4670-8193-0e9a90e88d78"). InnerVolumeSpecName "kube-api-access-cc6gh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 07:54:52 crc kubenswrapper[4612]: I0227 07:54:52.842023 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7faf5006-23b1-4ef4-ad29-e0e676340a7c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7faf5006-23b1-4ef4-ad29-e0e676340a7c" (UID: "7faf5006-23b1-4ef4-ad29-e0e676340a7c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 07:54:52 crc kubenswrapper[4612]: I0227 07:54:52.859484 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0bed0a85-ad8c-4670-8193-0e9a90e88d78-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0bed0a85-ad8c-4670-8193-0e9a90e88d78" (UID: "0bed0a85-ad8c-4670-8193-0e9a90e88d78"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 07:54:52 crc kubenswrapper[4612]: I0227 07:54:52.928826 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/6d061f42-06de-43e5-b4d0-dbca09e48630-marketplace-operator-metrics\") pod \"6d061f42-06de-43e5-b4d0-dbca09e48630\" (UID: \"6d061f42-06de-43e5-b4d0-dbca09e48630\") " Feb 27 07:54:52 crc kubenswrapper[4612]: I0227 07:54:52.928865 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ctqts\" (UniqueName: \"kubernetes.io/projected/6d061f42-06de-43e5-b4d0-dbca09e48630-kube-api-access-ctqts\") pod \"6d061f42-06de-43e5-b4d0-dbca09e48630\" (UID: \"6d061f42-06de-43e5-b4d0-dbca09e48630\") " Feb 27 07:54:52 crc kubenswrapper[4612]: I0227 07:54:52.928911 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/6d061f42-06de-43e5-b4d0-dbca09e48630-marketplace-trusted-ca\") pod \"6d061f42-06de-43e5-b4d0-dbca09e48630\" (UID: \"6d061f42-06de-43e5-b4d0-dbca09e48630\") " Feb 27 07:54:52 crc kubenswrapper[4612]: I0227 07:54:52.928944 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a321b7e1-dbb6-4839-b112-14d7674639c0-catalog-content\") pod \"a321b7e1-dbb6-4839-b112-14d7674639c0\" (UID: \"a321b7e1-dbb6-4839-b112-14d7674639c0\") " Feb 27 07:54:52 crc kubenswrapper[4612]: I0227 07:54:52.928972 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5zpf5\" (UniqueName: \"kubernetes.io/projected/a321b7e1-dbb6-4839-b112-14d7674639c0-kube-api-access-5zpf5\") pod \"a321b7e1-dbb6-4839-b112-14d7674639c0\" (UID: \"a321b7e1-dbb6-4839-b112-14d7674639c0\") " Feb 27 07:54:52 crc kubenswrapper[4612]: I0227 07:54:52.928992 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a321b7e1-dbb6-4839-b112-14d7674639c0-utilities\") pod \"a321b7e1-dbb6-4839-b112-14d7674639c0\" (UID: \"a321b7e1-dbb6-4839-b112-14d7674639c0\") " Feb 27 07:54:52 crc kubenswrapper[4612]: I0227 07:54:52.929163 4612 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7faf5006-23b1-4ef4-ad29-e0e676340a7c-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 27 07:54:52 crc kubenswrapper[4612]: I0227 07:54:52.929179 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cc6gh\" (UniqueName: \"kubernetes.io/projected/0bed0a85-ad8c-4670-8193-0e9a90e88d78-kube-api-access-cc6gh\") on node \"crc\" DevicePath \"\"" Feb 27 07:54:52 crc kubenswrapper[4612]: I0227 07:54:52.929190 4612 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0bed0a85-ad8c-4670-8193-0e9a90e88d78-utilities\") on node \"crc\" DevicePath \"\"" Feb 27 07:54:52 crc kubenswrapper[4612]: I0227 07:54:52.929199 4612 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0bed0a85-ad8c-4670-8193-0e9a90e88d78-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 27 07:54:52 crc kubenswrapper[4612]: I0227 07:54:52.929947 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a321b7e1-dbb6-4839-b112-14d7674639c0-utilities" (OuterVolumeSpecName: "utilities") pod "a321b7e1-dbb6-4839-b112-14d7674639c0" (UID: "a321b7e1-dbb6-4839-b112-14d7674639c0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 07:54:52 crc kubenswrapper[4612]: I0227 07:54:52.930948 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6d061f42-06de-43e5-b4d0-dbca09e48630-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "6d061f42-06de-43e5-b4d0-dbca09e48630" (UID: "6d061f42-06de-43e5-b4d0-dbca09e48630"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 07:54:52 crc kubenswrapper[4612]: I0227 07:54:52.933036 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6d061f42-06de-43e5-b4d0-dbca09e48630-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "6d061f42-06de-43e5-b4d0-dbca09e48630" (UID: "6d061f42-06de-43e5-b4d0-dbca09e48630"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 07:54:52 crc kubenswrapper[4612]: I0227 07:54:52.933082 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6d061f42-06de-43e5-b4d0-dbca09e48630-kube-api-access-ctqts" (OuterVolumeSpecName: "kube-api-access-ctqts") pod "6d061f42-06de-43e5-b4d0-dbca09e48630" (UID: "6d061f42-06de-43e5-b4d0-dbca09e48630"). InnerVolumeSpecName "kube-api-access-ctqts". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 07:54:52 crc kubenswrapper[4612]: I0227 07:54:52.933330 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a321b7e1-dbb6-4839-b112-14d7674639c0-kube-api-access-5zpf5" (OuterVolumeSpecName: "kube-api-access-5zpf5") pod "a321b7e1-dbb6-4839-b112-14d7674639c0" (UID: "a321b7e1-dbb6-4839-b112-14d7674639c0"). InnerVolumeSpecName "kube-api-access-5zpf5". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 07:54:53 crc kubenswrapper[4612]: I0227 07:54:53.016623 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-88rfh"] Feb 27 07:54:53 crc kubenswrapper[4612]: I0227 07:54:53.030826 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5zpf5\" (UniqueName: \"kubernetes.io/projected/a321b7e1-dbb6-4839-b112-14d7674639c0-kube-api-access-5zpf5\") on node \"crc\" DevicePath \"\"" Feb 27 07:54:53 crc kubenswrapper[4612]: I0227 07:54:53.030864 4612 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a321b7e1-dbb6-4839-b112-14d7674639c0-utilities\") on node \"crc\" DevicePath \"\"" Feb 27 07:54:53 crc kubenswrapper[4612]: I0227 07:54:53.030879 4612 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/6d061f42-06de-43e5-b4d0-dbca09e48630-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Feb 27 07:54:53 crc kubenswrapper[4612]: I0227 07:54:53.030889 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ctqts\" (UniqueName: \"kubernetes.io/projected/6d061f42-06de-43e5-b4d0-dbca09e48630-kube-api-access-ctqts\") on node \"crc\" DevicePath \"\"" Feb 27 07:54:53 crc kubenswrapper[4612]: I0227 07:54:53.030899 4612 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/6d061f42-06de-43e5-b4d0-dbca09e48630-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Feb 27 07:54:53 crc kubenswrapper[4612]: I0227 07:54:53.072985 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a321b7e1-dbb6-4839-b112-14d7674639c0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a321b7e1-dbb6-4839-b112-14d7674639c0" (UID: "a321b7e1-dbb6-4839-b112-14d7674639c0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 07:54:53 crc kubenswrapper[4612]: I0227 07:54:53.131969 4612 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a321b7e1-dbb6-4839-b112-14d7674639c0-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 27 07:54:53 crc kubenswrapper[4612]: I0227 07:54:53.251596 4612 generic.go:334] "Generic (PLEG): container finished" podID="6d061f42-06de-43e5-b4d0-dbca09e48630" containerID="33022ff99e1ecbfeccfd65a15ffbd81939f7629ee36cb5c0959129bde7ef1c56" exitCode=0 Feb 27 07:54:53 crc kubenswrapper[4612]: I0227 07:54:53.251669 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-hp72h" event={"ID":"6d061f42-06de-43e5-b4d0-dbca09e48630","Type":"ContainerDied","Data":"33022ff99e1ecbfeccfd65a15ffbd81939f7629ee36cb5c0959129bde7ef1c56"} Feb 27 07:54:53 crc kubenswrapper[4612]: I0227 07:54:53.251710 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-hp72h" event={"ID":"6d061f42-06de-43e5-b4d0-dbca09e48630","Type":"ContainerDied","Data":"b00ebd717acd7c4583cb33e8ef55fa92362ffe216db3eda89fe149afef661031"} Feb 27 07:54:53 crc kubenswrapper[4612]: I0227 07:54:53.251728 4612 scope.go:117] "RemoveContainer" containerID="33022ff99e1ecbfeccfd65a15ffbd81939f7629ee36cb5c0959129bde7ef1c56" Feb 27 07:54:53 crc kubenswrapper[4612]: I0227 07:54:53.251842 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-hp72h" Feb 27 07:54:53 crc kubenswrapper[4612]: I0227 07:54:53.268797 4612 scope.go:117] "RemoveContainer" containerID="872f6d6ba2cff41e30dff82c7bfa288c746d738f5f9784ceca408e98c3730656" Feb 27 07:54:53 crc kubenswrapper[4612]: I0227 07:54:53.268934 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-88rfh" event={"ID":"6f0a0683-7e38-410b-8c31-137e9d3a841a","Type":"ContainerStarted","Data":"fcd79176685f6ce40788da798152532715c779ce55f45beeb04cbb677fe75c37"} Feb 27 07:54:53 crc kubenswrapper[4612]: I0227 07:54:53.276407 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-hp72h"] Feb 27 07:54:53 crc kubenswrapper[4612]: I0227 07:54:53.279857 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-hp72h"] Feb 27 07:54:53 crc kubenswrapper[4612]: I0227 07:54:53.283622 4612 generic.go:334] "Generic (PLEG): container finished" podID="201de829-c0b7-4d50-9a3e-ef65d3c06916" containerID="a93c39714c2e5eda46c61ea58eae7a9e3f1ba38218d4c700a99e6cc7b1c55c31" exitCode=0 Feb 27 07:54:53 crc kubenswrapper[4612]: I0227 07:54:53.283671 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p5fp8" event={"ID":"201de829-c0b7-4d50-9a3e-ef65d3c06916","Type":"ContainerDied","Data":"a93c39714c2e5eda46c61ea58eae7a9e3f1ba38218d4c700a99e6cc7b1c55c31"} Feb 27 07:54:53 crc kubenswrapper[4612]: I0227 07:54:53.283704 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p5fp8" event={"ID":"201de829-c0b7-4d50-9a3e-ef65d3c06916","Type":"ContainerDied","Data":"0521755a1fd13d9dfd37e8c466d6d6f38844dd6dff978b91c3083d738fc1e1a4"} Feb 27 07:54:53 crc kubenswrapper[4612]: I0227 07:54:53.283762 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-p5fp8" Feb 27 07:54:53 crc kubenswrapper[4612]: I0227 07:54:53.286748 4612 generic.go:334] "Generic (PLEG): container finished" podID="a321b7e1-dbb6-4839-b112-14d7674639c0" containerID="5957f89e071a4c4332d217cff26adc7774782198f5c9c6fb6298b97137125ca0" exitCode=0 Feb 27 07:54:53 crc kubenswrapper[4612]: I0227 07:54:53.286799 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fwkxc" event={"ID":"a321b7e1-dbb6-4839-b112-14d7674639c0","Type":"ContainerDied","Data":"5957f89e071a4c4332d217cff26adc7774782198f5c9c6fb6298b97137125ca0"} Feb 27 07:54:53 crc kubenswrapper[4612]: I0227 07:54:53.286819 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fwkxc" event={"ID":"a321b7e1-dbb6-4839-b112-14d7674639c0","Type":"ContainerDied","Data":"6c30c404089e69affb635ab954734cef84c8d90b79d3e9221b069d1534b0cf54"} Feb 27 07:54:53 crc kubenswrapper[4612]: I0227 07:54:53.286879 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-fwkxc" Feb 27 07:54:53 crc kubenswrapper[4612]: I0227 07:54:53.303397 4612 generic.go:334] "Generic (PLEG): container finished" podID="0bed0a85-ad8c-4670-8193-0e9a90e88d78" containerID="d9b35ac517f10a4d23b4a3238292b60f5e90ded7b9e5bd0bb0cecd7ba88e541b" exitCode=0 Feb 27 07:54:53 crc kubenswrapper[4612]: I0227 07:54:53.303974 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-w4brj" event={"ID":"0bed0a85-ad8c-4670-8193-0e9a90e88d78","Type":"ContainerDied","Data":"d9b35ac517f10a4d23b4a3238292b60f5e90ded7b9e5bd0bb0cecd7ba88e541b"} Feb 27 07:54:53 crc kubenswrapper[4612]: I0227 07:54:53.304721 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-w4brj" event={"ID":"0bed0a85-ad8c-4670-8193-0e9a90e88d78","Type":"ContainerDied","Data":"1e00925de47b036507b4e069657a360fb4493d49c80db12572efba3e2366c650"} Feb 27 07:54:53 crc kubenswrapper[4612]: I0227 07:54:53.305801 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-w4brj" Feb 27 07:54:53 crc kubenswrapper[4612]: I0227 07:54:53.313073 4612 generic.go:334] "Generic (PLEG): container finished" podID="7faf5006-23b1-4ef4-ad29-e0e676340a7c" containerID="bc3d2a71143833b070c3078c2c2694503dc694341f36a9dad42134afa1a0c8c9" exitCode=0 Feb 27 07:54:53 crc kubenswrapper[4612]: I0227 07:54:53.313373 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qlhr9" event={"ID":"7faf5006-23b1-4ef4-ad29-e0e676340a7c","Type":"ContainerDied","Data":"bc3d2a71143833b070c3078c2c2694503dc694341f36a9dad42134afa1a0c8c9"} Feb 27 07:54:53 crc kubenswrapper[4612]: I0227 07:54:53.313408 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-qlhr9" event={"ID":"7faf5006-23b1-4ef4-ad29-e0e676340a7c","Type":"ContainerDied","Data":"3909413230e119c86dc9abe3e069de2ff1342a9e4a7331713645d53104392a1a"} Feb 27 07:54:53 crc kubenswrapper[4612]: I0227 07:54:53.313480 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-qlhr9" Feb 27 07:54:53 crc kubenswrapper[4612]: I0227 07:54:53.319445 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-p5fp8"] Feb 27 07:54:53 crc kubenswrapper[4612]: I0227 07:54:53.330999 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-p5fp8"] Feb 27 07:54:53 crc kubenswrapper[4612]: I0227 07:54:53.334638 4612 scope.go:117] "RemoveContainer" containerID="33022ff99e1ecbfeccfd65a15ffbd81939f7629ee36cb5c0959129bde7ef1c56" Feb 27 07:54:53 crc kubenswrapper[4612]: E0227 07:54:53.335130 4612 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"33022ff99e1ecbfeccfd65a15ffbd81939f7629ee36cb5c0959129bde7ef1c56\": container with ID starting with 33022ff99e1ecbfeccfd65a15ffbd81939f7629ee36cb5c0959129bde7ef1c56 not found: ID does not exist" containerID="33022ff99e1ecbfeccfd65a15ffbd81939f7629ee36cb5c0959129bde7ef1c56" Feb 27 07:54:53 crc kubenswrapper[4612]: I0227 07:54:53.335176 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"33022ff99e1ecbfeccfd65a15ffbd81939f7629ee36cb5c0959129bde7ef1c56"} err="failed to get container status \"33022ff99e1ecbfeccfd65a15ffbd81939f7629ee36cb5c0959129bde7ef1c56\": rpc error: code = NotFound desc = could not find container \"33022ff99e1ecbfeccfd65a15ffbd81939f7629ee36cb5c0959129bde7ef1c56\": container with ID starting with 33022ff99e1ecbfeccfd65a15ffbd81939f7629ee36cb5c0959129bde7ef1c56 not found: ID does not exist" Feb 27 07:54:53 crc kubenswrapper[4612]: I0227 07:54:53.335210 4612 scope.go:117] "RemoveContainer" containerID="872f6d6ba2cff41e30dff82c7bfa288c746d738f5f9784ceca408e98c3730656" Feb 27 07:54:53 crc kubenswrapper[4612]: E0227 07:54:53.335522 4612 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"872f6d6ba2cff41e30dff82c7bfa288c746d738f5f9784ceca408e98c3730656\": container with ID starting with 872f6d6ba2cff41e30dff82c7bfa288c746d738f5f9784ceca408e98c3730656 not found: ID does not exist" containerID="872f6d6ba2cff41e30dff82c7bfa288c746d738f5f9784ceca408e98c3730656" Feb 27 07:54:53 crc kubenswrapper[4612]: I0227 07:54:53.335565 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"872f6d6ba2cff41e30dff82c7bfa288c746d738f5f9784ceca408e98c3730656"} err="failed to get container status \"872f6d6ba2cff41e30dff82c7bfa288c746d738f5f9784ceca408e98c3730656\": rpc error: code = NotFound desc = could not find container \"872f6d6ba2cff41e30dff82c7bfa288c746d738f5f9784ceca408e98c3730656\": container with ID starting with 872f6d6ba2cff41e30dff82c7bfa288c746d738f5f9784ceca408e98c3730656 not found: ID does not exist" Feb 27 07:54:53 crc kubenswrapper[4612]: I0227 07:54:53.335592 4612 scope.go:117] "RemoveContainer" containerID="a93c39714c2e5eda46c61ea58eae7a9e3f1ba38218d4c700a99e6cc7b1c55c31" Feb 27 07:54:53 crc kubenswrapper[4612]: I0227 07:54:53.369212 4612 scope.go:117] "RemoveContainer" containerID="e7d182d8908ed3789a5ec0e42333b53967e23f78c8f66f9498c6a40440374b5a" Feb 27 07:54:53 crc kubenswrapper[4612]: I0227 07:54:53.377749 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-w4brj"] Feb 27 07:54:53 crc kubenswrapper[4612]: I0227 07:54:53.384778 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-w4brj"] Feb 27 07:54:53 crc kubenswrapper[4612]: I0227 07:54:53.384806 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-fwkxc"] Feb 27 07:54:53 crc kubenswrapper[4612]: I0227 07:54:53.391522 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-fwkxc"] Feb 27 07:54:53 crc kubenswrapper[4612]: I0227 07:54:53.406585 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-qlhr9"] Feb 27 07:54:53 crc kubenswrapper[4612]: I0227 07:54:53.407589 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-qlhr9"] Feb 27 07:54:53 crc kubenswrapper[4612]: I0227 07:54:53.414068 4612 scope.go:117] "RemoveContainer" containerID="c3a1bd90b4e84308b9a8a12e1bcd3c87b094f875095f113890440ccd5ae909ad" Feb 27 07:54:53 crc kubenswrapper[4612]: I0227 07:54:53.438294 4612 scope.go:117] "RemoveContainer" containerID="a93c39714c2e5eda46c61ea58eae7a9e3f1ba38218d4c700a99e6cc7b1c55c31" Feb 27 07:54:53 crc kubenswrapper[4612]: E0227 07:54:53.438821 4612 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a93c39714c2e5eda46c61ea58eae7a9e3f1ba38218d4c700a99e6cc7b1c55c31\": container with ID starting with a93c39714c2e5eda46c61ea58eae7a9e3f1ba38218d4c700a99e6cc7b1c55c31 not found: ID does not exist" containerID="a93c39714c2e5eda46c61ea58eae7a9e3f1ba38218d4c700a99e6cc7b1c55c31" Feb 27 07:54:53 crc kubenswrapper[4612]: I0227 07:54:53.438875 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a93c39714c2e5eda46c61ea58eae7a9e3f1ba38218d4c700a99e6cc7b1c55c31"} err="failed to get container status \"a93c39714c2e5eda46c61ea58eae7a9e3f1ba38218d4c700a99e6cc7b1c55c31\": rpc error: code = NotFound desc = could not find container \"a93c39714c2e5eda46c61ea58eae7a9e3f1ba38218d4c700a99e6cc7b1c55c31\": container with ID starting with a93c39714c2e5eda46c61ea58eae7a9e3f1ba38218d4c700a99e6cc7b1c55c31 not found: ID does not exist" Feb 27 07:54:53 crc kubenswrapper[4612]: I0227 07:54:53.438903 4612 scope.go:117] "RemoveContainer" containerID="e7d182d8908ed3789a5ec0e42333b53967e23f78c8f66f9498c6a40440374b5a" Feb 27 07:54:53 crc kubenswrapper[4612]: E0227 07:54:53.442139 4612 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e7d182d8908ed3789a5ec0e42333b53967e23f78c8f66f9498c6a40440374b5a\": container with ID starting with e7d182d8908ed3789a5ec0e42333b53967e23f78c8f66f9498c6a40440374b5a not found: ID does not exist" containerID="e7d182d8908ed3789a5ec0e42333b53967e23f78c8f66f9498c6a40440374b5a" Feb 27 07:54:53 crc kubenswrapper[4612]: I0227 07:54:53.442179 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e7d182d8908ed3789a5ec0e42333b53967e23f78c8f66f9498c6a40440374b5a"} err="failed to get container status \"e7d182d8908ed3789a5ec0e42333b53967e23f78c8f66f9498c6a40440374b5a\": rpc error: code = NotFound desc = could not find container \"e7d182d8908ed3789a5ec0e42333b53967e23f78c8f66f9498c6a40440374b5a\": container with ID starting with e7d182d8908ed3789a5ec0e42333b53967e23f78c8f66f9498c6a40440374b5a not found: ID does not exist" Feb 27 07:54:53 crc kubenswrapper[4612]: I0227 07:54:53.442206 4612 scope.go:117] "RemoveContainer" containerID="c3a1bd90b4e84308b9a8a12e1bcd3c87b094f875095f113890440ccd5ae909ad" Feb 27 07:54:53 crc kubenswrapper[4612]: E0227 07:54:53.446073 4612 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c3a1bd90b4e84308b9a8a12e1bcd3c87b094f875095f113890440ccd5ae909ad\": container with ID starting with c3a1bd90b4e84308b9a8a12e1bcd3c87b094f875095f113890440ccd5ae909ad not found: ID does not exist" containerID="c3a1bd90b4e84308b9a8a12e1bcd3c87b094f875095f113890440ccd5ae909ad" Feb 27 07:54:53 crc kubenswrapper[4612]: I0227 07:54:53.446101 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c3a1bd90b4e84308b9a8a12e1bcd3c87b094f875095f113890440ccd5ae909ad"} err="failed to get container status \"c3a1bd90b4e84308b9a8a12e1bcd3c87b094f875095f113890440ccd5ae909ad\": rpc error: code = NotFound desc = could not find container \"c3a1bd90b4e84308b9a8a12e1bcd3c87b094f875095f113890440ccd5ae909ad\": container with ID starting with c3a1bd90b4e84308b9a8a12e1bcd3c87b094f875095f113890440ccd5ae909ad not found: ID does not exist" Feb 27 07:54:53 crc kubenswrapper[4612]: I0227 07:54:53.446117 4612 scope.go:117] "RemoveContainer" containerID="5957f89e071a4c4332d217cff26adc7774782198f5c9c6fb6298b97137125ca0" Feb 27 07:54:53 crc kubenswrapper[4612]: I0227 07:54:53.474586 4612 scope.go:117] "RemoveContainer" containerID="07539f761b9f4d1ae3d06f040cbd6f19e2153f9ebf9c0cfd92de620b1132d10d" Feb 27 07:54:53 crc kubenswrapper[4612]: I0227 07:54:53.490076 4612 scope.go:117] "RemoveContainer" containerID="041be02357fc54b1f0d14a282c3a15a136c5b290322b49ee5403ee58e1b3e221" Feb 27 07:54:53 crc kubenswrapper[4612]: I0227 07:54:53.506531 4612 scope.go:117] "RemoveContainer" containerID="5957f89e071a4c4332d217cff26adc7774782198f5c9c6fb6298b97137125ca0" Feb 27 07:54:53 crc kubenswrapper[4612]: E0227 07:54:53.506930 4612 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5957f89e071a4c4332d217cff26adc7774782198f5c9c6fb6298b97137125ca0\": container with ID starting with 5957f89e071a4c4332d217cff26adc7774782198f5c9c6fb6298b97137125ca0 not found: ID does not exist" containerID="5957f89e071a4c4332d217cff26adc7774782198f5c9c6fb6298b97137125ca0" Feb 27 07:54:53 crc kubenswrapper[4612]: I0227 07:54:53.506969 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5957f89e071a4c4332d217cff26adc7774782198f5c9c6fb6298b97137125ca0"} err="failed to get container status \"5957f89e071a4c4332d217cff26adc7774782198f5c9c6fb6298b97137125ca0\": rpc error: code = NotFound desc = could not find container \"5957f89e071a4c4332d217cff26adc7774782198f5c9c6fb6298b97137125ca0\": container with ID starting with 5957f89e071a4c4332d217cff26adc7774782198f5c9c6fb6298b97137125ca0 not found: ID does not exist" Feb 27 07:54:53 crc kubenswrapper[4612]: I0227 07:54:53.506998 4612 scope.go:117] "RemoveContainer" containerID="07539f761b9f4d1ae3d06f040cbd6f19e2153f9ebf9c0cfd92de620b1132d10d" Feb 27 07:54:53 crc kubenswrapper[4612]: E0227 07:54:53.507270 4612 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"07539f761b9f4d1ae3d06f040cbd6f19e2153f9ebf9c0cfd92de620b1132d10d\": container with ID starting with 07539f761b9f4d1ae3d06f040cbd6f19e2153f9ebf9c0cfd92de620b1132d10d not found: ID does not exist" containerID="07539f761b9f4d1ae3d06f040cbd6f19e2153f9ebf9c0cfd92de620b1132d10d" Feb 27 07:54:53 crc kubenswrapper[4612]: I0227 07:54:53.507293 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"07539f761b9f4d1ae3d06f040cbd6f19e2153f9ebf9c0cfd92de620b1132d10d"} err="failed to get container status \"07539f761b9f4d1ae3d06f040cbd6f19e2153f9ebf9c0cfd92de620b1132d10d\": rpc error: code = NotFound desc = could not find container \"07539f761b9f4d1ae3d06f040cbd6f19e2153f9ebf9c0cfd92de620b1132d10d\": container with ID starting with 07539f761b9f4d1ae3d06f040cbd6f19e2153f9ebf9c0cfd92de620b1132d10d not found: ID does not exist" Feb 27 07:54:53 crc kubenswrapper[4612]: I0227 07:54:53.507306 4612 scope.go:117] "RemoveContainer" containerID="041be02357fc54b1f0d14a282c3a15a136c5b290322b49ee5403ee58e1b3e221" Feb 27 07:54:53 crc kubenswrapper[4612]: E0227 07:54:53.507527 4612 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"041be02357fc54b1f0d14a282c3a15a136c5b290322b49ee5403ee58e1b3e221\": container with ID starting with 041be02357fc54b1f0d14a282c3a15a136c5b290322b49ee5403ee58e1b3e221 not found: ID does not exist" containerID="041be02357fc54b1f0d14a282c3a15a136c5b290322b49ee5403ee58e1b3e221" Feb 27 07:54:53 crc kubenswrapper[4612]: I0227 07:54:53.507547 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"041be02357fc54b1f0d14a282c3a15a136c5b290322b49ee5403ee58e1b3e221"} err="failed to get container status \"041be02357fc54b1f0d14a282c3a15a136c5b290322b49ee5403ee58e1b3e221\": rpc error: code = NotFound desc = could not find container \"041be02357fc54b1f0d14a282c3a15a136c5b290322b49ee5403ee58e1b3e221\": container with ID starting with 041be02357fc54b1f0d14a282c3a15a136c5b290322b49ee5403ee58e1b3e221 not found: ID does not exist" Feb 27 07:54:53 crc kubenswrapper[4612]: I0227 07:54:53.507566 4612 scope.go:117] "RemoveContainer" containerID="d9b35ac517f10a4d23b4a3238292b60f5e90ded7b9e5bd0bb0cecd7ba88e541b" Feb 27 07:54:53 crc kubenswrapper[4612]: I0227 07:54:53.525479 4612 scope.go:117] "RemoveContainer" containerID="db569021060effab133a7d67b77b4bf9c38c1f256f73e7d7ec963c63bd23482d" Feb 27 07:54:53 crc kubenswrapper[4612]: I0227 07:54:53.537984 4612 scope.go:117] "RemoveContainer" containerID="a4f3f08c42a2b8f5989c0e5f57f73f47697425566b9354deab7746a5d4af3958" Feb 27 07:54:53 crc kubenswrapper[4612]: I0227 07:54:53.552303 4612 scope.go:117] "RemoveContainer" containerID="d9b35ac517f10a4d23b4a3238292b60f5e90ded7b9e5bd0bb0cecd7ba88e541b" Feb 27 07:54:53 crc kubenswrapper[4612]: E0227 07:54:53.552656 4612 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d9b35ac517f10a4d23b4a3238292b60f5e90ded7b9e5bd0bb0cecd7ba88e541b\": container with ID starting with d9b35ac517f10a4d23b4a3238292b60f5e90ded7b9e5bd0bb0cecd7ba88e541b not found: ID does not exist" containerID="d9b35ac517f10a4d23b4a3238292b60f5e90ded7b9e5bd0bb0cecd7ba88e541b" Feb 27 07:54:53 crc kubenswrapper[4612]: I0227 07:54:53.552700 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d9b35ac517f10a4d23b4a3238292b60f5e90ded7b9e5bd0bb0cecd7ba88e541b"} err="failed to get container status \"d9b35ac517f10a4d23b4a3238292b60f5e90ded7b9e5bd0bb0cecd7ba88e541b\": rpc error: code = NotFound desc = could not find container \"d9b35ac517f10a4d23b4a3238292b60f5e90ded7b9e5bd0bb0cecd7ba88e541b\": container with ID starting with d9b35ac517f10a4d23b4a3238292b60f5e90ded7b9e5bd0bb0cecd7ba88e541b not found: ID does not exist" Feb 27 07:54:53 crc kubenswrapper[4612]: I0227 07:54:53.552749 4612 scope.go:117] "RemoveContainer" containerID="db569021060effab133a7d67b77b4bf9c38c1f256f73e7d7ec963c63bd23482d" Feb 27 07:54:53 crc kubenswrapper[4612]: E0227 07:54:53.552950 4612 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"db569021060effab133a7d67b77b4bf9c38c1f256f73e7d7ec963c63bd23482d\": container with ID starting with db569021060effab133a7d67b77b4bf9c38c1f256f73e7d7ec963c63bd23482d not found: ID does not exist" containerID="db569021060effab133a7d67b77b4bf9c38c1f256f73e7d7ec963c63bd23482d" Feb 27 07:54:53 crc kubenswrapper[4612]: I0227 07:54:53.552981 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"db569021060effab133a7d67b77b4bf9c38c1f256f73e7d7ec963c63bd23482d"} err="failed to get container status \"db569021060effab133a7d67b77b4bf9c38c1f256f73e7d7ec963c63bd23482d\": rpc error: code = NotFound desc = could not find container \"db569021060effab133a7d67b77b4bf9c38c1f256f73e7d7ec963c63bd23482d\": container with ID starting with db569021060effab133a7d67b77b4bf9c38c1f256f73e7d7ec963c63bd23482d not found: ID does not exist" Feb 27 07:54:53 crc kubenswrapper[4612]: I0227 07:54:53.552999 4612 scope.go:117] "RemoveContainer" containerID="a4f3f08c42a2b8f5989c0e5f57f73f47697425566b9354deab7746a5d4af3958" Feb 27 07:54:53 crc kubenswrapper[4612]: E0227 07:54:53.553172 4612 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a4f3f08c42a2b8f5989c0e5f57f73f47697425566b9354deab7746a5d4af3958\": container with ID starting with a4f3f08c42a2b8f5989c0e5f57f73f47697425566b9354deab7746a5d4af3958 not found: ID does not exist" containerID="a4f3f08c42a2b8f5989c0e5f57f73f47697425566b9354deab7746a5d4af3958" Feb 27 07:54:53 crc kubenswrapper[4612]: I0227 07:54:53.553195 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a4f3f08c42a2b8f5989c0e5f57f73f47697425566b9354deab7746a5d4af3958"} err="failed to get container status \"a4f3f08c42a2b8f5989c0e5f57f73f47697425566b9354deab7746a5d4af3958\": rpc error: code = NotFound desc = could not find container \"a4f3f08c42a2b8f5989c0e5f57f73f47697425566b9354deab7746a5d4af3958\": container with ID starting with a4f3f08c42a2b8f5989c0e5f57f73f47697425566b9354deab7746a5d4af3958 not found: ID does not exist" Feb 27 07:54:53 crc kubenswrapper[4612]: I0227 07:54:53.553211 4612 scope.go:117] "RemoveContainer" containerID="bc3d2a71143833b070c3078c2c2694503dc694341f36a9dad42134afa1a0c8c9" Feb 27 07:54:53 crc kubenswrapper[4612]: I0227 07:54:53.566935 4612 scope.go:117] "RemoveContainer" containerID="a2e0af366c381d79253fa4c1ba8882fa2b8590183c06bfe972f9171eca234e7a" Feb 27 07:54:53 crc kubenswrapper[4612]: I0227 07:54:53.580091 4612 scope.go:117] "RemoveContainer" containerID="2fb6c4cf5724b631a0b34f2088cfbfa5e0a7c694db0d32f4f330dca0cc39c7a2" Feb 27 07:54:53 crc kubenswrapper[4612]: I0227 07:54:53.591980 4612 scope.go:117] "RemoveContainer" containerID="bc3d2a71143833b070c3078c2c2694503dc694341f36a9dad42134afa1a0c8c9" Feb 27 07:54:53 crc kubenswrapper[4612]: E0227 07:54:53.592446 4612 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bc3d2a71143833b070c3078c2c2694503dc694341f36a9dad42134afa1a0c8c9\": container with ID starting with bc3d2a71143833b070c3078c2c2694503dc694341f36a9dad42134afa1a0c8c9 not found: ID does not exist" containerID="bc3d2a71143833b070c3078c2c2694503dc694341f36a9dad42134afa1a0c8c9" Feb 27 07:54:53 crc kubenswrapper[4612]: I0227 07:54:53.592483 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bc3d2a71143833b070c3078c2c2694503dc694341f36a9dad42134afa1a0c8c9"} err="failed to get container status \"bc3d2a71143833b070c3078c2c2694503dc694341f36a9dad42134afa1a0c8c9\": rpc error: code = NotFound desc = could not find container \"bc3d2a71143833b070c3078c2c2694503dc694341f36a9dad42134afa1a0c8c9\": container with ID starting with bc3d2a71143833b070c3078c2c2694503dc694341f36a9dad42134afa1a0c8c9 not found: ID does not exist" Feb 27 07:54:53 crc kubenswrapper[4612]: I0227 07:54:53.593367 4612 scope.go:117] "RemoveContainer" containerID="a2e0af366c381d79253fa4c1ba8882fa2b8590183c06bfe972f9171eca234e7a" Feb 27 07:54:53 crc kubenswrapper[4612]: E0227 07:54:53.593893 4612 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a2e0af366c381d79253fa4c1ba8882fa2b8590183c06bfe972f9171eca234e7a\": container with ID starting with a2e0af366c381d79253fa4c1ba8882fa2b8590183c06bfe972f9171eca234e7a not found: ID does not exist" containerID="a2e0af366c381d79253fa4c1ba8882fa2b8590183c06bfe972f9171eca234e7a" Feb 27 07:54:53 crc kubenswrapper[4612]: I0227 07:54:53.593921 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a2e0af366c381d79253fa4c1ba8882fa2b8590183c06bfe972f9171eca234e7a"} err="failed to get container status \"a2e0af366c381d79253fa4c1ba8882fa2b8590183c06bfe972f9171eca234e7a\": rpc error: code = NotFound desc = could not find container \"a2e0af366c381d79253fa4c1ba8882fa2b8590183c06bfe972f9171eca234e7a\": container with ID starting with a2e0af366c381d79253fa4c1ba8882fa2b8590183c06bfe972f9171eca234e7a not found: ID does not exist" Feb 27 07:54:53 crc kubenswrapper[4612]: I0227 07:54:53.593943 4612 scope.go:117] "RemoveContainer" containerID="2fb6c4cf5724b631a0b34f2088cfbfa5e0a7c694db0d32f4f330dca0cc39c7a2" Feb 27 07:54:53 crc kubenswrapper[4612]: E0227 07:54:53.594171 4612 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2fb6c4cf5724b631a0b34f2088cfbfa5e0a7c694db0d32f4f330dca0cc39c7a2\": container with ID starting with 2fb6c4cf5724b631a0b34f2088cfbfa5e0a7c694db0d32f4f330dca0cc39c7a2 not found: ID does not exist" containerID="2fb6c4cf5724b631a0b34f2088cfbfa5e0a7c694db0d32f4f330dca0cc39c7a2" Feb 27 07:54:53 crc kubenswrapper[4612]: I0227 07:54:53.594195 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2fb6c4cf5724b631a0b34f2088cfbfa5e0a7c694db0d32f4f330dca0cc39c7a2"} err="failed to get container status \"2fb6c4cf5724b631a0b34f2088cfbfa5e0a7c694db0d32f4f330dca0cc39c7a2\": rpc error: code = NotFound desc = could not find container \"2fb6c4cf5724b631a0b34f2088cfbfa5e0a7c694db0d32f4f330dca0cc39c7a2\": container with ID starting with 2fb6c4cf5724b631a0b34f2088cfbfa5e0a7c694db0d32f4f330dca0cc39c7a2 not found: ID does not exist" Feb 27 07:54:54 crc kubenswrapper[4612]: I0227 07:54:54.332130 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-88rfh" event={"ID":"6f0a0683-7e38-410b-8c31-137e9d3a841a","Type":"ContainerStarted","Data":"0a521d98a83bc76e73d78bb2ed390913ec039227a0a44eb476cc08583a65ddff"} Feb 27 07:54:54 crc kubenswrapper[4612]: I0227 07:54:54.332793 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-88rfh" Feb 27 07:54:54 crc kubenswrapper[4612]: I0227 07:54:54.336420 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-88rfh" Feb 27 07:54:54 crc kubenswrapper[4612]: I0227 07:54:54.359367 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-88rfh" podStartSLOduration=2.359347728 podStartE2EDuration="2.359347728s" podCreationTimestamp="2026-02-27 07:54:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 07:54:54.351183303 +0000 UTC m=+352.205113341" watchObservedRunningTime="2026-02-27 07:54:54.359347728 +0000 UTC m=+352.213277726" Feb 27 07:54:54 crc kubenswrapper[4612]: I0227 07:54:54.443720 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-8dbvm"] Feb 27 07:54:54 crc kubenswrapper[4612]: E0227 07:54:54.444311 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7faf5006-23b1-4ef4-ad29-e0e676340a7c" containerName="extract-utilities" Feb 27 07:54:54 crc kubenswrapper[4612]: I0227 07:54:54.444336 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="7faf5006-23b1-4ef4-ad29-e0e676340a7c" containerName="extract-utilities" Feb 27 07:54:54 crc kubenswrapper[4612]: E0227 07:54:54.444387 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a321b7e1-dbb6-4839-b112-14d7674639c0" containerName="registry-server" Feb 27 07:54:54 crc kubenswrapper[4612]: I0227 07:54:54.444400 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="a321b7e1-dbb6-4839-b112-14d7674639c0" containerName="registry-server" Feb 27 07:54:54 crc kubenswrapper[4612]: E0227 07:54:54.444447 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0bed0a85-ad8c-4670-8193-0e9a90e88d78" containerName="extract-utilities" Feb 27 07:54:54 crc kubenswrapper[4612]: I0227 07:54:54.444458 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="0bed0a85-ad8c-4670-8193-0e9a90e88d78" containerName="extract-utilities" Feb 27 07:54:54 crc kubenswrapper[4612]: E0227 07:54:54.444474 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="201de829-c0b7-4d50-9a3e-ef65d3c06916" containerName="extract-utilities" Feb 27 07:54:54 crc kubenswrapper[4612]: I0227 07:54:54.444484 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="201de829-c0b7-4d50-9a3e-ef65d3c06916" containerName="extract-utilities" Feb 27 07:54:54 crc kubenswrapper[4612]: E0227 07:54:54.444499 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7faf5006-23b1-4ef4-ad29-e0e676340a7c" containerName="registry-server" Feb 27 07:54:54 crc kubenswrapper[4612]: I0227 07:54:54.444528 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="7faf5006-23b1-4ef4-ad29-e0e676340a7c" containerName="registry-server" Feb 27 07:54:54 crc kubenswrapper[4612]: E0227 07:54:54.444546 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="201de829-c0b7-4d50-9a3e-ef65d3c06916" containerName="extract-content" Feb 27 07:54:54 crc kubenswrapper[4612]: I0227 07:54:54.444554 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="201de829-c0b7-4d50-9a3e-ef65d3c06916" containerName="extract-content" Feb 27 07:54:54 crc kubenswrapper[4612]: E0227 07:54:54.444563 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a321b7e1-dbb6-4839-b112-14d7674639c0" containerName="extract-content" Feb 27 07:54:54 crc kubenswrapper[4612]: I0227 07:54:54.444569 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="a321b7e1-dbb6-4839-b112-14d7674639c0" containerName="extract-content" Feb 27 07:54:54 crc kubenswrapper[4612]: E0227 07:54:54.444608 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7faf5006-23b1-4ef4-ad29-e0e676340a7c" containerName="extract-content" Feb 27 07:54:54 crc kubenswrapper[4612]: I0227 07:54:54.444618 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="7faf5006-23b1-4ef4-ad29-e0e676340a7c" containerName="extract-content" Feb 27 07:54:54 crc kubenswrapper[4612]: E0227 07:54:54.444639 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="201de829-c0b7-4d50-9a3e-ef65d3c06916" containerName="registry-server" Feb 27 07:54:54 crc kubenswrapper[4612]: I0227 07:54:54.444647 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="201de829-c0b7-4d50-9a3e-ef65d3c06916" containerName="registry-server" Feb 27 07:54:54 crc kubenswrapper[4612]: E0227 07:54:54.444681 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d061f42-06de-43e5-b4d0-dbca09e48630" containerName="marketplace-operator" Feb 27 07:54:54 crc kubenswrapper[4612]: I0227 07:54:54.444702 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d061f42-06de-43e5-b4d0-dbca09e48630" containerName="marketplace-operator" Feb 27 07:54:54 crc kubenswrapper[4612]: E0227 07:54:54.444721 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0bed0a85-ad8c-4670-8193-0e9a90e88d78" containerName="registry-server" Feb 27 07:54:54 crc kubenswrapper[4612]: I0227 07:54:54.444729 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="0bed0a85-ad8c-4670-8193-0e9a90e88d78" containerName="registry-server" Feb 27 07:54:54 crc kubenswrapper[4612]: E0227 07:54:54.444771 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0bed0a85-ad8c-4670-8193-0e9a90e88d78" containerName="extract-content" Feb 27 07:54:54 crc kubenswrapper[4612]: I0227 07:54:54.444781 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="0bed0a85-ad8c-4670-8193-0e9a90e88d78" containerName="extract-content" Feb 27 07:54:54 crc kubenswrapper[4612]: E0227 07:54:54.444797 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a321b7e1-dbb6-4839-b112-14d7674639c0" containerName="extract-utilities" Feb 27 07:54:54 crc kubenswrapper[4612]: I0227 07:54:54.444805 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="a321b7e1-dbb6-4839-b112-14d7674639c0" containerName="extract-utilities" Feb 27 07:54:54 crc kubenswrapper[4612]: E0227 07:54:54.445068 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d061f42-06de-43e5-b4d0-dbca09e48630" containerName="marketplace-operator" Feb 27 07:54:54 crc kubenswrapper[4612]: I0227 07:54:54.445083 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d061f42-06de-43e5-b4d0-dbca09e48630" containerName="marketplace-operator" Feb 27 07:54:54 crc kubenswrapper[4612]: I0227 07:54:54.445419 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="0bed0a85-ad8c-4670-8193-0e9a90e88d78" containerName="registry-server" Feb 27 07:54:54 crc kubenswrapper[4612]: I0227 07:54:54.445447 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="7faf5006-23b1-4ef4-ad29-e0e676340a7c" containerName="registry-server" Feb 27 07:54:54 crc kubenswrapper[4612]: I0227 07:54:54.445502 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="201de829-c0b7-4d50-9a3e-ef65d3c06916" containerName="registry-server" Feb 27 07:54:54 crc kubenswrapper[4612]: I0227 07:54:54.445521 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="6d061f42-06de-43e5-b4d0-dbca09e48630" containerName="marketplace-operator" Feb 27 07:54:54 crc kubenswrapper[4612]: I0227 07:54:54.445531 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="6d061f42-06de-43e5-b4d0-dbca09e48630" containerName="marketplace-operator" Feb 27 07:54:54 crc kubenswrapper[4612]: I0227 07:54:54.445578 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="a321b7e1-dbb6-4839-b112-14d7674639c0" containerName="registry-server" Feb 27 07:54:54 crc kubenswrapper[4612]: I0227 07:54:54.449768 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8dbvm" Feb 27 07:54:54 crc kubenswrapper[4612]: I0227 07:54:54.455716 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Feb 27 07:54:54 crc kubenswrapper[4612]: I0227 07:54:54.456589 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-8dbvm"] Feb 27 07:54:54 crc kubenswrapper[4612]: I0227 07:54:54.552252 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eb785264-58b2-4d87-89f3-8dc86f784419-utilities\") pod \"redhat-marketplace-8dbvm\" (UID: \"eb785264-58b2-4d87-89f3-8dc86f784419\") " pod="openshift-marketplace/redhat-marketplace-8dbvm" Feb 27 07:54:54 crc kubenswrapper[4612]: I0227 07:54:54.552581 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9vrbw\" (UniqueName: \"kubernetes.io/projected/eb785264-58b2-4d87-89f3-8dc86f784419-kube-api-access-9vrbw\") pod \"redhat-marketplace-8dbvm\" (UID: \"eb785264-58b2-4d87-89f3-8dc86f784419\") " pod="openshift-marketplace/redhat-marketplace-8dbvm" Feb 27 07:54:54 crc kubenswrapper[4612]: I0227 07:54:54.552798 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eb785264-58b2-4d87-89f3-8dc86f784419-catalog-content\") pod \"redhat-marketplace-8dbvm\" (UID: \"eb785264-58b2-4d87-89f3-8dc86f784419\") " pod="openshift-marketplace/redhat-marketplace-8dbvm" Feb 27 07:54:54 crc kubenswrapper[4612]: I0227 07:54:54.599626 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-x56jg"] Feb 27 07:54:54 crc kubenswrapper[4612]: I0227 07:54:54.600813 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-x56jg" Feb 27 07:54:54 crc kubenswrapper[4612]: I0227 07:54:54.602793 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Feb 27 07:54:54 crc kubenswrapper[4612]: I0227 07:54:54.612952 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-x56jg"] Feb 27 07:54:54 crc kubenswrapper[4612]: I0227 07:54:54.653905 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eb785264-58b2-4d87-89f3-8dc86f784419-catalog-content\") pod \"redhat-marketplace-8dbvm\" (UID: \"eb785264-58b2-4d87-89f3-8dc86f784419\") " pod="openshift-marketplace/redhat-marketplace-8dbvm" Feb 27 07:54:54 crc kubenswrapper[4612]: I0227 07:54:54.653991 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eb785264-58b2-4d87-89f3-8dc86f784419-utilities\") pod \"redhat-marketplace-8dbvm\" (UID: \"eb785264-58b2-4d87-89f3-8dc86f784419\") " pod="openshift-marketplace/redhat-marketplace-8dbvm" Feb 27 07:54:54 crc kubenswrapper[4612]: I0227 07:54:54.654055 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9vrbw\" (UniqueName: \"kubernetes.io/projected/eb785264-58b2-4d87-89f3-8dc86f784419-kube-api-access-9vrbw\") pod \"redhat-marketplace-8dbvm\" (UID: \"eb785264-58b2-4d87-89f3-8dc86f784419\") " pod="openshift-marketplace/redhat-marketplace-8dbvm" Feb 27 07:54:54 crc kubenswrapper[4612]: I0227 07:54:54.654412 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eb785264-58b2-4d87-89f3-8dc86f784419-catalog-content\") pod \"redhat-marketplace-8dbvm\" (UID: \"eb785264-58b2-4d87-89f3-8dc86f784419\") " pod="openshift-marketplace/redhat-marketplace-8dbvm" Feb 27 07:54:54 crc kubenswrapper[4612]: I0227 07:54:54.654586 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eb785264-58b2-4d87-89f3-8dc86f784419-utilities\") pod \"redhat-marketplace-8dbvm\" (UID: \"eb785264-58b2-4d87-89f3-8dc86f784419\") " pod="openshift-marketplace/redhat-marketplace-8dbvm" Feb 27 07:54:54 crc kubenswrapper[4612]: I0227 07:54:54.679942 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9vrbw\" (UniqueName: \"kubernetes.io/projected/eb785264-58b2-4d87-89f3-8dc86f784419-kube-api-access-9vrbw\") pod \"redhat-marketplace-8dbvm\" (UID: \"eb785264-58b2-4d87-89f3-8dc86f784419\") " pod="openshift-marketplace/redhat-marketplace-8dbvm" Feb 27 07:54:54 crc kubenswrapper[4612]: I0227 07:54:54.755928 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cba13562-dcb0-4896-94eb-5d5246ffa1c2-utilities\") pod \"redhat-operators-x56jg\" (UID: \"cba13562-dcb0-4896-94eb-5d5246ffa1c2\") " pod="openshift-marketplace/redhat-operators-x56jg" Feb 27 07:54:54 crc kubenswrapper[4612]: I0227 07:54:54.756004 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m89z4\" (UniqueName: \"kubernetes.io/projected/cba13562-dcb0-4896-94eb-5d5246ffa1c2-kube-api-access-m89z4\") pod \"redhat-operators-x56jg\" (UID: \"cba13562-dcb0-4896-94eb-5d5246ffa1c2\") " pod="openshift-marketplace/redhat-operators-x56jg" Feb 27 07:54:54 crc kubenswrapper[4612]: I0227 07:54:54.756058 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cba13562-dcb0-4896-94eb-5d5246ffa1c2-catalog-content\") pod \"redhat-operators-x56jg\" (UID: \"cba13562-dcb0-4896-94eb-5d5246ffa1c2\") " pod="openshift-marketplace/redhat-operators-x56jg" Feb 27 07:54:54 crc kubenswrapper[4612]: I0227 07:54:54.778841 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8dbvm" Feb 27 07:54:54 crc kubenswrapper[4612]: I0227 07:54:54.856823 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cba13562-dcb0-4896-94eb-5d5246ffa1c2-catalog-content\") pod \"redhat-operators-x56jg\" (UID: \"cba13562-dcb0-4896-94eb-5d5246ffa1c2\") " pod="openshift-marketplace/redhat-operators-x56jg" Feb 27 07:54:54 crc kubenswrapper[4612]: I0227 07:54:54.857104 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cba13562-dcb0-4896-94eb-5d5246ffa1c2-utilities\") pod \"redhat-operators-x56jg\" (UID: \"cba13562-dcb0-4896-94eb-5d5246ffa1c2\") " pod="openshift-marketplace/redhat-operators-x56jg" Feb 27 07:54:54 crc kubenswrapper[4612]: I0227 07:54:54.857132 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m89z4\" (UniqueName: \"kubernetes.io/projected/cba13562-dcb0-4896-94eb-5d5246ffa1c2-kube-api-access-m89z4\") pod \"redhat-operators-x56jg\" (UID: \"cba13562-dcb0-4896-94eb-5d5246ffa1c2\") " pod="openshift-marketplace/redhat-operators-x56jg" Feb 27 07:54:54 crc kubenswrapper[4612]: I0227 07:54:54.857780 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cba13562-dcb0-4896-94eb-5d5246ffa1c2-catalog-content\") pod \"redhat-operators-x56jg\" (UID: \"cba13562-dcb0-4896-94eb-5d5246ffa1c2\") " pod="openshift-marketplace/redhat-operators-x56jg" Feb 27 07:54:54 crc kubenswrapper[4612]: I0227 07:54:54.857982 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cba13562-dcb0-4896-94eb-5d5246ffa1c2-utilities\") pod \"redhat-operators-x56jg\" (UID: \"cba13562-dcb0-4896-94eb-5d5246ffa1c2\") " pod="openshift-marketplace/redhat-operators-x56jg" Feb 27 07:54:54 crc kubenswrapper[4612]: I0227 07:54:54.869865 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0bed0a85-ad8c-4670-8193-0e9a90e88d78" path="/var/lib/kubelet/pods/0bed0a85-ad8c-4670-8193-0e9a90e88d78/volumes" Feb 27 07:54:54 crc kubenswrapper[4612]: I0227 07:54:54.870665 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="201de829-c0b7-4d50-9a3e-ef65d3c06916" path="/var/lib/kubelet/pods/201de829-c0b7-4d50-9a3e-ef65d3c06916/volumes" Feb 27 07:54:54 crc kubenswrapper[4612]: I0227 07:54:54.871431 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6d061f42-06de-43e5-b4d0-dbca09e48630" path="/var/lib/kubelet/pods/6d061f42-06de-43e5-b4d0-dbca09e48630/volumes" Feb 27 07:54:54 crc kubenswrapper[4612]: I0227 07:54:54.873977 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7faf5006-23b1-4ef4-ad29-e0e676340a7c" path="/var/lib/kubelet/pods/7faf5006-23b1-4ef4-ad29-e0e676340a7c/volumes" Feb 27 07:54:54 crc kubenswrapper[4612]: I0227 07:54:54.875246 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a321b7e1-dbb6-4839-b112-14d7674639c0" path="/var/lib/kubelet/pods/a321b7e1-dbb6-4839-b112-14d7674639c0/volumes" Feb 27 07:54:54 crc kubenswrapper[4612]: I0227 07:54:54.894458 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m89z4\" (UniqueName: \"kubernetes.io/projected/cba13562-dcb0-4896-94eb-5d5246ffa1c2-kube-api-access-m89z4\") pod \"redhat-operators-x56jg\" (UID: \"cba13562-dcb0-4896-94eb-5d5246ffa1c2\") " pod="openshift-marketplace/redhat-operators-x56jg" Feb 27 07:54:54 crc kubenswrapper[4612]: I0227 07:54:54.920316 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-x56jg" Feb 27 07:54:54 crc kubenswrapper[4612]: I0227 07:54:54.952746 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-8dbvm"] Feb 27 07:54:54 crc kubenswrapper[4612]: W0227 07:54:54.960391 4612 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podeb785264_58b2_4d87_89f3_8dc86f784419.slice/crio-57420dbc4b640d625d51cfbb800aa10bc1702a6a73e392677edd12f757e611e9 WatchSource:0}: Error finding container 57420dbc4b640d625d51cfbb800aa10bc1702a6a73e392677edd12f757e611e9: Status 404 returned error can't find the container with id 57420dbc4b640d625d51cfbb800aa10bc1702a6a73e392677edd12f757e611e9 Feb 27 07:54:55 crc kubenswrapper[4612]: I0227 07:54:55.105044 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-x56jg"] Feb 27 07:54:55 crc kubenswrapper[4612]: W0227 07:54:55.109441 4612 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcba13562_dcb0_4896_94eb_5d5246ffa1c2.slice/crio-335eca84cd8cc747dc210e1e7eb667e66c09a37df1a5626a2dfd5e652f804ed0 WatchSource:0}: Error finding container 335eca84cd8cc747dc210e1e7eb667e66c09a37df1a5626a2dfd5e652f804ed0: Status 404 returned error can't find the container with id 335eca84cd8cc747dc210e1e7eb667e66c09a37df1a5626a2dfd5e652f804ed0 Feb 27 07:54:55 crc kubenswrapper[4612]: I0227 07:54:55.340904 4612 generic.go:334] "Generic (PLEG): container finished" podID="cba13562-dcb0-4896-94eb-5d5246ffa1c2" containerID="330db48510988c344c158c483bd8c38e75806a7aa272bffd37a84633fac9a0f4" exitCode=0 Feb 27 07:54:55 crc kubenswrapper[4612]: I0227 07:54:55.340979 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x56jg" event={"ID":"cba13562-dcb0-4896-94eb-5d5246ffa1c2","Type":"ContainerDied","Data":"330db48510988c344c158c483bd8c38e75806a7aa272bffd37a84633fac9a0f4"} Feb 27 07:54:55 crc kubenswrapper[4612]: I0227 07:54:55.341009 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x56jg" event={"ID":"cba13562-dcb0-4896-94eb-5d5246ffa1c2","Type":"ContainerStarted","Data":"335eca84cd8cc747dc210e1e7eb667e66c09a37df1a5626a2dfd5e652f804ed0"} Feb 27 07:54:55 crc kubenswrapper[4612]: I0227 07:54:55.342710 4612 generic.go:334] "Generic (PLEG): container finished" podID="eb785264-58b2-4d87-89f3-8dc86f784419" containerID="3b5ecdd980bcca739ac294dd240aa7e3906d4eecf735e89c6d0fed8b0b00450b" exitCode=0 Feb 27 07:54:55 crc kubenswrapper[4612]: I0227 07:54:55.342784 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8dbvm" event={"ID":"eb785264-58b2-4d87-89f3-8dc86f784419","Type":"ContainerDied","Data":"3b5ecdd980bcca739ac294dd240aa7e3906d4eecf735e89c6d0fed8b0b00450b"} Feb 27 07:54:55 crc kubenswrapper[4612]: I0227 07:54:55.342850 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8dbvm" event={"ID":"eb785264-58b2-4d87-89f3-8dc86f784419","Type":"ContainerStarted","Data":"57420dbc4b640d625d51cfbb800aa10bc1702a6a73e392677edd12f757e611e9"} Feb 27 07:54:56 crc kubenswrapper[4612]: I0227 07:54:56.812797 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-ln97g"] Feb 27 07:54:56 crc kubenswrapper[4612]: I0227 07:54:56.815176 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ln97g" Feb 27 07:54:56 crc kubenswrapper[4612]: I0227 07:54:56.818823 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-ln97g"] Feb 27 07:54:56 crc kubenswrapper[4612]: I0227 07:54:56.819995 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Feb 27 07:54:56 crc kubenswrapper[4612]: I0227 07:54:56.985163 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tpgzq\" (UniqueName: \"kubernetes.io/projected/4a605168-76bd-4605-9a60-5ea574a0b2c1-kube-api-access-tpgzq\") pod \"certified-operators-ln97g\" (UID: \"4a605168-76bd-4605-9a60-5ea574a0b2c1\") " pod="openshift-marketplace/certified-operators-ln97g" Feb 27 07:54:56 crc kubenswrapper[4612]: I0227 07:54:56.985416 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4a605168-76bd-4605-9a60-5ea574a0b2c1-catalog-content\") pod \"certified-operators-ln97g\" (UID: \"4a605168-76bd-4605-9a60-5ea574a0b2c1\") " pod="openshift-marketplace/certified-operators-ln97g" Feb 27 07:54:56 crc kubenswrapper[4612]: I0227 07:54:56.985563 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4a605168-76bd-4605-9a60-5ea574a0b2c1-utilities\") pod \"certified-operators-ln97g\" (UID: \"4a605168-76bd-4605-9a60-5ea574a0b2c1\") " pod="openshift-marketplace/certified-operators-ln97g" Feb 27 07:54:57 crc kubenswrapper[4612]: I0227 07:54:57.002018 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-ck8cc"] Feb 27 07:54:57 crc kubenswrapper[4612]: I0227 07:54:57.005848 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ck8cc" Feb 27 07:54:57 crc kubenswrapper[4612]: I0227 07:54:57.012488 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Feb 27 07:54:57 crc kubenswrapper[4612]: I0227 07:54:57.036350 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-ck8cc"] Feb 27 07:54:57 crc kubenswrapper[4612]: I0227 07:54:57.086920 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tpgzq\" (UniqueName: \"kubernetes.io/projected/4a605168-76bd-4605-9a60-5ea574a0b2c1-kube-api-access-tpgzq\") pod \"certified-operators-ln97g\" (UID: \"4a605168-76bd-4605-9a60-5ea574a0b2c1\") " pod="openshift-marketplace/certified-operators-ln97g" Feb 27 07:54:57 crc kubenswrapper[4612]: I0227 07:54:57.086978 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4a605168-76bd-4605-9a60-5ea574a0b2c1-catalog-content\") pod \"certified-operators-ln97g\" (UID: \"4a605168-76bd-4605-9a60-5ea574a0b2c1\") " pod="openshift-marketplace/certified-operators-ln97g" Feb 27 07:54:57 crc kubenswrapper[4612]: I0227 07:54:57.087024 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4a605168-76bd-4605-9a60-5ea574a0b2c1-utilities\") pod \"certified-operators-ln97g\" (UID: \"4a605168-76bd-4605-9a60-5ea574a0b2c1\") " pod="openshift-marketplace/certified-operators-ln97g" Feb 27 07:54:57 crc kubenswrapper[4612]: I0227 07:54:57.087826 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4a605168-76bd-4605-9a60-5ea574a0b2c1-catalog-content\") pod \"certified-operators-ln97g\" (UID: \"4a605168-76bd-4605-9a60-5ea574a0b2c1\") " pod="openshift-marketplace/certified-operators-ln97g" Feb 27 07:54:57 crc kubenswrapper[4612]: I0227 07:54:57.088373 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4a605168-76bd-4605-9a60-5ea574a0b2c1-utilities\") pod \"certified-operators-ln97g\" (UID: \"4a605168-76bd-4605-9a60-5ea574a0b2c1\") " pod="openshift-marketplace/certified-operators-ln97g" Feb 27 07:54:57 crc kubenswrapper[4612]: I0227 07:54:57.107688 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tpgzq\" (UniqueName: \"kubernetes.io/projected/4a605168-76bd-4605-9a60-5ea574a0b2c1-kube-api-access-tpgzq\") pod \"certified-operators-ln97g\" (UID: \"4a605168-76bd-4605-9a60-5ea574a0b2c1\") " pod="openshift-marketplace/certified-operators-ln97g" Feb 27 07:54:57 crc kubenswrapper[4612]: I0227 07:54:57.133777 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ln97g" Feb 27 07:54:57 crc kubenswrapper[4612]: I0227 07:54:57.188289 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f2b71124-3e61-48fc-a493-60c759124936-utilities\") pod \"community-operators-ck8cc\" (UID: \"f2b71124-3e61-48fc-a493-60c759124936\") " pod="openshift-marketplace/community-operators-ck8cc" Feb 27 07:54:57 crc kubenswrapper[4612]: I0227 07:54:57.188339 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f2b71124-3e61-48fc-a493-60c759124936-catalog-content\") pod \"community-operators-ck8cc\" (UID: \"f2b71124-3e61-48fc-a493-60c759124936\") " pod="openshift-marketplace/community-operators-ck8cc" Feb 27 07:54:57 crc kubenswrapper[4612]: I0227 07:54:57.188425 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jgbd9\" (UniqueName: \"kubernetes.io/projected/f2b71124-3e61-48fc-a493-60c759124936-kube-api-access-jgbd9\") pod \"community-operators-ck8cc\" (UID: \"f2b71124-3e61-48fc-a493-60c759124936\") " pod="openshift-marketplace/community-operators-ck8cc" Feb 27 07:54:57 crc kubenswrapper[4612]: I0227 07:54:57.293085 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f2b71124-3e61-48fc-a493-60c759124936-utilities\") pod \"community-operators-ck8cc\" (UID: \"f2b71124-3e61-48fc-a493-60c759124936\") " pod="openshift-marketplace/community-operators-ck8cc" Feb 27 07:54:57 crc kubenswrapper[4612]: I0227 07:54:57.293343 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f2b71124-3e61-48fc-a493-60c759124936-catalog-content\") pod \"community-operators-ck8cc\" (UID: \"f2b71124-3e61-48fc-a493-60c759124936\") " pod="openshift-marketplace/community-operators-ck8cc" Feb 27 07:54:57 crc kubenswrapper[4612]: I0227 07:54:57.293409 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jgbd9\" (UniqueName: \"kubernetes.io/projected/f2b71124-3e61-48fc-a493-60c759124936-kube-api-access-jgbd9\") pod \"community-operators-ck8cc\" (UID: \"f2b71124-3e61-48fc-a493-60c759124936\") " pod="openshift-marketplace/community-operators-ck8cc" Feb 27 07:54:57 crc kubenswrapper[4612]: I0227 07:54:57.294435 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f2b71124-3e61-48fc-a493-60c759124936-utilities\") pod \"community-operators-ck8cc\" (UID: \"f2b71124-3e61-48fc-a493-60c759124936\") " pod="openshift-marketplace/community-operators-ck8cc" Feb 27 07:54:57 crc kubenswrapper[4612]: I0227 07:54:57.294538 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f2b71124-3e61-48fc-a493-60c759124936-catalog-content\") pod \"community-operators-ck8cc\" (UID: \"f2b71124-3e61-48fc-a493-60c759124936\") " pod="openshift-marketplace/community-operators-ck8cc" Feb 27 07:54:57 crc kubenswrapper[4612]: I0227 07:54:57.324588 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jgbd9\" (UniqueName: \"kubernetes.io/projected/f2b71124-3e61-48fc-a493-60c759124936-kube-api-access-jgbd9\") pod \"community-operators-ck8cc\" (UID: \"f2b71124-3e61-48fc-a493-60c759124936\") " pod="openshift-marketplace/community-operators-ck8cc" Feb 27 07:54:57 crc kubenswrapper[4612]: I0227 07:54:57.325083 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ck8cc" Feb 27 07:54:57 crc kubenswrapper[4612]: I0227 07:54:57.356426 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x56jg" event={"ID":"cba13562-dcb0-4896-94eb-5d5246ffa1c2","Type":"ContainerStarted","Data":"5e28abb2cb881b397063c473d743fcb9261940a864617577bb65938d9055ccaf"} Feb 27 07:54:57 crc kubenswrapper[4612]: I0227 07:54:57.358494 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8dbvm" event={"ID":"eb785264-58b2-4d87-89f3-8dc86f784419","Type":"ContainerDied","Data":"92c4b2659f5758a68e63faf70c9b1a15807cc87e34fa7f8284b0e157b2ebfcdc"} Feb 27 07:54:57 crc kubenswrapper[4612]: I0227 07:54:57.358491 4612 generic.go:334] "Generic (PLEG): container finished" podID="eb785264-58b2-4d87-89f3-8dc86f784419" containerID="92c4b2659f5758a68e63faf70c9b1a15807cc87e34fa7f8284b0e157b2ebfcdc" exitCode=0 Feb 27 07:54:57 crc kubenswrapper[4612]: E0227 07:54:57.415768 4612 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podeb785264_58b2_4d87_89f3_8dc86f784419.slice/crio-conmon-92c4b2659f5758a68e63faf70c9b1a15807cc87e34fa7f8284b0e157b2ebfcdc.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podeb785264_58b2_4d87_89f3_8dc86f784419.slice/crio-92c4b2659f5758a68e63faf70c9b1a15807cc87e34fa7f8284b0e157b2ebfcdc.scope\": RecentStats: unable to find data in memory cache]" Feb 27 07:54:57 crc kubenswrapper[4612]: I0227 07:54:57.497494 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-ck8cc"] Feb 27 07:54:57 crc kubenswrapper[4612]: W0227 07:54:57.512776 4612 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf2b71124_3e61_48fc_a493_60c759124936.slice/crio-d519cc8b5b0bf9379197b1bce79305ca8e2d37dc382b4b09b49bc3d3f2d2138f WatchSource:0}: Error finding container d519cc8b5b0bf9379197b1bce79305ca8e2d37dc382b4b09b49bc3d3f2d2138f: Status 404 returned error can't find the container with id d519cc8b5b0bf9379197b1bce79305ca8e2d37dc382b4b09b49bc3d3f2d2138f Feb 27 07:54:57 crc kubenswrapper[4612]: I0227 07:54:57.599924 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-ln97g"] Feb 27 07:54:57 crc kubenswrapper[4612]: W0227 07:54:57.619038 4612 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4a605168_76bd_4605_9a60_5ea574a0b2c1.slice/crio-d77b2b78500e4721bdc594849b42fc57f0b097d6f610d191555ab960059e1bab WatchSource:0}: Error finding container d77b2b78500e4721bdc594849b42fc57f0b097d6f610d191555ab960059e1bab: Status 404 returned error can't find the container with id d77b2b78500e4721bdc594849b42fc57f0b097d6f610d191555ab960059e1bab Feb 27 07:54:58 crc kubenswrapper[4612]: I0227 07:54:58.366393 4612 generic.go:334] "Generic (PLEG): container finished" podID="f2b71124-3e61-48fc-a493-60c759124936" containerID="ac674729d26ee09b2f1a54d09f874c1e253783277fa0a9b325269c8d90e0e241" exitCode=0 Feb 27 07:54:58 crc kubenswrapper[4612]: I0227 07:54:58.366481 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ck8cc" event={"ID":"f2b71124-3e61-48fc-a493-60c759124936","Type":"ContainerDied","Data":"ac674729d26ee09b2f1a54d09f874c1e253783277fa0a9b325269c8d90e0e241"} Feb 27 07:54:58 crc kubenswrapper[4612]: I0227 07:54:58.366794 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ck8cc" event={"ID":"f2b71124-3e61-48fc-a493-60c759124936","Type":"ContainerStarted","Data":"d519cc8b5b0bf9379197b1bce79305ca8e2d37dc382b4b09b49bc3d3f2d2138f"} Feb 27 07:54:58 crc kubenswrapper[4612]: I0227 07:54:58.369884 4612 generic.go:334] "Generic (PLEG): container finished" podID="cba13562-dcb0-4896-94eb-5d5246ffa1c2" containerID="5e28abb2cb881b397063c473d743fcb9261940a864617577bb65938d9055ccaf" exitCode=0 Feb 27 07:54:58 crc kubenswrapper[4612]: I0227 07:54:58.369935 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x56jg" event={"ID":"cba13562-dcb0-4896-94eb-5d5246ffa1c2","Type":"ContainerDied","Data":"5e28abb2cb881b397063c473d743fcb9261940a864617577bb65938d9055ccaf"} Feb 27 07:54:58 crc kubenswrapper[4612]: I0227 07:54:58.374100 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8dbvm" event={"ID":"eb785264-58b2-4d87-89f3-8dc86f784419","Type":"ContainerStarted","Data":"83e2efd0cce6f6947cdf7f6884717df49cc17267ff5482e138e1f36a8c074365"} Feb 27 07:54:58 crc kubenswrapper[4612]: I0227 07:54:58.376250 4612 generic.go:334] "Generic (PLEG): container finished" podID="4a605168-76bd-4605-9a60-5ea574a0b2c1" containerID="f64524b49e1d0afca88c5737bcec021db257d3f6cd97e8eeceb26f87ebd7f503" exitCode=0 Feb 27 07:54:58 crc kubenswrapper[4612]: I0227 07:54:58.376299 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ln97g" event={"ID":"4a605168-76bd-4605-9a60-5ea574a0b2c1","Type":"ContainerDied","Data":"f64524b49e1d0afca88c5737bcec021db257d3f6cd97e8eeceb26f87ebd7f503"} Feb 27 07:54:58 crc kubenswrapper[4612]: I0227 07:54:58.376324 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ln97g" event={"ID":"4a605168-76bd-4605-9a60-5ea574a0b2c1","Type":"ContainerStarted","Data":"d77b2b78500e4721bdc594849b42fc57f0b097d6f610d191555ab960059e1bab"} Feb 27 07:54:58 crc kubenswrapper[4612]: I0227 07:54:58.435160 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-8dbvm" podStartSLOduration=2.043189319 podStartE2EDuration="4.435146153s" podCreationTimestamp="2026-02-27 07:54:54 +0000 UTC" firstStartedPulling="2026-02-27 07:54:55.344947303 +0000 UTC m=+353.198877311" lastFinishedPulling="2026-02-27 07:54:57.736904147 +0000 UTC m=+355.590834145" observedRunningTime="2026-02-27 07:54:58.432561155 +0000 UTC m=+356.286491153" watchObservedRunningTime="2026-02-27 07:54:58.435146153 +0000 UTC m=+356.289076151" Feb 27 07:54:59 crc kubenswrapper[4612]: I0227 07:54:59.383851 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x56jg" event={"ID":"cba13562-dcb0-4896-94eb-5d5246ffa1c2","Type":"ContainerStarted","Data":"f395d86e6d2744ac940efd5ea0ceb2bef047b485b2db58f64e91457ee5b166aa"} Feb 27 07:54:59 crc kubenswrapper[4612]: I0227 07:54:59.386774 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ln97g" event={"ID":"4a605168-76bd-4605-9a60-5ea574a0b2c1","Type":"ContainerStarted","Data":"feda41e89d77b055b6cd3dd3b948a1a4091d460c6b4f5572ecd79ce9838c8e21"} Feb 27 07:54:59 crc kubenswrapper[4612]: I0227 07:54:59.389363 4612 generic.go:334] "Generic (PLEG): container finished" podID="f2b71124-3e61-48fc-a493-60c759124936" containerID="126f26db60b09716c669667fd580133de63d0ea20cf6695228d29ec06db97991" exitCode=0 Feb 27 07:54:59 crc kubenswrapper[4612]: I0227 07:54:59.389405 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ck8cc" event={"ID":"f2b71124-3e61-48fc-a493-60c759124936","Type":"ContainerDied","Data":"126f26db60b09716c669667fd580133de63d0ea20cf6695228d29ec06db97991"} Feb 27 07:54:59 crc kubenswrapper[4612]: I0227 07:54:59.406300 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-x56jg" podStartSLOduration=2.01636457 podStartE2EDuration="5.406277452s" podCreationTimestamp="2026-02-27 07:54:54 +0000 UTC" firstStartedPulling="2026-02-27 07:54:55.343074356 +0000 UTC m=+353.197004384" lastFinishedPulling="2026-02-27 07:54:58.732987268 +0000 UTC m=+356.586917266" observedRunningTime="2026-02-27 07:54:59.40354773 +0000 UTC m=+357.257477728" watchObservedRunningTime="2026-02-27 07:54:59.406277452 +0000 UTC m=+357.260207460" Feb 27 07:55:00 crc kubenswrapper[4612]: I0227 07:55:00.397319 4612 generic.go:334] "Generic (PLEG): container finished" podID="4a605168-76bd-4605-9a60-5ea574a0b2c1" containerID="feda41e89d77b055b6cd3dd3b948a1a4091d460c6b4f5572ecd79ce9838c8e21" exitCode=0 Feb 27 07:55:00 crc kubenswrapper[4612]: I0227 07:55:00.397384 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ln97g" event={"ID":"4a605168-76bd-4605-9a60-5ea574a0b2c1","Type":"ContainerDied","Data":"feda41e89d77b055b6cd3dd3b948a1a4091d460c6b4f5572ecd79ce9838c8e21"} Feb 27 07:55:00 crc kubenswrapper[4612]: I0227 07:55:00.400215 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ck8cc" event={"ID":"f2b71124-3e61-48fc-a493-60c759124936","Type":"ContainerStarted","Data":"e23394db55e865166fb5e3e87432c1ff4e818b304dcf664a3ede7c1af06879c4"} Feb 27 07:55:00 crc kubenswrapper[4612]: I0227 07:55:00.450552 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-ck8cc" podStartSLOduration=2.915279354 podStartE2EDuration="4.450521092s" podCreationTimestamp="2026-02-27 07:54:56 +0000 UTC" firstStartedPulling="2026-02-27 07:54:58.368109285 +0000 UTC m=+356.222039293" lastFinishedPulling="2026-02-27 07:54:59.903351033 +0000 UTC m=+357.757281031" observedRunningTime="2026-02-27 07:55:00.444014896 +0000 UTC m=+358.297944904" watchObservedRunningTime="2026-02-27 07:55:00.450521092 +0000 UTC m=+358.304451110" Feb 27 07:55:01 crc kubenswrapper[4612]: I0227 07:55:01.413057 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ln97g" event={"ID":"4a605168-76bd-4605-9a60-5ea574a0b2c1","Type":"ContainerStarted","Data":"96b9c5352b982c73f6230b55624bbbc342ce824cfad1a912895e019ab29441f4"} Feb 27 07:55:04 crc kubenswrapper[4612]: I0227 07:55:04.780753 4612 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-8dbvm" Feb 27 07:55:04 crc kubenswrapper[4612]: I0227 07:55:04.780972 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-8dbvm" Feb 27 07:55:04 crc kubenswrapper[4612]: I0227 07:55:04.845592 4612 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-8dbvm" Feb 27 07:55:04 crc kubenswrapper[4612]: I0227 07:55:04.864579 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-ln97g" podStartSLOduration=6.088737509 podStartE2EDuration="8.864558197s" podCreationTimestamp="2026-02-27 07:54:56 +0000 UTC" firstStartedPulling="2026-02-27 07:54:58.377737735 +0000 UTC m=+356.231667733" lastFinishedPulling="2026-02-27 07:55:01.153558413 +0000 UTC m=+359.007488421" observedRunningTime="2026-02-27 07:55:01.441727646 +0000 UTC m=+359.295657654" watchObservedRunningTime="2026-02-27 07:55:04.864558197 +0000 UTC m=+362.718488205" Feb 27 07:55:04 crc kubenswrapper[4612]: I0227 07:55:04.921715 4612 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-x56jg" Feb 27 07:55:04 crc kubenswrapper[4612]: I0227 07:55:04.921775 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-x56jg" Feb 27 07:55:05 crc kubenswrapper[4612]: I0227 07:55:05.477469 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-8dbvm" Feb 27 07:55:05 crc kubenswrapper[4612]: I0227 07:55:05.964214 4612 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-x56jg" podUID="cba13562-dcb0-4896-94eb-5d5246ffa1c2" containerName="registry-server" probeResult="failure" output=< Feb 27 07:55:05 crc kubenswrapper[4612]: timeout: failed to connect service ":50051" within 1s Feb 27 07:55:05 crc kubenswrapper[4612]: > Feb 27 07:55:07 crc kubenswrapper[4612]: I0227 07:55:07.134619 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-ln97g" Feb 27 07:55:07 crc kubenswrapper[4612]: I0227 07:55:07.134968 4612 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-ln97g" Feb 27 07:55:07 crc kubenswrapper[4612]: I0227 07:55:07.183374 4612 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-ln97g" Feb 27 07:55:07 crc kubenswrapper[4612]: I0227 07:55:07.326574 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-ck8cc" Feb 27 07:55:07 crc kubenswrapper[4612]: I0227 07:55:07.327894 4612 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-ck8cc" Feb 27 07:55:07 crc kubenswrapper[4612]: I0227 07:55:07.376822 4612 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-ck8cc" Feb 27 07:55:07 crc kubenswrapper[4612]: I0227 07:55:07.500427 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-ck8cc" Feb 27 07:55:07 crc kubenswrapper[4612]: I0227 07:55:07.500501 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-ln97g" Feb 27 07:55:14 crc kubenswrapper[4612]: I0227 07:55:14.976346 4612 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-x56jg" Feb 27 07:55:15 crc kubenswrapper[4612]: I0227 07:55:15.037845 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-x56jg" Feb 27 07:55:46 crc kubenswrapper[4612]: I0227 07:55:46.027458 4612 patch_prober.go:28] interesting pod/machine-config-daemon-924vb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 27 07:55:46 crc kubenswrapper[4612]: I0227 07:55:46.028848 4612 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 27 07:56:00 crc kubenswrapper[4612]: I0227 07:56:00.142764 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29536316-5mt4v"] Feb 27 07:56:00 crc kubenswrapper[4612]: I0227 07:56:00.144197 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536316-5mt4v" Feb 27 07:56:00 crc kubenswrapper[4612]: I0227 07:56:00.151562 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 27 07:56:00 crc kubenswrapper[4612]: I0227 07:56:00.151900 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 27 07:56:00 crc kubenswrapper[4612]: I0227 07:56:00.152910 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-5zwfx" Feb 27 07:56:00 crc kubenswrapper[4612]: I0227 07:56:00.226545 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536316-5mt4v"] Feb 27 07:56:00 crc kubenswrapper[4612]: I0227 07:56:00.246682 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zxpz2\" (UniqueName: \"kubernetes.io/projected/863bc847-68ff-4ddc-92c6-eb4f95a00347-kube-api-access-zxpz2\") pod \"auto-csr-approver-29536316-5mt4v\" (UID: \"863bc847-68ff-4ddc-92c6-eb4f95a00347\") " pod="openshift-infra/auto-csr-approver-29536316-5mt4v" Feb 27 07:56:00 crc kubenswrapper[4612]: I0227 07:56:00.348548 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zxpz2\" (UniqueName: \"kubernetes.io/projected/863bc847-68ff-4ddc-92c6-eb4f95a00347-kube-api-access-zxpz2\") pod \"auto-csr-approver-29536316-5mt4v\" (UID: \"863bc847-68ff-4ddc-92c6-eb4f95a00347\") " pod="openshift-infra/auto-csr-approver-29536316-5mt4v" Feb 27 07:56:00 crc kubenswrapper[4612]: I0227 07:56:00.379348 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zxpz2\" (UniqueName: \"kubernetes.io/projected/863bc847-68ff-4ddc-92c6-eb4f95a00347-kube-api-access-zxpz2\") pod \"auto-csr-approver-29536316-5mt4v\" (UID: \"863bc847-68ff-4ddc-92c6-eb4f95a00347\") " pod="openshift-infra/auto-csr-approver-29536316-5mt4v" Feb 27 07:56:00 crc kubenswrapper[4612]: I0227 07:56:00.506490 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536316-5mt4v" Feb 27 07:56:00 crc kubenswrapper[4612]: I0227 07:56:00.747666 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536316-5mt4v"] Feb 27 07:56:00 crc kubenswrapper[4612]: I0227 07:56:00.850922 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536316-5mt4v" event={"ID":"863bc847-68ff-4ddc-92c6-eb4f95a00347","Type":"ContainerStarted","Data":"40a4ba39a25efd52ea919c83a39e1f376fe4766af7aad29346ac12a1a5324eb8"} Feb 27 07:56:02 crc kubenswrapper[4612]: I0227 07:56:02.869552 4612 generic.go:334] "Generic (PLEG): container finished" podID="863bc847-68ff-4ddc-92c6-eb4f95a00347" containerID="6f038fa34ec7b8bedc6054ae35fc4e63cb940f189520a2151c51ca811dee4347" exitCode=0 Feb 27 07:56:02 crc kubenswrapper[4612]: I0227 07:56:02.869637 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536316-5mt4v" event={"ID":"863bc847-68ff-4ddc-92c6-eb4f95a00347","Type":"ContainerDied","Data":"6f038fa34ec7b8bedc6054ae35fc4e63cb940f189520a2151c51ca811dee4347"} Feb 27 07:56:04 crc kubenswrapper[4612]: I0227 07:56:04.177248 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536316-5mt4v" Feb 27 07:56:04 crc kubenswrapper[4612]: I0227 07:56:04.304606 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zxpz2\" (UniqueName: \"kubernetes.io/projected/863bc847-68ff-4ddc-92c6-eb4f95a00347-kube-api-access-zxpz2\") pod \"863bc847-68ff-4ddc-92c6-eb4f95a00347\" (UID: \"863bc847-68ff-4ddc-92c6-eb4f95a00347\") " Feb 27 07:56:04 crc kubenswrapper[4612]: I0227 07:56:04.314059 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/863bc847-68ff-4ddc-92c6-eb4f95a00347-kube-api-access-zxpz2" (OuterVolumeSpecName: "kube-api-access-zxpz2") pod "863bc847-68ff-4ddc-92c6-eb4f95a00347" (UID: "863bc847-68ff-4ddc-92c6-eb4f95a00347"). InnerVolumeSpecName "kube-api-access-zxpz2". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 07:56:04 crc kubenswrapper[4612]: I0227 07:56:04.406750 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zxpz2\" (UniqueName: \"kubernetes.io/projected/863bc847-68ff-4ddc-92c6-eb4f95a00347-kube-api-access-zxpz2\") on node \"crc\" DevicePath \"\"" Feb 27 07:56:04 crc kubenswrapper[4612]: I0227 07:56:04.883452 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536316-5mt4v" event={"ID":"863bc847-68ff-4ddc-92c6-eb4f95a00347","Type":"ContainerDied","Data":"40a4ba39a25efd52ea919c83a39e1f376fe4766af7aad29346ac12a1a5324eb8"} Feb 27 07:56:04 crc kubenswrapper[4612]: I0227 07:56:04.883817 4612 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="40a4ba39a25efd52ea919c83a39e1f376fe4766af7aad29346ac12a1a5324eb8" Feb 27 07:56:04 crc kubenswrapper[4612]: I0227 07:56:04.883517 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536316-5mt4v" Feb 27 07:56:16 crc kubenswrapper[4612]: I0227 07:56:16.027568 4612 patch_prober.go:28] interesting pod/machine-config-daemon-924vb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 27 07:56:16 crc kubenswrapper[4612]: I0227 07:56:16.028244 4612 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 27 07:56:46 crc kubenswrapper[4612]: I0227 07:56:46.027084 4612 patch_prober.go:28] interesting pod/machine-config-daemon-924vb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 27 07:56:46 crc kubenswrapper[4612]: I0227 07:56:46.027645 4612 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 27 07:56:46 crc kubenswrapper[4612]: I0227 07:56:46.027726 4612 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-924vb" Feb 27 07:56:46 crc kubenswrapper[4612]: I0227 07:56:46.028429 4612 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"50feff0d2db2ed217d5a022d814ec77b6e80057f575cba0e92f90bcac1d31a76"} pod="openshift-machine-config-operator/machine-config-daemon-924vb" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 27 07:56:46 crc kubenswrapper[4612]: I0227 07:56:46.028518 4612 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" containerName="machine-config-daemon" containerID="cri-o://50feff0d2db2ed217d5a022d814ec77b6e80057f575cba0e92f90bcac1d31a76" gracePeriod=600 Feb 27 07:56:47 crc kubenswrapper[4612]: I0227 07:56:47.159229 4612 generic.go:334] "Generic (PLEG): container finished" podID="24599373-7adc-4a1b-8bb4-797bf726f43d" containerID="50feff0d2db2ed217d5a022d814ec77b6e80057f575cba0e92f90bcac1d31a76" exitCode=0 Feb 27 07:56:47 crc kubenswrapper[4612]: I0227 07:56:47.159337 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-924vb" event={"ID":"24599373-7adc-4a1b-8bb4-797bf726f43d","Type":"ContainerDied","Data":"50feff0d2db2ed217d5a022d814ec77b6e80057f575cba0e92f90bcac1d31a76"} Feb 27 07:56:47 crc kubenswrapper[4612]: I0227 07:56:47.159684 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-924vb" event={"ID":"24599373-7adc-4a1b-8bb4-797bf726f43d","Type":"ContainerStarted","Data":"778719935b065f111a6a6df70937016b6e803a5b231a05b52794a3dcd1909559"} Feb 27 07:56:47 crc kubenswrapper[4612]: I0227 07:56:47.159733 4612 scope.go:117] "RemoveContainer" containerID="02aa743ba66d91783716a3af076f70d769d77dce3c83913aa953859928a3c90f" Feb 27 07:58:00 crc kubenswrapper[4612]: I0227 07:58:00.145238 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29536318-gjl85"] Feb 27 07:58:00 crc kubenswrapper[4612]: E0227 07:58:00.146341 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="863bc847-68ff-4ddc-92c6-eb4f95a00347" containerName="oc" Feb 27 07:58:00 crc kubenswrapper[4612]: I0227 07:58:00.146359 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="863bc847-68ff-4ddc-92c6-eb4f95a00347" containerName="oc" Feb 27 07:58:00 crc kubenswrapper[4612]: I0227 07:58:00.146528 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="863bc847-68ff-4ddc-92c6-eb4f95a00347" containerName="oc" Feb 27 07:58:00 crc kubenswrapper[4612]: I0227 07:58:00.147127 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536318-gjl85" Feb 27 07:58:00 crc kubenswrapper[4612]: I0227 07:58:00.151383 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-5zwfx" Feb 27 07:58:00 crc kubenswrapper[4612]: I0227 07:58:00.152013 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 27 07:58:00 crc kubenswrapper[4612]: I0227 07:58:00.155738 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 27 07:58:00 crc kubenswrapper[4612]: I0227 07:58:00.162467 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536318-gjl85"] Feb 27 07:58:00 crc kubenswrapper[4612]: I0227 07:58:00.186624 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-89cp4\" (UniqueName: \"kubernetes.io/projected/d76961be-0bf0-45b8-afed-089eaa480485-kube-api-access-89cp4\") pod \"auto-csr-approver-29536318-gjl85\" (UID: \"d76961be-0bf0-45b8-afed-089eaa480485\") " pod="openshift-infra/auto-csr-approver-29536318-gjl85" Feb 27 07:58:00 crc kubenswrapper[4612]: I0227 07:58:00.287817 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-89cp4\" (UniqueName: \"kubernetes.io/projected/d76961be-0bf0-45b8-afed-089eaa480485-kube-api-access-89cp4\") pod \"auto-csr-approver-29536318-gjl85\" (UID: \"d76961be-0bf0-45b8-afed-089eaa480485\") " pod="openshift-infra/auto-csr-approver-29536318-gjl85" Feb 27 07:58:00 crc kubenswrapper[4612]: I0227 07:58:00.306825 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-89cp4\" (UniqueName: \"kubernetes.io/projected/d76961be-0bf0-45b8-afed-089eaa480485-kube-api-access-89cp4\") pod \"auto-csr-approver-29536318-gjl85\" (UID: \"d76961be-0bf0-45b8-afed-089eaa480485\") " pod="openshift-infra/auto-csr-approver-29536318-gjl85" Feb 27 07:58:00 crc kubenswrapper[4612]: I0227 07:58:00.471991 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536318-gjl85" Feb 27 07:58:00 crc kubenswrapper[4612]: I0227 07:58:00.938841 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536318-gjl85"] Feb 27 07:58:00 crc kubenswrapper[4612]: I0227 07:58:00.945208 4612 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 27 07:58:01 crc kubenswrapper[4612]: I0227 07:58:01.657161 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536318-gjl85" event={"ID":"d76961be-0bf0-45b8-afed-089eaa480485","Type":"ContainerStarted","Data":"8942f63904e50db74ae7e9732d553bc9aac2cec50ae3334044ed5326caf4d95a"} Feb 27 07:58:02 crc kubenswrapper[4612]: I0227 07:58:02.667441 4612 generic.go:334] "Generic (PLEG): container finished" podID="d76961be-0bf0-45b8-afed-089eaa480485" containerID="5cd4755114c2182b6ac4cf17cfea7c8aa357aac1c10adadb9c49b8bf54cea10e" exitCode=0 Feb 27 07:58:02 crc kubenswrapper[4612]: I0227 07:58:02.667483 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536318-gjl85" event={"ID":"d76961be-0bf0-45b8-afed-089eaa480485","Type":"ContainerDied","Data":"5cd4755114c2182b6ac4cf17cfea7c8aa357aac1c10adadb9c49b8bf54cea10e"} Feb 27 07:58:03 crc kubenswrapper[4612]: I0227 07:58:03.985936 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536318-gjl85" Feb 27 07:58:04 crc kubenswrapper[4612]: I0227 07:58:04.046416 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-89cp4\" (UniqueName: \"kubernetes.io/projected/d76961be-0bf0-45b8-afed-089eaa480485-kube-api-access-89cp4\") pod \"d76961be-0bf0-45b8-afed-089eaa480485\" (UID: \"d76961be-0bf0-45b8-afed-089eaa480485\") " Feb 27 07:58:04 crc kubenswrapper[4612]: I0227 07:58:04.054904 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d76961be-0bf0-45b8-afed-089eaa480485-kube-api-access-89cp4" (OuterVolumeSpecName: "kube-api-access-89cp4") pod "d76961be-0bf0-45b8-afed-089eaa480485" (UID: "d76961be-0bf0-45b8-afed-089eaa480485"). InnerVolumeSpecName "kube-api-access-89cp4". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 07:58:04 crc kubenswrapper[4612]: I0227 07:58:04.147093 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-89cp4\" (UniqueName: \"kubernetes.io/projected/d76961be-0bf0-45b8-afed-089eaa480485-kube-api-access-89cp4\") on node \"crc\" DevicePath \"\"" Feb 27 07:58:04 crc kubenswrapper[4612]: I0227 07:58:04.683055 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536318-gjl85" event={"ID":"d76961be-0bf0-45b8-afed-089eaa480485","Type":"ContainerDied","Data":"8942f63904e50db74ae7e9732d553bc9aac2cec50ae3334044ed5326caf4d95a"} Feb 27 07:58:04 crc kubenswrapper[4612]: I0227 07:58:04.683114 4612 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8942f63904e50db74ae7e9732d553bc9aac2cec50ae3334044ed5326caf4d95a" Feb 27 07:58:04 crc kubenswrapper[4612]: I0227 07:58:04.683199 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536318-gjl85" Feb 27 07:58:05 crc kubenswrapper[4612]: I0227 07:58:05.051583 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29536312-tjtt6"] Feb 27 07:58:05 crc kubenswrapper[4612]: I0227 07:58:05.056002 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29536312-tjtt6"] Feb 27 07:58:06 crc kubenswrapper[4612]: I0227 07:58:06.862818 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="396f1b18-f225-4d67-a98b-87b2605406d4" path="/var/lib/kubelet/pods/396f1b18-f225-4d67-a98b-87b2605406d4/volumes" Feb 27 07:58:46 crc kubenswrapper[4612]: I0227 07:58:46.026991 4612 patch_prober.go:28] interesting pod/machine-config-daemon-924vb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 27 07:58:46 crc kubenswrapper[4612]: I0227 07:58:46.027606 4612 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 27 07:59:03 crc kubenswrapper[4612]: I0227 07:59:03.460094 4612 scope.go:117] "RemoveContainer" containerID="fe48465b2a2f1e8415bd353b3cca066d922ee14bd35e47161e5458f20f4f96d3" Feb 27 07:59:16 crc kubenswrapper[4612]: I0227 07:59:16.027646 4612 patch_prober.go:28] interesting pod/machine-config-daemon-924vb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 27 07:59:16 crc kubenswrapper[4612]: I0227 07:59:16.029035 4612 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 27 07:59:46 crc kubenswrapper[4612]: I0227 07:59:46.027460 4612 patch_prober.go:28] interesting pod/machine-config-daemon-924vb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 27 07:59:46 crc kubenswrapper[4612]: I0227 07:59:46.028090 4612 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 27 07:59:46 crc kubenswrapper[4612]: I0227 07:59:46.028151 4612 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-924vb" Feb 27 07:59:46 crc kubenswrapper[4612]: I0227 07:59:46.028939 4612 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"778719935b065f111a6a6df70937016b6e803a5b231a05b52794a3dcd1909559"} pod="openshift-machine-config-operator/machine-config-daemon-924vb" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 27 07:59:46 crc kubenswrapper[4612]: I0227 07:59:46.029003 4612 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" containerName="machine-config-daemon" containerID="cri-o://778719935b065f111a6a6df70937016b6e803a5b231a05b52794a3dcd1909559" gracePeriod=600 Feb 27 07:59:46 crc kubenswrapper[4612]: I0227 07:59:46.352759 4612 generic.go:334] "Generic (PLEG): container finished" podID="24599373-7adc-4a1b-8bb4-797bf726f43d" containerID="778719935b065f111a6a6df70937016b6e803a5b231a05b52794a3dcd1909559" exitCode=0 Feb 27 07:59:46 crc kubenswrapper[4612]: I0227 07:59:46.352813 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-924vb" event={"ID":"24599373-7adc-4a1b-8bb4-797bf726f43d","Type":"ContainerDied","Data":"778719935b065f111a6a6df70937016b6e803a5b231a05b52794a3dcd1909559"} Feb 27 07:59:46 crc kubenswrapper[4612]: I0227 07:59:46.353097 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-924vb" event={"ID":"24599373-7adc-4a1b-8bb4-797bf726f43d","Type":"ContainerStarted","Data":"b0982c79ab30408b51bb931dbdd8a46865a7a758c57125e452728efbaccdd1ec"} Feb 27 07:59:46 crc kubenswrapper[4612]: I0227 07:59:46.353120 4612 scope.go:117] "RemoveContainer" containerID="50feff0d2db2ed217d5a022d814ec77b6e80057f575cba0e92f90bcac1d31a76" Feb 27 08:00:00 crc kubenswrapper[4612]: I0227 08:00:00.157059 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29536320-2skmc"] Feb 27 08:00:00 crc kubenswrapper[4612]: E0227 08:00:00.158076 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d76961be-0bf0-45b8-afed-089eaa480485" containerName="oc" Feb 27 08:00:00 crc kubenswrapper[4612]: I0227 08:00:00.158097 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="d76961be-0bf0-45b8-afed-089eaa480485" containerName="oc" Feb 27 08:00:00 crc kubenswrapper[4612]: I0227 08:00:00.158279 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="d76961be-0bf0-45b8-afed-089eaa480485" containerName="oc" Feb 27 08:00:00 crc kubenswrapper[4612]: I0227 08:00:00.158874 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29536320-2skmc" Feb 27 08:00:00 crc kubenswrapper[4612]: I0227 08:00:00.161134 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Feb 27 08:00:00 crc kubenswrapper[4612]: I0227 08:00:00.161428 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Feb 27 08:00:00 crc kubenswrapper[4612]: I0227 08:00:00.171420 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29536320-2skmc"] Feb 27 08:00:00 crc kubenswrapper[4612]: I0227 08:00:00.239315 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d2x75\" (UniqueName: \"kubernetes.io/projected/52c3c89b-5721-4bc5-b1af-c6354236896f-kube-api-access-d2x75\") pod \"collect-profiles-29536320-2skmc\" (UID: \"52c3c89b-5721-4bc5-b1af-c6354236896f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29536320-2skmc" Feb 27 08:00:00 crc kubenswrapper[4612]: I0227 08:00:00.239375 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/52c3c89b-5721-4bc5-b1af-c6354236896f-secret-volume\") pod \"collect-profiles-29536320-2skmc\" (UID: \"52c3c89b-5721-4bc5-b1af-c6354236896f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29536320-2skmc" Feb 27 08:00:00 crc kubenswrapper[4612]: I0227 08:00:00.239640 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/52c3c89b-5721-4bc5-b1af-c6354236896f-config-volume\") pod \"collect-profiles-29536320-2skmc\" (UID: \"52c3c89b-5721-4bc5-b1af-c6354236896f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29536320-2skmc" Feb 27 08:00:00 crc kubenswrapper[4612]: I0227 08:00:00.244256 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29536320-n4vgt"] Feb 27 08:00:00 crc kubenswrapper[4612]: I0227 08:00:00.246470 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536320-n4vgt" Feb 27 08:00:00 crc kubenswrapper[4612]: I0227 08:00:00.250813 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536320-n4vgt"] Feb 27 08:00:00 crc kubenswrapper[4612]: I0227 08:00:00.251260 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 27 08:00:00 crc kubenswrapper[4612]: I0227 08:00:00.251506 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 27 08:00:00 crc kubenswrapper[4612]: I0227 08:00:00.251625 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-5zwfx" Feb 27 08:00:00 crc kubenswrapper[4612]: I0227 08:00:00.340673 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d2x75\" (UniqueName: \"kubernetes.io/projected/52c3c89b-5721-4bc5-b1af-c6354236896f-kube-api-access-d2x75\") pod \"collect-profiles-29536320-2skmc\" (UID: \"52c3c89b-5721-4bc5-b1af-c6354236896f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29536320-2skmc" Feb 27 08:00:00 crc kubenswrapper[4612]: I0227 08:00:00.340805 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/52c3c89b-5721-4bc5-b1af-c6354236896f-secret-volume\") pod \"collect-profiles-29536320-2skmc\" (UID: \"52c3c89b-5721-4bc5-b1af-c6354236896f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29536320-2skmc" Feb 27 08:00:00 crc kubenswrapper[4612]: I0227 08:00:00.340897 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2k5vz\" (UniqueName: \"kubernetes.io/projected/fa6e044a-296c-448c-909f-9b81a8d01f15-kube-api-access-2k5vz\") pod \"auto-csr-approver-29536320-n4vgt\" (UID: \"fa6e044a-296c-448c-909f-9b81a8d01f15\") " pod="openshift-infra/auto-csr-approver-29536320-n4vgt" Feb 27 08:00:00 crc kubenswrapper[4612]: I0227 08:00:00.341096 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/52c3c89b-5721-4bc5-b1af-c6354236896f-config-volume\") pod \"collect-profiles-29536320-2skmc\" (UID: \"52c3c89b-5721-4bc5-b1af-c6354236896f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29536320-2skmc" Feb 27 08:00:00 crc kubenswrapper[4612]: I0227 08:00:00.341975 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/52c3c89b-5721-4bc5-b1af-c6354236896f-config-volume\") pod \"collect-profiles-29536320-2skmc\" (UID: \"52c3c89b-5721-4bc5-b1af-c6354236896f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29536320-2skmc" Feb 27 08:00:00 crc kubenswrapper[4612]: I0227 08:00:00.346787 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/52c3c89b-5721-4bc5-b1af-c6354236896f-secret-volume\") pod \"collect-profiles-29536320-2skmc\" (UID: \"52c3c89b-5721-4bc5-b1af-c6354236896f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29536320-2skmc" Feb 27 08:00:00 crc kubenswrapper[4612]: I0227 08:00:00.358135 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d2x75\" (UniqueName: \"kubernetes.io/projected/52c3c89b-5721-4bc5-b1af-c6354236896f-kube-api-access-d2x75\") pod \"collect-profiles-29536320-2skmc\" (UID: \"52c3c89b-5721-4bc5-b1af-c6354236896f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29536320-2skmc" Feb 27 08:00:00 crc kubenswrapper[4612]: I0227 08:00:00.442464 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2k5vz\" (UniqueName: \"kubernetes.io/projected/fa6e044a-296c-448c-909f-9b81a8d01f15-kube-api-access-2k5vz\") pod \"auto-csr-approver-29536320-n4vgt\" (UID: \"fa6e044a-296c-448c-909f-9b81a8d01f15\") " pod="openshift-infra/auto-csr-approver-29536320-n4vgt" Feb 27 08:00:00 crc kubenswrapper[4612]: I0227 08:00:00.472146 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2k5vz\" (UniqueName: \"kubernetes.io/projected/fa6e044a-296c-448c-909f-9b81a8d01f15-kube-api-access-2k5vz\") pod \"auto-csr-approver-29536320-n4vgt\" (UID: \"fa6e044a-296c-448c-909f-9b81a8d01f15\") " pod="openshift-infra/auto-csr-approver-29536320-n4vgt" Feb 27 08:00:00 crc kubenswrapper[4612]: I0227 08:00:00.492983 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29536320-2skmc" Feb 27 08:00:00 crc kubenswrapper[4612]: I0227 08:00:00.571026 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536320-n4vgt" Feb 27 08:00:00 crc kubenswrapper[4612]: I0227 08:00:00.757812 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536320-n4vgt"] Feb 27 08:00:00 crc kubenswrapper[4612]: I0227 08:00:00.928113 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29536320-2skmc"] Feb 27 08:00:00 crc kubenswrapper[4612]: W0227 08:00:00.932236 4612 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod52c3c89b_5721_4bc5_b1af_c6354236896f.slice/crio-6b41b5de46991a8b10e607db9a7cccca8d922764d1a5b743472390167b66a1f0 WatchSource:0}: Error finding container 6b41b5de46991a8b10e607db9a7cccca8d922764d1a5b743472390167b66a1f0: Status 404 returned error can't find the container with id 6b41b5de46991a8b10e607db9a7cccca8d922764d1a5b743472390167b66a1f0 Feb 27 08:00:01 crc kubenswrapper[4612]: I0227 08:00:01.453396 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536320-n4vgt" event={"ID":"fa6e044a-296c-448c-909f-9b81a8d01f15","Type":"ContainerStarted","Data":"6c3b66d20b96fdb9d2b67ec44e2c3c66931c327853e6b06fd91c718ff9392132"} Feb 27 08:00:01 crc kubenswrapper[4612]: I0227 08:00:01.455035 4612 generic.go:334] "Generic (PLEG): container finished" podID="52c3c89b-5721-4bc5-b1af-c6354236896f" containerID="873b6e2e81a9933f9218fb5e670f8181c4c24427e897138f061f66159a964999" exitCode=0 Feb 27 08:00:01 crc kubenswrapper[4612]: I0227 08:00:01.455064 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29536320-2skmc" event={"ID":"52c3c89b-5721-4bc5-b1af-c6354236896f","Type":"ContainerDied","Data":"873b6e2e81a9933f9218fb5e670f8181c4c24427e897138f061f66159a964999"} Feb 27 08:00:01 crc kubenswrapper[4612]: I0227 08:00:01.455078 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29536320-2skmc" event={"ID":"52c3c89b-5721-4bc5-b1af-c6354236896f","Type":"ContainerStarted","Data":"6b41b5de46991a8b10e607db9a7cccca8d922764d1a5b743472390167b66a1f0"} Feb 27 08:00:02 crc kubenswrapper[4612]: I0227 08:00:02.752162 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29536320-2skmc" Feb 27 08:00:02 crc kubenswrapper[4612]: I0227 08:00:02.878025 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/52c3c89b-5721-4bc5-b1af-c6354236896f-secret-volume\") pod \"52c3c89b-5721-4bc5-b1af-c6354236896f\" (UID: \"52c3c89b-5721-4bc5-b1af-c6354236896f\") " Feb 27 08:00:02 crc kubenswrapper[4612]: I0227 08:00:02.878139 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d2x75\" (UniqueName: \"kubernetes.io/projected/52c3c89b-5721-4bc5-b1af-c6354236896f-kube-api-access-d2x75\") pod \"52c3c89b-5721-4bc5-b1af-c6354236896f\" (UID: \"52c3c89b-5721-4bc5-b1af-c6354236896f\") " Feb 27 08:00:02 crc kubenswrapper[4612]: I0227 08:00:02.878211 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/52c3c89b-5721-4bc5-b1af-c6354236896f-config-volume\") pod \"52c3c89b-5721-4bc5-b1af-c6354236896f\" (UID: \"52c3c89b-5721-4bc5-b1af-c6354236896f\") " Feb 27 08:00:02 crc kubenswrapper[4612]: I0227 08:00:02.879134 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/52c3c89b-5721-4bc5-b1af-c6354236896f-config-volume" (OuterVolumeSpecName: "config-volume") pod "52c3c89b-5721-4bc5-b1af-c6354236896f" (UID: "52c3c89b-5721-4bc5-b1af-c6354236896f"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 08:00:02 crc kubenswrapper[4612]: I0227 08:00:02.885007 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/52c3c89b-5721-4bc5-b1af-c6354236896f-kube-api-access-d2x75" (OuterVolumeSpecName: "kube-api-access-d2x75") pod "52c3c89b-5721-4bc5-b1af-c6354236896f" (UID: "52c3c89b-5721-4bc5-b1af-c6354236896f"). InnerVolumeSpecName "kube-api-access-d2x75". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 08:00:02 crc kubenswrapper[4612]: I0227 08:00:02.885433 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/52c3c89b-5721-4bc5-b1af-c6354236896f-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "52c3c89b-5721-4bc5-b1af-c6354236896f" (UID: "52c3c89b-5721-4bc5-b1af-c6354236896f"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:00:02 crc kubenswrapper[4612]: I0227 08:00:02.979601 4612 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/52c3c89b-5721-4bc5-b1af-c6354236896f-secret-volume\") on node \"crc\" DevicePath \"\"" Feb 27 08:00:02 crc kubenswrapper[4612]: I0227 08:00:02.979648 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d2x75\" (UniqueName: \"kubernetes.io/projected/52c3c89b-5721-4bc5-b1af-c6354236896f-kube-api-access-d2x75\") on node \"crc\" DevicePath \"\"" Feb 27 08:00:02 crc kubenswrapper[4612]: I0227 08:00:02.979661 4612 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/52c3c89b-5721-4bc5-b1af-c6354236896f-config-volume\") on node \"crc\" DevicePath \"\"" Feb 27 08:00:03 crc kubenswrapper[4612]: I0227 08:00:03.468013 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29536320-2skmc" event={"ID":"52c3c89b-5721-4bc5-b1af-c6354236896f","Type":"ContainerDied","Data":"6b41b5de46991a8b10e607db9a7cccca8d922764d1a5b743472390167b66a1f0"} Feb 27 08:00:03 crc kubenswrapper[4612]: I0227 08:00:03.468055 4612 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6b41b5de46991a8b10e607db9a7cccca8d922764d1a5b743472390167b66a1f0" Feb 27 08:00:03 crc kubenswrapper[4612]: I0227 08:00:03.468075 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29536320-2skmc" Feb 27 08:00:14 crc kubenswrapper[4612]: I0227 08:00:14.542657 4612 generic.go:334] "Generic (PLEG): container finished" podID="fa6e044a-296c-448c-909f-9b81a8d01f15" containerID="0f22ee3abd5aaa17b4be1aeea0a2da15f2f6708489f9b954cfb0662bf9fb8a7b" exitCode=0 Feb 27 08:00:14 crc kubenswrapper[4612]: I0227 08:00:14.542772 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536320-n4vgt" event={"ID":"fa6e044a-296c-448c-909f-9b81a8d01f15","Type":"ContainerDied","Data":"0f22ee3abd5aaa17b4be1aeea0a2da15f2f6708489f9b954cfb0662bf9fb8a7b"} Feb 27 08:00:15 crc kubenswrapper[4612]: I0227 08:00:15.758424 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536320-n4vgt" Feb 27 08:00:15 crc kubenswrapper[4612]: I0227 08:00:15.876165 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2k5vz\" (UniqueName: \"kubernetes.io/projected/fa6e044a-296c-448c-909f-9b81a8d01f15-kube-api-access-2k5vz\") pod \"fa6e044a-296c-448c-909f-9b81a8d01f15\" (UID: \"fa6e044a-296c-448c-909f-9b81a8d01f15\") " Feb 27 08:00:15 crc kubenswrapper[4612]: I0227 08:00:15.885301 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fa6e044a-296c-448c-909f-9b81a8d01f15-kube-api-access-2k5vz" (OuterVolumeSpecName: "kube-api-access-2k5vz") pod "fa6e044a-296c-448c-909f-9b81a8d01f15" (UID: "fa6e044a-296c-448c-909f-9b81a8d01f15"). InnerVolumeSpecName "kube-api-access-2k5vz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 08:00:15 crc kubenswrapper[4612]: I0227 08:00:15.978155 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2k5vz\" (UniqueName: \"kubernetes.io/projected/fa6e044a-296c-448c-909f-9b81a8d01f15-kube-api-access-2k5vz\") on node \"crc\" DevicePath \"\"" Feb 27 08:00:16 crc kubenswrapper[4612]: I0227 08:00:16.560923 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536320-n4vgt" event={"ID":"fa6e044a-296c-448c-909f-9b81a8d01f15","Type":"ContainerDied","Data":"6c3b66d20b96fdb9d2b67ec44e2c3c66931c327853e6b06fd91c718ff9392132"} Feb 27 08:00:16 crc kubenswrapper[4612]: I0227 08:00:16.560978 4612 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6c3b66d20b96fdb9d2b67ec44e2c3c66931c327853e6b06fd91c718ff9392132" Feb 27 08:00:16 crc kubenswrapper[4612]: I0227 08:00:16.561006 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536320-n4vgt" Feb 27 08:00:16 crc kubenswrapper[4612]: I0227 08:00:16.834457 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29536314-5bct5"] Feb 27 08:00:16 crc kubenswrapper[4612]: I0227 08:00:16.839117 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29536314-5bct5"] Feb 27 08:00:16 crc kubenswrapper[4612]: I0227 08:00:16.863510 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01363531-9d69-4f11-90f2-d4ee8a952b1a" path="/var/lib/kubelet/pods/01363531-9d69-4f11-90f2-d4ee8a952b1a/volumes" Feb 27 08:01:03 crc kubenswrapper[4612]: I0227 08:01:03.523234 4612 scope.go:117] "RemoveContainer" containerID="be89e07ad393203e501bac58eb8537477d3eb47a0b38a39c6448750bdc8f791a" Feb 27 08:01:13 crc kubenswrapper[4612]: I0227 08:01:13.078830 4612 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Feb 27 08:01:46 crc kubenswrapper[4612]: I0227 08:01:46.027380 4612 patch_prober.go:28] interesting pod/machine-config-daemon-924vb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 27 08:01:46 crc kubenswrapper[4612]: I0227 08:01:46.028647 4612 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 27 08:01:51 crc kubenswrapper[4612]: I0227 08:01:51.391237 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-cf98fcc89-gk4r5"] Feb 27 08:01:51 crc kubenswrapper[4612]: E0227 08:01:51.391706 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa6e044a-296c-448c-909f-9b81a8d01f15" containerName="oc" Feb 27 08:01:51 crc kubenswrapper[4612]: I0227 08:01:51.391720 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa6e044a-296c-448c-909f-9b81a8d01f15" containerName="oc" Feb 27 08:01:51 crc kubenswrapper[4612]: E0227 08:01:51.391749 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="52c3c89b-5721-4bc5-b1af-c6354236896f" containerName="collect-profiles" Feb 27 08:01:51 crc kubenswrapper[4612]: I0227 08:01:51.391757 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="52c3c89b-5721-4bc5-b1af-c6354236896f" containerName="collect-profiles" Feb 27 08:01:51 crc kubenswrapper[4612]: I0227 08:01:51.391863 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="fa6e044a-296c-448c-909f-9b81a8d01f15" containerName="oc" Feb 27 08:01:51 crc kubenswrapper[4612]: I0227 08:01:51.391884 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="52c3c89b-5721-4bc5-b1af-c6354236896f" containerName="collect-profiles" Feb 27 08:01:51 crc kubenswrapper[4612]: I0227 08:01:51.392324 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-cf98fcc89-gk4r5" Feb 27 08:01:51 crc kubenswrapper[4612]: I0227 08:01:51.394544 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Feb 27 08:01:51 crc kubenswrapper[4612]: I0227 08:01:51.395205 4612 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-q2hnr" Feb 27 08:01:51 crc kubenswrapper[4612]: I0227 08:01:51.395276 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Feb 27 08:01:51 crc kubenswrapper[4612]: I0227 08:01:51.411806 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-cf98fcc89-gk4r5"] Feb 27 08:01:51 crc kubenswrapper[4612]: I0227 08:01:51.418261 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-858654f9db-lq868"] Feb 27 08:01:51 crc kubenswrapper[4612]: I0227 08:01:51.419241 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-858654f9db-lq868" Feb 27 08:01:51 crc kubenswrapper[4612]: I0227 08:01:51.421503 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-687f57d79b-krsbj"] Feb 27 08:01:51 crc kubenswrapper[4612]: I0227 08:01:51.422207 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-687f57d79b-krsbj" Feb 27 08:01:51 crc kubenswrapper[4612]: I0227 08:01:51.427247 4612 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-dp9dl" Feb 27 08:01:51 crc kubenswrapper[4612]: I0227 08:01:51.427347 4612 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-q2s27" Feb 27 08:01:51 crc kubenswrapper[4612]: I0227 08:01:51.437288 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-858654f9db-lq868"] Feb 27 08:01:51 crc kubenswrapper[4612]: I0227 08:01:51.445025 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-687f57d79b-krsbj"] Feb 27 08:01:51 crc kubenswrapper[4612]: I0227 08:01:51.479445 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9q2lj\" (UniqueName: \"kubernetes.io/projected/4df7eb22-d428-4e88-bdc5-0bab2b4806a3-kube-api-access-9q2lj\") pod \"cert-manager-cainjector-cf98fcc89-gk4r5\" (UID: \"4df7eb22-d428-4e88-bdc5-0bab2b4806a3\") " pod="cert-manager/cert-manager-cainjector-cf98fcc89-gk4r5" Feb 27 08:01:51 crc kubenswrapper[4612]: I0227 08:01:51.580253 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f5q9k\" (UniqueName: \"kubernetes.io/projected/095685d4-89ee-421e-ba88-a2df75c3a160-kube-api-access-f5q9k\") pod \"cert-manager-858654f9db-lq868\" (UID: \"095685d4-89ee-421e-ba88-a2df75c3a160\") " pod="cert-manager/cert-manager-858654f9db-lq868" Feb 27 08:01:51 crc kubenswrapper[4612]: I0227 08:01:51.580336 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9q2lj\" (UniqueName: \"kubernetes.io/projected/4df7eb22-d428-4e88-bdc5-0bab2b4806a3-kube-api-access-9q2lj\") pod \"cert-manager-cainjector-cf98fcc89-gk4r5\" (UID: \"4df7eb22-d428-4e88-bdc5-0bab2b4806a3\") " pod="cert-manager/cert-manager-cainjector-cf98fcc89-gk4r5" Feb 27 08:01:51 crc kubenswrapper[4612]: I0227 08:01:51.580414 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xnttc\" (UniqueName: \"kubernetes.io/projected/5bba164a-78bf-4a9a-9147-729a1a91740c-kube-api-access-xnttc\") pod \"cert-manager-webhook-687f57d79b-krsbj\" (UID: \"5bba164a-78bf-4a9a-9147-729a1a91740c\") " pod="cert-manager/cert-manager-webhook-687f57d79b-krsbj" Feb 27 08:01:51 crc kubenswrapper[4612]: I0227 08:01:51.598800 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9q2lj\" (UniqueName: \"kubernetes.io/projected/4df7eb22-d428-4e88-bdc5-0bab2b4806a3-kube-api-access-9q2lj\") pod \"cert-manager-cainjector-cf98fcc89-gk4r5\" (UID: \"4df7eb22-d428-4e88-bdc5-0bab2b4806a3\") " pod="cert-manager/cert-manager-cainjector-cf98fcc89-gk4r5" Feb 27 08:01:51 crc kubenswrapper[4612]: I0227 08:01:51.681871 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f5q9k\" (UniqueName: \"kubernetes.io/projected/095685d4-89ee-421e-ba88-a2df75c3a160-kube-api-access-f5q9k\") pod \"cert-manager-858654f9db-lq868\" (UID: \"095685d4-89ee-421e-ba88-a2df75c3a160\") " pod="cert-manager/cert-manager-858654f9db-lq868" Feb 27 08:01:51 crc kubenswrapper[4612]: I0227 08:01:51.681991 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xnttc\" (UniqueName: \"kubernetes.io/projected/5bba164a-78bf-4a9a-9147-729a1a91740c-kube-api-access-xnttc\") pod \"cert-manager-webhook-687f57d79b-krsbj\" (UID: \"5bba164a-78bf-4a9a-9147-729a1a91740c\") " pod="cert-manager/cert-manager-webhook-687f57d79b-krsbj" Feb 27 08:01:51 crc kubenswrapper[4612]: I0227 08:01:51.706563 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-cf98fcc89-gk4r5" Feb 27 08:01:51 crc kubenswrapper[4612]: I0227 08:01:51.712990 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xnttc\" (UniqueName: \"kubernetes.io/projected/5bba164a-78bf-4a9a-9147-729a1a91740c-kube-api-access-xnttc\") pod \"cert-manager-webhook-687f57d79b-krsbj\" (UID: \"5bba164a-78bf-4a9a-9147-729a1a91740c\") " pod="cert-manager/cert-manager-webhook-687f57d79b-krsbj" Feb 27 08:01:51 crc kubenswrapper[4612]: I0227 08:01:51.714943 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f5q9k\" (UniqueName: \"kubernetes.io/projected/095685d4-89ee-421e-ba88-a2df75c3a160-kube-api-access-f5q9k\") pod \"cert-manager-858654f9db-lq868\" (UID: \"095685d4-89ee-421e-ba88-a2df75c3a160\") " pod="cert-manager/cert-manager-858654f9db-lq868" Feb 27 08:01:51 crc kubenswrapper[4612]: I0227 08:01:51.734772 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-858654f9db-lq868" Feb 27 08:01:51 crc kubenswrapper[4612]: I0227 08:01:51.748119 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-687f57d79b-krsbj" Feb 27 08:01:52 crc kubenswrapper[4612]: I0227 08:01:52.127323 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-cf98fcc89-gk4r5"] Feb 27 08:01:52 crc kubenswrapper[4612]: I0227 08:01:52.173633 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-858654f9db-lq868"] Feb 27 08:01:52 crc kubenswrapper[4612]: W0227 08:01:52.177128 4612 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod095685d4_89ee_421e_ba88_a2df75c3a160.slice/crio-544a48cd65ebd5601df1fae8133162a104babd7a91d90300823ec528a2305261 WatchSource:0}: Error finding container 544a48cd65ebd5601df1fae8133162a104babd7a91d90300823ec528a2305261: Status 404 returned error can't find the container with id 544a48cd65ebd5601df1fae8133162a104babd7a91d90300823ec528a2305261 Feb 27 08:01:52 crc kubenswrapper[4612]: I0227 08:01:52.179365 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-687f57d79b-krsbj"] Feb 27 08:01:52 crc kubenswrapper[4612]: W0227 08:01:52.185566 4612 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5bba164a_78bf_4a9a_9147_729a1a91740c.slice/crio-b44b732c14a2edab95a2778d403733dd92ad732df05f9e0851479484eb6393f6 WatchSource:0}: Error finding container b44b732c14a2edab95a2778d403733dd92ad732df05f9e0851479484eb6393f6: Status 404 returned error can't find the container with id b44b732c14a2edab95a2778d403733dd92ad732df05f9e0851479484eb6393f6 Feb 27 08:01:52 crc kubenswrapper[4612]: I0227 08:01:52.353122 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-687f57d79b-krsbj" event={"ID":"5bba164a-78bf-4a9a-9147-729a1a91740c","Type":"ContainerStarted","Data":"b44b732c14a2edab95a2778d403733dd92ad732df05f9e0851479484eb6393f6"} Feb 27 08:01:52 crc kubenswrapper[4612]: I0227 08:01:52.354049 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-858654f9db-lq868" event={"ID":"095685d4-89ee-421e-ba88-a2df75c3a160","Type":"ContainerStarted","Data":"544a48cd65ebd5601df1fae8133162a104babd7a91d90300823ec528a2305261"} Feb 27 08:01:52 crc kubenswrapper[4612]: I0227 08:01:52.354990 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-cf98fcc89-gk4r5" event={"ID":"4df7eb22-d428-4e88-bdc5-0bab2b4806a3","Type":"ContainerStarted","Data":"3e0724fabf999d9dd060f4a5995aad05fdc51c8baf730337da4f822fc88a78b6"} Feb 27 08:01:57 crc kubenswrapper[4612]: I0227 08:01:57.396146 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-cf98fcc89-gk4r5" event={"ID":"4df7eb22-d428-4e88-bdc5-0bab2b4806a3","Type":"ContainerStarted","Data":"737172fbf30b60b9e5375675124bfb5bfe1a36502c1959fb4d1bf2c6a9517287"} Feb 27 08:01:57 crc kubenswrapper[4612]: I0227 08:01:57.399949 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-687f57d79b-krsbj" event={"ID":"5bba164a-78bf-4a9a-9147-729a1a91740c","Type":"ContainerStarted","Data":"9f46b6682ba61559c5846cf8ad30edfea21964c8219704f97f20b97170325083"} Feb 27 08:01:57 crc kubenswrapper[4612]: I0227 08:01:57.400141 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-687f57d79b-krsbj" Feb 27 08:01:57 crc kubenswrapper[4612]: I0227 08:01:57.403092 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-858654f9db-lq868" event={"ID":"095685d4-89ee-421e-ba88-a2df75c3a160","Type":"ContainerStarted","Data":"3dee910408e2e506dfd32bc69e50f342c9db3f5e8603129a247e981587897294"} Feb 27 08:01:57 crc kubenswrapper[4612]: I0227 08:01:57.424456 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-cf98fcc89-gk4r5" podStartSLOduration=2.2764417 podStartE2EDuration="6.424438338s" podCreationTimestamp="2026-02-27 08:01:51 +0000 UTC" firstStartedPulling="2026-02-27 08:01:52.134415541 +0000 UTC m=+769.988345539" lastFinishedPulling="2026-02-27 08:01:56.282412189 +0000 UTC m=+774.136342177" observedRunningTime="2026-02-27 08:01:57.421830743 +0000 UTC m=+775.275760791" watchObservedRunningTime="2026-02-27 08:01:57.424438338 +0000 UTC m=+775.278368326" Feb 27 08:01:57 crc kubenswrapper[4612]: I0227 08:01:57.453288 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-687f57d79b-krsbj" podStartSLOduration=2.353046368 podStartE2EDuration="6.453270695s" podCreationTimestamp="2026-02-27 08:01:51 +0000 UTC" firstStartedPulling="2026-02-27 08:01:52.188876144 +0000 UTC m=+770.042806142" lastFinishedPulling="2026-02-27 08:01:56.289100471 +0000 UTC m=+774.143030469" observedRunningTime="2026-02-27 08:01:57.45238831 +0000 UTC m=+775.306318308" watchObservedRunningTime="2026-02-27 08:01:57.453270695 +0000 UTC m=+775.307200693" Feb 27 08:01:57 crc kubenswrapper[4612]: I0227 08:01:57.472600 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-858654f9db-lq868" podStartSLOduration=2.2960639130000002 podStartE2EDuration="6.472579289s" podCreationTimestamp="2026-02-27 08:01:51 +0000 UTC" firstStartedPulling="2026-02-27 08:01:52.179086573 +0000 UTC m=+770.033016571" lastFinishedPulling="2026-02-27 08:01:56.355601909 +0000 UTC m=+774.209531947" observedRunningTime="2026-02-27 08:01:57.471131217 +0000 UTC m=+775.325061235" watchObservedRunningTime="2026-02-27 08:01:57.472579289 +0000 UTC m=+775.326509307" Feb 27 08:02:00 crc kubenswrapper[4612]: I0227 08:02:00.139913 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29536322-w8jkp"] Feb 27 08:02:00 crc kubenswrapper[4612]: I0227 08:02:00.141877 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536322-w8jkp" Feb 27 08:02:00 crc kubenswrapper[4612]: I0227 08:02:00.144766 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-5zwfx" Feb 27 08:02:00 crc kubenswrapper[4612]: I0227 08:02:00.145291 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 27 08:02:00 crc kubenswrapper[4612]: I0227 08:02:00.148930 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 27 08:02:00 crc kubenswrapper[4612]: I0227 08:02:00.152592 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536322-w8jkp"] Feb 27 08:02:00 crc kubenswrapper[4612]: I0227 08:02:00.297397 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-llzzn\" (UniqueName: \"kubernetes.io/projected/2114ee13-c2af-4646-a086-8491370cb05d-kube-api-access-llzzn\") pod \"auto-csr-approver-29536322-w8jkp\" (UID: \"2114ee13-c2af-4646-a086-8491370cb05d\") " pod="openshift-infra/auto-csr-approver-29536322-w8jkp" Feb 27 08:02:00 crc kubenswrapper[4612]: I0227 08:02:00.399595 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-llzzn\" (UniqueName: \"kubernetes.io/projected/2114ee13-c2af-4646-a086-8491370cb05d-kube-api-access-llzzn\") pod \"auto-csr-approver-29536322-w8jkp\" (UID: \"2114ee13-c2af-4646-a086-8491370cb05d\") " pod="openshift-infra/auto-csr-approver-29536322-w8jkp" Feb 27 08:02:00 crc kubenswrapper[4612]: I0227 08:02:00.436109 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-llzzn\" (UniqueName: \"kubernetes.io/projected/2114ee13-c2af-4646-a086-8491370cb05d-kube-api-access-llzzn\") pod \"auto-csr-approver-29536322-w8jkp\" (UID: \"2114ee13-c2af-4646-a086-8491370cb05d\") " pod="openshift-infra/auto-csr-approver-29536322-w8jkp" Feb 27 08:02:00 crc kubenswrapper[4612]: I0227 08:02:00.500020 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536322-w8jkp" Feb 27 08:02:00 crc kubenswrapper[4612]: I0227 08:02:00.756241 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536322-w8jkp"] Feb 27 08:02:00 crc kubenswrapper[4612]: W0227 08:02:00.764065 4612 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2114ee13_c2af_4646_a086_8491370cb05d.slice/crio-1baaa97756c08c8de8fb236c2812e4fd1485309cef2946ba1ceeb64fc4a4c5da WatchSource:0}: Error finding container 1baaa97756c08c8de8fb236c2812e4fd1485309cef2946ba1ceeb64fc4a4c5da: Status 404 returned error can't find the container with id 1baaa97756c08c8de8fb236c2812e4fd1485309cef2946ba1ceeb64fc4a4c5da Feb 27 08:02:01 crc kubenswrapper[4612]: I0227 08:02:01.436519 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536322-w8jkp" event={"ID":"2114ee13-c2af-4646-a086-8491370cb05d","Type":"ContainerStarted","Data":"1baaa97756c08c8de8fb236c2812e4fd1485309cef2946ba1ceeb64fc4a4c5da"} Feb 27 08:02:01 crc kubenswrapper[4612]: I0227 08:02:01.502021 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-hb4dm"] Feb 27 08:02:01 crc kubenswrapper[4612]: I0227 08:02:01.502451 4612 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-hb4dm" podUID="457f0d99-2681-403b-abc2-92af86fa76e5" containerName="ovn-controller" containerID="cri-o://272efda744ba67b324d44c1a9a0e60d9c2eede817ee274bdfa1a2adcef018517" gracePeriod=30 Feb 27 08:02:01 crc kubenswrapper[4612]: I0227 08:02:01.502513 4612 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-hb4dm" podUID="457f0d99-2681-403b-abc2-92af86fa76e5" containerName="northd" containerID="cri-o://afbd64ee95f7e2f410dc87e9f75d523cc8c37d0567edc382c2d4bd06eaba651b" gracePeriod=30 Feb 27 08:02:01 crc kubenswrapper[4612]: I0227 08:02:01.502545 4612 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-hb4dm" podUID="457f0d99-2681-403b-abc2-92af86fa76e5" containerName="sbdb" containerID="cri-o://53ab32307d4bba133b74665292857b1dad285f3ef223f0758c3752cf1355d4da" gracePeriod=30 Feb 27 08:02:01 crc kubenswrapper[4612]: I0227 08:02:01.502554 4612 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-hb4dm" podUID="457f0d99-2681-403b-abc2-92af86fa76e5" containerName="kube-rbac-proxy-node" containerID="cri-o://d6a8a1c2c4961b0cae8d98041227bad13852d7dab2d4043eb13ebfe7f9e983b8" gracePeriod=30 Feb 27 08:02:01 crc kubenswrapper[4612]: I0227 08:02:01.502678 4612 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-hb4dm" podUID="457f0d99-2681-403b-abc2-92af86fa76e5" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://804ccd844cb7c9fe0448b798d55e5f40e8e4b949c76f7acaeba4be31bc805ea0" gracePeriod=30 Feb 27 08:02:01 crc kubenswrapper[4612]: I0227 08:02:01.502770 4612 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-hb4dm" podUID="457f0d99-2681-403b-abc2-92af86fa76e5" containerName="ovn-acl-logging" containerID="cri-o://3938aacd5fe03026f6a2c84286053cb2e237e9ebe6325bc6d4f1da844329e77e" gracePeriod=30 Feb 27 08:02:01 crc kubenswrapper[4612]: I0227 08:02:01.502882 4612 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-hb4dm" podUID="457f0d99-2681-403b-abc2-92af86fa76e5" containerName="nbdb" containerID="cri-o://f782a058f108da24717810364f65cf60ee65d23c78dbaba7cd8939c96eef7a6e" gracePeriod=30 Feb 27 08:02:01 crc kubenswrapper[4612]: I0227 08:02:01.561808 4612 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-hb4dm" podUID="457f0d99-2681-403b-abc2-92af86fa76e5" containerName="ovnkube-controller" containerID="cri-o://a920896c7db750f9e9ad9cd609c66d12c71145a4b4cc475ae07fbdc1d2694a7b" gracePeriod=30 Feb 27 08:02:01 crc kubenswrapper[4612]: I0227 08:02:01.750336 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-687f57d79b-krsbj" Feb 27 08:02:01 crc kubenswrapper[4612]: I0227 08:02:01.858151 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-hb4dm_457f0d99-2681-403b-abc2-92af86fa76e5/ovn-acl-logging/0.log" Feb 27 08:02:01 crc kubenswrapper[4612]: I0227 08:02:01.858921 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-hb4dm_457f0d99-2681-403b-abc2-92af86fa76e5/ovn-controller/0.log" Feb 27 08:02:01 crc kubenswrapper[4612]: I0227 08:02:01.859401 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-hb4dm" Feb 27 08:02:01 crc kubenswrapper[4612]: I0227 08:02:01.920398 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-ckhjd"] Feb 27 08:02:01 crc kubenswrapper[4612]: E0227 08:02:01.920617 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="457f0d99-2681-403b-abc2-92af86fa76e5" containerName="sbdb" Feb 27 08:02:01 crc kubenswrapper[4612]: I0227 08:02:01.920631 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="457f0d99-2681-403b-abc2-92af86fa76e5" containerName="sbdb" Feb 27 08:02:01 crc kubenswrapper[4612]: E0227 08:02:01.920642 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="457f0d99-2681-403b-abc2-92af86fa76e5" containerName="ovn-acl-logging" Feb 27 08:02:01 crc kubenswrapper[4612]: I0227 08:02:01.920649 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="457f0d99-2681-403b-abc2-92af86fa76e5" containerName="ovn-acl-logging" Feb 27 08:02:01 crc kubenswrapper[4612]: E0227 08:02:01.920661 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="457f0d99-2681-403b-abc2-92af86fa76e5" containerName="northd" Feb 27 08:02:01 crc kubenswrapper[4612]: I0227 08:02:01.920666 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="457f0d99-2681-403b-abc2-92af86fa76e5" containerName="northd" Feb 27 08:02:01 crc kubenswrapper[4612]: E0227 08:02:01.920677 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="457f0d99-2681-403b-abc2-92af86fa76e5" containerName="ovnkube-controller" Feb 27 08:02:01 crc kubenswrapper[4612]: I0227 08:02:01.920684 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="457f0d99-2681-403b-abc2-92af86fa76e5" containerName="ovnkube-controller" Feb 27 08:02:01 crc kubenswrapper[4612]: E0227 08:02:01.921177 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="457f0d99-2681-403b-abc2-92af86fa76e5" containerName="ovn-controller" Feb 27 08:02:01 crc kubenswrapper[4612]: I0227 08:02:01.921186 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="457f0d99-2681-403b-abc2-92af86fa76e5" containerName="ovn-controller" Feb 27 08:02:01 crc kubenswrapper[4612]: E0227 08:02:01.921194 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="457f0d99-2681-403b-abc2-92af86fa76e5" containerName="kube-rbac-proxy-node" Feb 27 08:02:01 crc kubenswrapper[4612]: I0227 08:02:01.921200 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="457f0d99-2681-403b-abc2-92af86fa76e5" containerName="kube-rbac-proxy-node" Feb 27 08:02:01 crc kubenswrapper[4612]: E0227 08:02:01.921207 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="457f0d99-2681-403b-abc2-92af86fa76e5" containerName="kubecfg-setup" Feb 27 08:02:01 crc kubenswrapper[4612]: I0227 08:02:01.921213 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="457f0d99-2681-403b-abc2-92af86fa76e5" containerName="kubecfg-setup" Feb 27 08:02:01 crc kubenswrapper[4612]: E0227 08:02:01.921222 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="457f0d99-2681-403b-abc2-92af86fa76e5" containerName="nbdb" Feb 27 08:02:01 crc kubenswrapper[4612]: I0227 08:02:01.921227 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="457f0d99-2681-403b-abc2-92af86fa76e5" containerName="nbdb" Feb 27 08:02:01 crc kubenswrapper[4612]: E0227 08:02:01.921239 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="457f0d99-2681-403b-abc2-92af86fa76e5" containerName="kube-rbac-proxy-ovn-metrics" Feb 27 08:02:01 crc kubenswrapper[4612]: I0227 08:02:01.921244 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="457f0d99-2681-403b-abc2-92af86fa76e5" containerName="kube-rbac-proxy-ovn-metrics" Feb 27 08:02:01 crc kubenswrapper[4612]: I0227 08:02:01.921351 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="457f0d99-2681-403b-abc2-92af86fa76e5" containerName="kube-rbac-proxy-node" Feb 27 08:02:01 crc kubenswrapper[4612]: I0227 08:02:01.921359 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="457f0d99-2681-403b-abc2-92af86fa76e5" containerName="nbdb" Feb 27 08:02:01 crc kubenswrapper[4612]: I0227 08:02:01.921365 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="457f0d99-2681-403b-abc2-92af86fa76e5" containerName="northd" Feb 27 08:02:01 crc kubenswrapper[4612]: I0227 08:02:01.921373 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="457f0d99-2681-403b-abc2-92af86fa76e5" containerName="sbdb" Feb 27 08:02:01 crc kubenswrapper[4612]: I0227 08:02:01.921382 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="457f0d99-2681-403b-abc2-92af86fa76e5" containerName="kube-rbac-proxy-ovn-metrics" Feb 27 08:02:01 crc kubenswrapper[4612]: I0227 08:02:01.921390 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="457f0d99-2681-403b-abc2-92af86fa76e5" containerName="ovn-controller" Feb 27 08:02:01 crc kubenswrapper[4612]: I0227 08:02:01.921396 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="457f0d99-2681-403b-abc2-92af86fa76e5" containerName="ovnkube-controller" Feb 27 08:02:01 crc kubenswrapper[4612]: I0227 08:02:01.921403 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="457f0d99-2681-403b-abc2-92af86fa76e5" containerName="ovn-acl-logging" Feb 27 08:02:01 crc kubenswrapper[4612]: I0227 08:02:01.923319 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-ckhjd" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.025262 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/457f0d99-2681-403b-abc2-92af86fa76e5-etc-openvswitch\") pod \"457f0d99-2681-403b-abc2-92af86fa76e5\" (UID: \"457f0d99-2681-403b-abc2-92af86fa76e5\") " Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.025325 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/457f0d99-2681-403b-abc2-92af86fa76e5-host-var-lib-cni-networks-ovn-kubernetes\") pod \"457f0d99-2681-403b-abc2-92af86fa76e5\" (UID: \"457f0d99-2681-403b-abc2-92af86fa76e5\") " Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.025351 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/457f0d99-2681-403b-abc2-92af86fa76e5-host-cni-bin\") pod \"457f0d99-2681-403b-abc2-92af86fa76e5\" (UID: \"457f0d99-2681-403b-abc2-92af86fa76e5\") " Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.025376 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/457f0d99-2681-403b-abc2-92af86fa76e5-host-kubelet\") pod \"457f0d99-2681-403b-abc2-92af86fa76e5\" (UID: \"457f0d99-2681-403b-abc2-92af86fa76e5\") " Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.025370 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/457f0d99-2681-403b-abc2-92af86fa76e5-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "457f0d99-2681-403b-abc2-92af86fa76e5" (UID: "457f0d99-2681-403b-abc2-92af86fa76e5"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.025429 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/457f0d99-2681-403b-abc2-92af86fa76e5-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "457f0d99-2681-403b-abc2-92af86fa76e5" (UID: "457f0d99-2681-403b-abc2-92af86fa76e5"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.025443 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/457f0d99-2681-403b-abc2-92af86fa76e5-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "457f0d99-2681-403b-abc2-92af86fa76e5" (UID: "457f0d99-2681-403b-abc2-92af86fa76e5"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.025397 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/457f0d99-2681-403b-abc2-92af86fa76e5-ovnkube-config\") pod \"457f0d99-2681-403b-abc2-92af86fa76e5\" (UID: \"457f0d99-2681-403b-abc2-92af86fa76e5\") " Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.025468 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/457f0d99-2681-403b-abc2-92af86fa76e5-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "457f0d99-2681-403b-abc2-92af86fa76e5" (UID: "457f0d99-2681-403b-abc2-92af86fa76e5"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.025546 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5zllp\" (UniqueName: \"kubernetes.io/projected/457f0d99-2681-403b-abc2-92af86fa76e5-kube-api-access-5zllp\") pod \"457f0d99-2681-403b-abc2-92af86fa76e5\" (UID: \"457f0d99-2681-403b-abc2-92af86fa76e5\") " Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.025603 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/457f0d99-2681-403b-abc2-92af86fa76e5-node-log\") pod \"457f0d99-2681-403b-abc2-92af86fa76e5\" (UID: \"457f0d99-2681-403b-abc2-92af86fa76e5\") " Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.025629 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/457f0d99-2681-403b-abc2-92af86fa76e5-run-systemd\") pod \"457f0d99-2681-403b-abc2-92af86fa76e5\" (UID: \"457f0d99-2681-403b-abc2-92af86fa76e5\") " Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.025675 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/457f0d99-2681-403b-abc2-92af86fa76e5-node-log" (OuterVolumeSpecName: "node-log") pod "457f0d99-2681-403b-abc2-92af86fa76e5" (UID: "457f0d99-2681-403b-abc2-92af86fa76e5"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.025685 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/457f0d99-2681-403b-abc2-92af86fa76e5-ovnkube-script-lib\") pod \"457f0d99-2681-403b-abc2-92af86fa76e5\" (UID: \"457f0d99-2681-403b-abc2-92af86fa76e5\") " Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.025725 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/457f0d99-2681-403b-abc2-92af86fa76e5-run-ovn\") pod \"457f0d99-2681-403b-abc2-92af86fa76e5\" (UID: \"457f0d99-2681-403b-abc2-92af86fa76e5\") " Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.025772 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/457f0d99-2681-403b-abc2-92af86fa76e5-run-openvswitch\") pod \"457f0d99-2681-403b-abc2-92af86fa76e5\" (UID: \"457f0d99-2681-403b-abc2-92af86fa76e5\") " Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.025787 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/457f0d99-2681-403b-abc2-92af86fa76e5-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "457f0d99-2681-403b-abc2-92af86fa76e5" (UID: "457f0d99-2681-403b-abc2-92af86fa76e5"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.025799 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/457f0d99-2681-403b-abc2-92af86fa76e5-host-slash\") pod \"457f0d99-2681-403b-abc2-92af86fa76e5\" (UID: \"457f0d99-2681-403b-abc2-92af86fa76e5\") " Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.025825 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/457f0d99-2681-403b-abc2-92af86fa76e5-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "457f0d99-2681-403b-abc2-92af86fa76e5" (UID: "457f0d99-2681-403b-abc2-92af86fa76e5"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.025822 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/457f0d99-2681-403b-abc2-92af86fa76e5-host-run-ovn-kubernetes\") pod \"457f0d99-2681-403b-abc2-92af86fa76e5\" (UID: \"457f0d99-2681-403b-abc2-92af86fa76e5\") " Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.025883 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/457f0d99-2681-403b-abc2-92af86fa76e5-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "457f0d99-2681-403b-abc2-92af86fa76e5" (UID: "457f0d99-2681-403b-abc2-92af86fa76e5"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.025889 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/457f0d99-2681-403b-abc2-92af86fa76e5-host-run-netns\") pod \"457f0d99-2681-403b-abc2-92af86fa76e5\" (UID: \"457f0d99-2681-403b-abc2-92af86fa76e5\") " Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.025929 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/457f0d99-2681-403b-abc2-92af86fa76e5-systemd-units\") pod \"457f0d99-2681-403b-abc2-92af86fa76e5\" (UID: \"457f0d99-2681-403b-abc2-92af86fa76e5\") " Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.025966 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/457f0d99-2681-403b-abc2-92af86fa76e5-ovn-node-metrics-cert\") pod \"457f0d99-2681-403b-abc2-92af86fa76e5\" (UID: \"457f0d99-2681-403b-abc2-92af86fa76e5\") " Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.026032 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/457f0d99-2681-403b-abc2-92af86fa76e5-host-cni-netd\") pod \"457f0d99-2681-403b-abc2-92af86fa76e5\" (UID: \"457f0d99-2681-403b-abc2-92af86fa76e5\") " Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.026057 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/457f0d99-2681-403b-abc2-92af86fa76e5-var-lib-openvswitch\") pod \"457f0d99-2681-403b-abc2-92af86fa76e5\" (UID: \"457f0d99-2681-403b-abc2-92af86fa76e5\") " Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.026087 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/457f0d99-2681-403b-abc2-92af86fa76e5-env-overrides\") pod \"457f0d99-2681-403b-abc2-92af86fa76e5\" (UID: \"457f0d99-2681-403b-abc2-92af86fa76e5\") " Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.026113 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/457f0d99-2681-403b-abc2-92af86fa76e5-log-socket\") pod \"457f0d99-2681-403b-abc2-92af86fa76e5\" (UID: \"457f0d99-2681-403b-abc2-92af86fa76e5\") " Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.025905 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/457f0d99-2681-403b-abc2-92af86fa76e5-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "457f0d99-2681-403b-abc2-92af86fa76e5" (UID: "457f0d99-2681-403b-abc2-92af86fa76e5"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.026129 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/457f0d99-2681-403b-abc2-92af86fa76e5-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "457f0d99-2681-403b-abc2-92af86fa76e5" (UID: "457f0d99-2681-403b-abc2-92af86fa76e5"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.026151 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/457f0d99-2681-403b-abc2-92af86fa76e5-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "457f0d99-2681-403b-abc2-92af86fa76e5" (UID: "457f0d99-2681-403b-abc2-92af86fa76e5"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.026166 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/457f0d99-2681-403b-abc2-92af86fa76e5-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "457f0d99-2681-403b-abc2-92af86fa76e5" (UID: "457f0d99-2681-403b-abc2-92af86fa76e5"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.026179 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/457f0d99-2681-403b-abc2-92af86fa76e5-host-slash" (OuterVolumeSpecName: "host-slash") pod "457f0d99-2681-403b-abc2-92af86fa76e5" (UID: "457f0d99-2681-403b-abc2-92af86fa76e5"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.026220 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/457f0d99-2681-403b-abc2-92af86fa76e5-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "457f0d99-2681-403b-abc2-92af86fa76e5" (UID: "457f0d99-2681-403b-abc2-92af86fa76e5"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.026342 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/9c746ad6-c797-40c9-8199-0a84a258fed3-host-slash\") pod \"ovnkube-node-ckhjd\" (UID: \"9c746ad6-c797-40c9-8199-0a84a258fed3\") " pod="openshift-ovn-kubernetes/ovnkube-node-ckhjd" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.026349 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/457f0d99-2681-403b-abc2-92af86fa76e5-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "457f0d99-2681-403b-abc2-92af86fa76e5" (UID: "457f0d99-2681-403b-abc2-92af86fa76e5"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.026381 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/457f0d99-2681-403b-abc2-92af86fa76e5-log-socket" (OuterVolumeSpecName: "log-socket") pod "457f0d99-2681-403b-abc2-92af86fa76e5" (UID: "457f0d99-2681-403b-abc2-92af86fa76e5"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.026412 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/9c746ad6-c797-40c9-8199-0a84a258fed3-host-run-netns\") pod \"ovnkube-node-ckhjd\" (UID: \"9c746ad6-c797-40c9-8199-0a84a258fed3\") " pod="openshift-ovn-kubernetes/ovnkube-node-ckhjd" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.026440 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/9c746ad6-c797-40c9-8199-0a84a258fed3-run-systemd\") pod \"ovnkube-node-ckhjd\" (UID: \"9c746ad6-c797-40c9-8199-0a84a258fed3\") " pod="openshift-ovn-kubernetes/ovnkube-node-ckhjd" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.026466 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9c746ad6-c797-40c9-8199-0a84a258fed3-var-lib-openvswitch\") pod \"ovnkube-node-ckhjd\" (UID: \"9c746ad6-c797-40c9-8199-0a84a258fed3\") " pod="openshift-ovn-kubernetes/ovnkube-node-ckhjd" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.026559 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/9c746ad6-c797-40c9-8199-0a84a258fed3-ovn-node-metrics-cert\") pod \"ovnkube-node-ckhjd\" (UID: \"9c746ad6-c797-40c9-8199-0a84a258fed3\") " pod="openshift-ovn-kubernetes/ovnkube-node-ckhjd" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.026656 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/9c746ad6-c797-40c9-8199-0a84a258fed3-run-ovn\") pod \"ovnkube-node-ckhjd\" (UID: \"9c746ad6-c797-40c9-8199-0a84a258fed3\") " pod="openshift-ovn-kubernetes/ovnkube-node-ckhjd" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.026664 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/457f0d99-2681-403b-abc2-92af86fa76e5-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "457f0d99-2681-403b-abc2-92af86fa76e5" (UID: "457f0d99-2681-403b-abc2-92af86fa76e5"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.026730 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c2xrw\" (UniqueName: \"kubernetes.io/projected/9c746ad6-c797-40c9-8199-0a84a258fed3-kube-api-access-c2xrw\") pod \"ovnkube-node-ckhjd\" (UID: \"9c746ad6-c797-40c9-8199-0a84a258fed3\") " pod="openshift-ovn-kubernetes/ovnkube-node-ckhjd" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.026764 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/9c746ad6-c797-40c9-8199-0a84a258fed3-host-cni-bin\") pod \"ovnkube-node-ckhjd\" (UID: \"9c746ad6-c797-40c9-8199-0a84a258fed3\") " pod="openshift-ovn-kubernetes/ovnkube-node-ckhjd" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.026804 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/9c746ad6-c797-40c9-8199-0a84a258fed3-host-kubelet\") pod \"ovnkube-node-ckhjd\" (UID: \"9c746ad6-c797-40c9-8199-0a84a258fed3\") " pod="openshift-ovn-kubernetes/ovnkube-node-ckhjd" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.026833 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9c746ad6-c797-40c9-8199-0a84a258fed3-run-openvswitch\") pod \"ovnkube-node-ckhjd\" (UID: \"9c746ad6-c797-40c9-8199-0a84a258fed3\") " pod="openshift-ovn-kubernetes/ovnkube-node-ckhjd" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.026865 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9c746ad6-c797-40c9-8199-0a84a258fed3-host-run-ovn-kubernetes\") pod \"ovnkube-node-ckhjd\" (UID: \"9c746ad6-c797-40c9-8199-0a84a258fed3\") " pod="openshift-ovn-kubernetes/ovnkube-node-ckhjd" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.026898 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9c746ad6-c797-40c9-8199-0a84a258fed3-etc-openvswitch\") pod \"ovnkube-node-ckhjd\" (UID: \"9c746ad6-c797-40c9-8199-0a84a258fed3\") " pod="openshift-ovn-kubernetes/ovnkube-node-ckhjd" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.026930 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/9c746ad6-c797-40c9-8199-0a84a258fed3-ovnkube-config\") pod \"ovnkube-node-ckhjd\" (UID: \"9c746ad6-c797-40c9-8199-0a84a258fed3\") " pod="openshift-ovn-kubernetes/ovnkube-node-ckhjd" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.026976 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/9c746ad6-c797-40c9-8199-0a84a258fed3-ovnkube-script-lib\") pod \"ovnkube-node-ckhjd\" (UID: \"9c746ad6-c797-40c9-8199-0a84a258fed3\") " pod="openshift-ovn-kubernetes/ovnkube-node-ckhjd" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.027005 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/9c746ad6-c797-40c9-8199-0a84a258fed3-systemd-units\") pod \"ovnkube-node-ckhjd\" (UID: \"9c746ad6-c797-40c9-8199-0a84a258fed3\") " pod="openshift-ovn-kubernetes/ovnkube-node-ckhjd" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.027048 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9c746ad6-c797-40c9-8199-0a84a258fed3-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-ckhjd\" (UID: \"9c746ad6-c797-40c9-8199-0a84a258fed3\") " pod="openshift-ovn-kubernetes/ovnkube-node-ckhjd" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.027077 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/9c746ad6-c797-40c9-8199-0a84a258fed3-node-log\") pod \"ovnkube-node-ckhjd\" (UID: \"9c746ad6-c797-40c9-8199-0a84a258fed3\") " pod="openshift-ovn-kubernetes/ovnkube-node-ckhjd" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.027104 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/9c746ad6-c797-40c9-8199-0a84a258fed3-log-socket\") pod \"ovnkube-node-ckhjd\" (UID: \"9c746ad6-c797-40c9-8199-0a84a258fed3\") " pod="openshift-ovn-kubernetes/ovnkube-node-ckhjd" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.027126 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/9c746ad6-c797-40c9-8199-0a84a258fed3-host-cni-netd\") pod \"ovnkube-node-ckhjd\" (UID: \"9c746ad6-c797-40c9-8199-0a84a258fed3\") " pod="openshift-ovn-kubernetes/ovnkube-node-ckhjd" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.027155 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/9c746ad6-c797-40c9-8199-0a84a258fed3-env-overrides\") pod \"ovnkube-node-ckhjd\" (UID: \"9c746ad6-c797-40c9-8199-0a84a258fed3\") " pod="openshift-ovn-kubernetes/ovnkube-node-ckhjd" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.027289 4612 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/457f0d99-2681-403b-abc2-92af86fa76e5-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.027326 4612 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/457f0d99-2681-403b-abc2-92af86fa76e5-host-cni-bin\") on node \"crc\" DevicePath \"\"" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.027342 4612 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/457f0d99-2681-403b-abc2-92af86fa76e5-host-kubelet\") on node \"crc\" DevicePath \"\"" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.027354 4612 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/457f0d99-2681-403b-abc2-92af86fa76e5-ovnkube-config\") on node \"crc\" DevicePath \"\"" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.027366 4612 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/457f0d99-2681-403b-abc2-92af86fa76e5-node-log\") on node \"crc\" DevicePath \"\"" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.027380 4612 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/457f0d99-2681-403b-abc2-92af86fa76e5-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.027392 4612 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/457f0d99-2681-403b-abc2-92af86fa76e5-run-ovn\") on node \"crc\" DevicePath \"\"" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.027404 4612 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/457f0d99-2681-403b-abc2-92af86fa76e5-run-openvswitch\") on node \"crc\" DevicePath \"\"" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.027416 4612 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/457f0d99-2681-403b-abc2-92af86fa76e5-host-slash\") on node \"crc\" DevicePath \"\"" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.027428 4612 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/457f0d99-2681-403b-abc2-92af86fa76e5-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.027440 4612 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/457f0d99-2681-403b-abc2-92af86fa76e5-host-run-netns\") on node \"crc\" DevicePath \"\"" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.027453 4612 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/457f0d99-2681-403b-abc2-92af86fa76e5-systemd-units\") on node \"crc\" DevicePath \"\"" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.027465 4612 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/457f0d99-2681-403b-abc2-92af86fa76e5-host-cni-netd\") on node \"crc\" DevicePath \"\"" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.027477 4612 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/457f0d99-2681-403b-abc2-92af86fa76e5-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.027489 4612 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/457f0d99-2681-403b-abc2-92af86fa76e5-env-overrides\") on node \"crc\" DevicePath \"\"" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.027500 4612 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/457f0d99-2681-403b-abc2-92af86fa76e5-log-socket\") on node \"crc\" DevicePath \"\"" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.027513 4612 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/457f0d99-2681-403b-abc2-92af86fa76e5-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.030241 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/457f0d99-2681-403b-abc2-92af86fa76e5-kube-api-access-5zllp" (OuterVolumeSpecName: "kube-api-access-5zllp") pod "457f0d99-2681-403b-abc2-92af86fa76e5" (UID: "457f0d99-2681-403b-abc2-92af86fa76e5"). InnerVolumeSpecName "kube-api-access-5zllp". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.031736 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/457f0d99-2681-403b-abc2-92af86fa76e5-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "457f0d99-2681-403b-abc2-92af86fa76e5" (UID: "457f0d99-2681-403b-abc2-92af86fa76e5"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.041078 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/457f0d99-2681-403b-abc2-92af86fa76e5-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "457f0d99-2681-403b-abc2-92af86fa76e5" (UID: "457f0d99-2681-403b-abc2-92af86fa76e5"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.128749 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/9c746ad6-c797-40c9-8199-0a84a258fed3-ovn-node-metrics-cert\") pod \"ovnkube-node-ckhjd\" (UID: \"9c746ad6-c797-40c9-8199-0a84a258fed3\") " pod="openshift-ovn-kubernetes/ovnkube-node-ckhjd" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.128813 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/9c746ad6-c797-40c9-8199-0a84a258fed3-run-ovn\") pod \"ovnkube-node-ckhjd\" (UID: \"9c746ad6-c797-40c9-8199-0a84a258fed3\") " pod="openshift-ovn-kubernetes/ovnkube-node-ckhjd" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.128841 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c2xrw\" (UniqueName: \"kubernetes.io/projected/9c746ad6-c797-40c9-8199-0a84a258fed3-kube-api-access-c2xrw\") pod \"ovnkube-node-ckhjd\" (UID: \"9c746ad6-c797-40c9-8199-0a84a258fed3\") " pod="openshift-ovn-kubernetes/ovnkube-node-ckhjd" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.128867 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/9c746ad6-c797-40c9-8199-0a84a258fed3-host-cni-bin\") pod \"ovnkube-node-ckhjd\" (UID: \"9c746ad6-c797-40c9-8199-0a84a258fed3\") " pod="openshift-ovn-kubernetes/ovnkube-node-ckhjd" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.128889 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/9c746ad6-c797-40c9-8199-0a84a258fed3-host-kubelet\") pod \"ovnkube-node-ckhjd\" (UID: \"9c746ad6-c797-40c9-8199-0a84a258fed3\") " pod="openshift-ovn-kubernetes/ovnkube-node-ckhjd" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.128910 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9c746ad6-c797-40c9-8199-0a84a258fed3-run-openvswitch\") pod \"ovnkube-node-ckhjd\" (UID: \"9c746ad6-c797-40c9-8199-0a84a258fed3\") " pod="openshift-ovn-kubernetes/ovnkube-node-ckhjd" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.128932 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9c746ad6-c797-40c9-8199-0a84a258fed3-host-run-ovn-kubernetes\") pod \"ovnkube-node-ckhjd\" (UID: \"9c746ad6-c797-40c9-8199-0a84a258fed3\") " pod="openshift-ovn-kubernetes/ovnkube-node-ckhjd" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.128954 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9c746ad6-c797-40c9-8199-0a84a258fed3-etc-openvswitch\") pod \"ovnkube-node-ckhjd\" (UID: \"9c746ad6-c797-40c9-8199-0a84a258fed3\") " pod="openshift-ovn-kubernetes/ovnkube-node-ckhjd" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.128977 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/9c746ad6-c797-40c9-8199-0a84a258fed3-ovnkube-config\") pod \"ovnkube-node-ckhjd\" (UID: \"9c746ad6-c797-40c9-8199-0a84a258fed3\") " pod="openshift-ovn-kubernetes/ovnkube-node-ckhjd" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.129001 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/9c746ad6-c797-40c9-8199-0a84a258fed3-ovnkube-script-lib\") pod \"ovnkube-node-ckhjd\" (UID: \"9c746ad6-c797-40c9-8199-0a84a258fed3\") " pod="openshift-ovn-kubernetes/ovnkube-node-ckhjd" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.129024 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/9c746ad6-c797-40c9-8199-0a84a258fed3-systemd-units\") pod \"ovnkube-node-ckhjd\" (UID: \"9c746ad6-c797-40c9-8199-0a84a258fed3\") " pod="openshift-ovn-kubernetes/ovnkube-node-ckhjd" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.129051 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9c746ad6-c797-40c9-8199-0a84a258fed3-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-ckhjd\" (UID: \"9c746ad6-c797-40c9-8199-0a84a258fed3\") " pod="openshift-ovn-kubernetes/ovnkube-node-ckhjd" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.129073 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/9c746ad6-c797-40c9-8199-0a84a258fed3-node-log\") pod \"ovnkube-node-ckhjd\" (UID: \"9c746ad6-c797-40c9-8199-0a84a258fed3\") " pod="openshift-ovn-kubernetes/ovnkube-node-ckhjd" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.129091 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/9c746ad6-c797-40c9-8199-0a84a258fed3-log-socket\") pod \"ovnkube-node-ckhjd\" (UID: \"9c746ad6-c797-40c9-8199-0a84a258fed3\") " pod="openshift-ovn-kubernetes/ovnkube-node-ckhjd" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.129111 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/9c746ad6-c797-40c9-8199-0a84a258fed3-host-cni-netd\") pod \"ovnkube-node-ckhjd\" (UID: \"9c746ad6-c797-40c9-8199-0a84a258fed3\") " pod="openshift-ovn-kubernetes/ovnkube-node-ckhjd" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.129130 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/9c746ad6-c797-40c9-8199-0a84a258fed3-env-overrides\") pod \"ovnkube-node-ckhjd\" (UID: \"9c746ad6-c797-40c9-8199-0a84a258fed3\") " pod="openshift-ovn-kubernetes/ovnkube-node-ckhjd" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.129153 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/9c746ad6-c797-40c9-8199-0a84a258fed3-host-slash\") pod \"ovnkube-node-ckhjd\" (UID: \"9c746ad6-c797-40c9-8199-0a84a258fed3\") " pod="openshift-ovn-kubernetes/ovnkube-node-ckhjd" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.129180 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/9c746ad6-c797-40c9-8199-0a84a258fed3-host-run-netns\") pod \"ovnkube-node-ckhjd\" (UID: \"9c746ad6-c797-40c9-8199-0a84a258fed3\") " pod="openshift-ovn-kubernetes/ovnkube-node-ckhjd" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.129199 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/9c746ad6-c797-40c9-8199-0a84a258fed3-run-systemd\") pod \"ovnkube-node-ckhjd\" (UID: \"9c746ad6-c797-40c9-8199-0a84a258fed3\") " pod="openshift-ovn-kubernetes/ovnkube-node-ckhjd" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.129218 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9c746ad6-c797-40c9-8199-0a84a258fed3-var-lib-openvswitch\") pod \"ovnkube-node-ckhjd\" (UID: \"9c746ad6-c797-40c9-8199-0a84a258fed3\") " pod="openshift-ovn-kubernetes/ovnkube-node-ckhjd" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.129264 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5zllp\" (UniqueName: \"kubernetes.io/projected/457f0d99-2681-403b-abc2-92af86fa76e5-kube-api-access-5zllp\") on node \"crc\" DevicePath \"\"" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.129279 4612 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/457f0d99-2681-403b-abc2-92af86fa76e5-run-systemd\") on node \"crc\" DevicePath \"\"" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.129291 4612 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/457f0d99-2681-403b-abc2-92af86fa76e5-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.129337 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9c746ad6-c797-40c9-8199-0a84a258fed3-var-lib-openvswitch\") pod \"ovnkube-node-ckhjd\" (UID: \"9c746ad6-c797-40c9-8199-0a84a258fed3\") " pod="openshift-ovn-kubernetes/ovnkube-node-ckhjd" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.129925 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/9c746ad6-c797-40c9-8199-0a84a258fed3-run-ovn\") pod \"ovnkube-node-ckhjd\" (UID: \"9c746ad6-c797-40c9-8199-0a84a258fed3\") " pod="openshift-ovn-kubernetes/ovnkube-node-ckhjd" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.130034 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/9c746ad6-c797-40c9-8199-0a84a258fed3-systemd-units\") pod \"ovnkube-node-ckhjd\" (UID: \"9c746ad6-c797-40c9-8199-0a84a258fed3\") " pod="openshift-ovn-kubernetes/ovnkube-node-ckhjd" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.130092 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9c746ad6-c797-40c9-8199-0a84a258fed3-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-ckhjd\" (UID: \"9c746ad6-c797-40c9-8199-0a84a258fed3\") " pod="openshift-ovn-kubernetes/ovnkube-node-ckhjd" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.130137 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/9c746ad6-c797-40c9-8199-0a84a258fed3-node-log\") pod \"ovnkube-node-ckhjd\" (UID: \"9c746ad6-c797-40c9-8199-0a84a258fed3\") " pod="openshift-ovn-kubernetes/ovnkube-node-ckhjd" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.130184 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/9c746ad6-c797-40c9-8199-0a84a258fed3-log-socket\") pod \"ovnkube-node-ckhjd\" (UID: \"9c746ad6-c797-40c9-8199-0a84a258fed3\") " pod="openshift-ovn-kubernetes/ovnkube-node-ckhjd" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.130233 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/9c746ad6-c797-40c9-8199-0a84a258fed3-host-cni-netd\") pod \"ovnkube-node-ckhjd\" (UID: \"9c746ad6-c797-40c9-8199-0a84a258fed3\") " pod="openshift-ovn-kubernetes/ovnkube-node-ckhjd" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.130297 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9c746ad6-c797-40c9-8199-0a84a258fed3-etc-openvswitch\") pod \"ovnkube-node-ckhjd\" (UID: \"9c746ad6-c797-40c9-8199-0a84a258fed3\") " pod="openshift-ovn-kubernetes/ovnkube-node-ckhjd" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.130316 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9c746ad6-c797-40c9-8199-0a84a258fed3-host-run-ovn-kubernetes\") pod \"ovnkube-node-ckhjd\" (UID: \"9c746ad6-c797-40c9-8199-0a84a258fed3\") " pod="openshift-ovn-kubernetes/ovnkube-node-ckhjd" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.130404 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/9c746ad6-c797-40c9-8199-0a84a258fed3-host-cni-bin\") pod \"ovnkube-node-ckhjd\" (UID: \"9c746ad6-c797-40c9-8199-0a84a258fed3\") " pod="openshift-ovn-kubernetes/ovnkube-node-ckhjd" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.130438 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/9c746ad6-c797-40c9-8199-0a84a258fed3-host-kubelet\") pod \"ovnkube-node-ckhjd\" (UID: \"9c746ad6-c797-40c9-8199-0a84a258fed3\") " pod="openshift-ovn-kubernetes/ovnkube-node-ckhjd" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.130467 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9c746ad6-c797-40c9-8199-0a84a258fed3-run-openvswitch\") pod \"ovnkube-node-ckhjd\" (UID: \"9c746ad6-c797-40c9-8199-0a84a258fed3\") " pod="openshift-ovn-kubernetes/ovnkube-node-ckhjd" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.130648 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/9c746ad6-c797-40c9-8199-0a84a258fed3-host-slash\") pod \"ovnkube-node-ckhjd\" (UID: \"9c746ad6-c797-40c9-8199-0a84a258fed3\") " pod="openshift-ovn-kubernetes/ovnkube-node-ckhjd" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.130686 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/9c746ad6-c797-40c9-8199-0a84a258fed3-run-systemd\") pod \"ovnkube-node-ckhjd\" (UID: \"9c746ad6-c797-40c9-8199-0a84a258fed3\") " pod="openshift-ovn-kubernetes/ovnkube-node-ckhjd" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.131106 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/9c746ad6-c797-40c9-8199-0a84a258fed3-env-overrides\") pod \"ovnkube-node-ckhjd\" (UID: \"9c746ad6-c797-40c9-8199-0a84a258fed3\") " pod="openshift-ovn-kubernetes/ovnkube-node-ckhjd" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.131217 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/9c746ad6-c797-40c9-8199-0a84a258fed3-ovnkube-script-lib\") pod \"ovnkube-node-ckhjd\" (UID: \"9c746ad6-c797-40c9-8199-0a84a258fed3\") " pod="openshift-ovn-kubernetes/ovnkube-node-ckhjd" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.131289 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/9c746ad6-c797-40c9-8199-0a84a258fed3-host-run-netns\") pod \"ovnkube-node-ckhjd\" (UID: \"9c746ad6-c797-40c9-8199-0a84a258fed3\") " pod="openshift-ovn-kubernetes/ovnkube-node-ckhjd" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.131799 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/9c746ad6-c797-40c9-8199-0a84a258fed3-ovnkube-config\") pod \"ovnkube-node-ckhjd\" (UID: \"9c746ad6-c797-40c9-8199-0a84a258fed3\") " pod="openshift-ovn-kubernetes/ovnkube-node-ckhjd" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.133260 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/9c746ad6-c797-40c9-8199-0a84a258fed3-ovn-node-metrics-cert\") pod \"ovnkube-node-ckhjd\" (UID: \"9c746ad6-c797-40c9-8199-0a84a258fed3\") " pod="openshift-ovn-kubernetes/ovnkube-node-ckhjd" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.146780 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c2xrw\" (UniqueName: \"kubernetes.io/projected/9c746ad6-c797-40c9-8199-0a84a258fed3-kube-api-access-c2xrw\") pod \"ovnkube-node-ckhjd\" (UID: \"9c746ad6-c797-40c9-8199-0a84a258fed3\") " pod="openshift-ovn-kubernetes/ovnkube-node-ckhjd" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.242628 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-ckhjd" Feb 27 08:02:02 crc kubenswrapper[4612]: W0227 08:02:02.262229 4612 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9c746ad6_c797_40c9_8199_0a84a258fed3.slice/crio-8a4d2c06eec493089b56f486217eceb565b0b524c5c0fd1b5140b931ff509505 WatchSource:0}: Error finding container 8a4d2c06eec493089b56f486217eceb565b0b524c5c0fd1b5140b931ff509505: Status 404 returned error can't find the container with id 8a4d2c06eec493089b56f486217eceb565b0b524c5c0fd1b5140b931ff509505 Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.447016 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-hb4dm_457f0d99-2681-403b-abc2-92af86fa76e5/ovn-acl-logging/0.log" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.448335 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-hb4dm_457f0d99-2681-403b-abc2-92af86fa76e5/ovn-controller/0.log" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.448987 4612 generic.go:334] "Generic (PLEG): container finished" podID="457f0d99-2681-403b-abc2-92af86fa76e5" containerID="a920896c7db750f9e9ad9cd609c66d12c71145a4b4cc475ae07fbdc1d2694a7b" exitCode=0 Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.449013 4612 generic.go:334] "Generic (PLEG): container finished" podID="457f0d99-2681-403b-abc2-92af86fa76e5" containerID="53ab32307d4bba133b74665292857b1dad285f3ef223f0758c3752cf1355d4da" exitCode=0 Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.449021 4612 generic.go:334] "Generic (PLEG): container finished" podID="457f0d99-2681-403b-abc2-92af86fa76e5" containerID="f782a058f108da24717810364f65cf60ee65d23c78dbaba7cd8939c96eef7a6e" exitCode=0 Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.449028 4612 generic.go:334] "Generic (PLEG): container finished" podID="457f0d99-2681-403b-abc2-92af86fa76e5" containerID="afbd64ee95f7e2f410dc87e9f75d523cc8c37d0567edc382c2d4bd06eaba651b" exitCode=0 Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.449035 4612 generic.go:334] "Generic (PLEG): container finished" podID="457f0d99-2681-403b-abc2-92af86fa76e5" containerID="804ccd844cb7c9fe0448b798d55e5f40e8e4b949c76f7acaeba4be31bc805ea0" exitCode=0 Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.449042 4612 generic.go:334] "Generic (PLEG): container finished" podID="457f0d99-2681-403b-abc2-92af86fa76e5" containerID="d6a8a1c2c4961b0cae8d98041227bad13852d7dab2d4043eb13ebfe7f9e983b8" exitCode=0 Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.449049 4612 generic.go:334] "Generic (PLEG): container finished" podID="457f0d99-2681-403b-abc2-92af86fa76e5" containerID="3938aacd5fe03026f6a2c84286053cb2e237e9ebe6325bc6d4f1da844329e77e" exitCode=143 Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.449059 4612 generic.go:334] "Generic (PLEG): container finished" podID="457f0d99-2681-403b-abc2-92af86fa76e5" containerID="272efda744ba67b324d44c1a9a0e60d9c2eede817ee274bdfa1a2adcef018517" exitCode=143 Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.449191 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-hb4dm" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.449371 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hb4dm" event={"ID":"457f0d99-2681-403b-abc2-92af86fa76e5","Type":"ContainerDied","Data":"a920896c7db750f9e9ad9cd609c66d12c71145a4b4cc475ae07fbdc1d2694a7b"} Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.449418 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hb4dm" event={"ID":"457f0d99-2681-403b-abc2-92af86fa76e5","Type":"ContainerDied","Data":"53ab32307d4bba133b74665292857b1dad285f3ef223f0758c3752cf1355d4da"} Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.449435 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hb4dm" event={"ID":"457f0d99-2681-403b-abc2-92af86fa76e5","Type":"ContainerDied","Data":"f782a058f108da24717810364f65cf60ee65d23c78dbaba7cd8939c96eef7a6e"} Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.449446 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hb4dm" event={"ID":"457f0d99-2681-403b-abc2-92af86fa76e5","Type":"ContainerDied","Data":"afbd64ee95f7e2f410dc87e9f75d523cc8c37d0567edc382c2d4bd06eaba651b"} Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.449458 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hb4dm" event={"ID":"457f0d99-2681-403b-abc2-92af86fa76e5","Type":"ContainerDied","Data":"804ccd844cb7c9fe0448b798d55e5f40e8e4b949c76f7acaeba4be31bc805ea0"} Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.449471 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hb4dm" event={"ID":"457f0d99-2681-403b-abc2-92af86fa76e5","Type":"ContainerDied","Data":"d6a8a1c2c4961b0cae8d98041227bad13852d7dab2d4043eb13ebfe7f9e983b8"} Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.449487 4612 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"3938aacd5fe03026f6a2c84286053cb2e237e9ebe6325bc6d4f1da844329e77e"} Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.449501 4612 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"272efda744ba67b324d44c1a9a0e60d9c2eede817ee274bdfa1a2adcef018517"} Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.449509 4612 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"3f34f9c07a0fd73cf62299b31f3851cf759b6f85e30aa577427a8cd07a956aaf"} Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.449519 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hb4dm" event={"ID":"457f0d99-2681-403b-abc2-92af86fa76e5","Type":"ContainerDied","Data":"3938aacd5fe03026f6a2c84286053cb2e237e9ebe6325bc6d4f1da844329e77e"} Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.449530 4612 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a920896c7db750f9e9ad9cd609c66d12c71145a4b4cc475ae07fbdc1d2694a7b"} Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.449538 4612 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"53ab32307d4bba133b74665292857b1dad285f3ef223f0758c3752cf1355d4da"} Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.449544 4612 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f782a058f108da24717810364f65cf60ee65d23c78dbaba7cd8939c96eef7a6e"} Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.449551 4612 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"afbd64ee95f7e2f410dc87e9f75d523cc8c37d0567edc382c2d4bd06eaba651b"} Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.449558 4612 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"804ccd844cb7c9fe0448b798d55e5f40e8e4b949c76f7acaeba4be31bc805ea0"} Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.449565 4612 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d6a8a1c2c4961b0cae8d98041227bad13852d7dab2d4043eb13ebfe7f9e983b8"} Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.449565 4612 scope.go:117] "RemoveContainer" containerID="a920896c7db750f9e9ad9cd609c66d12c71145a4b4cc475ae07fbdc1d2694a7b" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.449572 4612 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"3938aacd5fe03026f6a2c84286053cb2e237e9ebe6325bc6d4f1da844329e77e"} Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.449723 4612 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"272efda744ba67b324d44c1a9a0e60d9c2eede817ee274bdfa1a2adcef018517"} Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.449732 4612 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"3f34f9c07a0fd73cf62299b31f3851cf759b6f85e30aa577427a8cd07a956aaf"} Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.449744 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hb4dm" event={"ID":"457f0d99-2681-403b-abc2-92af86fa76e5","Type":"ContainerDied","Data":"272efda744ba67b324d44c1a9a0e60d9c2eede817ee274bdfa1a2adcef018517"} Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.449758 4612 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a920896c7db750f9e9ad9cd609c66d12c71145a4b4cc475ae07fbdc1d2694a7b"} Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.449767 4612 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"53ab32307d4bba133b74665292857b1dad285f3ef223f0758c3752cf1355d4da"} Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.449774 4612 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f782a058f108da24717810364f65cf60ee65d23c78dbaba7cd8939c96eef7a6e"} Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.449781 4612 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"afbd64ee95f7e2f410dc87e9f75d523cc8c37d0567edc382c2d4bd06eaba651b"} Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.449787 4612 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"804ccd844cb7c9fe0448b798d55e5f40e8e4b949c76f7acaeba4be31bc805ea0"} Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.449793 4612 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d6a8a1c2c4961b0cae8d98041227bad13852d7dab2d4043eb13ebfe7f9e983b8"} Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.449800 4612 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"3938aacd5fe03026f6a2c84286053cb2e237e9ebe6325bc6d4f1da844329e77e"} Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.449806 4612 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"272efda744ba67b324d44c1a9a0e60d9c2eede817ee274bdfa1a2adcef018517"} Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.449811 4612 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"3f34f9c07a0fd73cf62299b31f3851cf759b6f85e30aa577427a8cd07a956aaf"} Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.449820 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-hb4dm" event={"ID":"457f0d99-2681-403b-abc2-92af86fa76e5","Type":"ContainerDied","Data":"212df7bdcf3f29364e20e1f6189ba971500b49282e1edd5512d243a568a05c08"} Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.449830 4612 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a920896c7db750f9e9ad9cd609c66d12c71145a4b4cc475ae07fbdc1d2694a7b"} Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.449839 4612 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"53ab32307d4bba133b74665292857b1dad285f3ef223f0758c3752cf1355d4da"} Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.449852 4612 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f782a058f108da24717810364f65cf60ee65d23c78dbaba7cd8939c96eef7a6e"} Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.449859 4612 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"afbd64ee95f7e2f410dc87e9f75d523cc8c37d0567edc382c2d4bd06eaba651b"} Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.449865 4612 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"804ccd844cb7c9fe0448b798d55e5f40e8e4b949c76f7acaeba4be31bc805ea0"} Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.449872 4612 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d6a8a1c2c4961b0cae8d98041227bad13852d7dab2d4043eb13ebfe7f9e983b8"} Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.449880 4612 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"3938aacd5fe03026f6a2c84286053cb2e237e9ebe6325bc6d4f1da844329e77e"} Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.449886 4612 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"272efda744ba67b324d44c1a9a0e60d9c2eede817ee274bdfa1a2adcef018517"} Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.449905 4612 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"3f34f9c07a0fd73cf62299b31f3851cf759b6f85e30aa577427a8cd07a956aaf"} Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.451359 4612 generic.go:334] "Generic (PLEG): container finished" podID="9c746ad6-c797-40c9-8199-0a84a258fed3" containerID="9692a84e36eb170d9894c8f2908f79ea90d2e71da207e3cb1b76d0a9b88c0c50" exitCode=0 Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.451416 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ckhjd" event={"ID":"9c746ad6-c797-40c9-8199-0a84a258fed3","Type":"ContainerDied","Data":"9692a84e36eb170d9894c8f2908f79ea90d2e71da207e3cb1b76d0a9b88c0c50"} Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.451436 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ckhjd" event={"ID":"9c746ad6-c797-40c9-8199-0a84a258fed3","Type":"ContainerStarted","Data":"8a4d2c06eec493089b56f486217eceb565b0b524c5c0fd1b5140b931ff509505"} Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.453198 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-4hcsd_5ec92260-2494-43dc-8d39-0c554e8e161b/kube-multus/0.log" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.453228 4612 generic.go:334] "Generic (PLEG): container finished" podID="5ec92260-2494-43dc-8d39-0c554e8e161b" containerID="143ad013598715c75c27c4f0f6b62bb97d5f8c8999150abd8b94039eb7296215" exitCode=2 Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.453263 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-4hcsd" event={"ID":"5ec92260-2494-43dc-8d39-0c554e8e161b","Type":"ContainerDied","Data":"143ad013598715c75c27c4f0f6b62bb97d5f8c8999150abd8b94039eb7296215"} Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.455025 4612 scope.go:117] "RemoveContainer" containerID="143ad013598715c75c27c4f0f6b62bb97d5f8c8999150abd8b94039eb7296215" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.457467 4612 generic.go:334] "Generic (PLEG): container finished" podID="2114ee13-c2af-4646-a086-8491370cb05d" containerID="524db539480c08de1db0e7eb21a0e4cc343c22d5c6b00e8818bb1778f73f9935" exitCode=0 Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.457502 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536322-w8jkp" event={"ID":"2114ee13-c2af-4646-a086-8491370cb05d","Type":"ContainerDied","Data":"524db539480c08de1db0e7eb21a0e4cc343c22d5c6b00e8818bb1778f73f9935"} Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.491907 4612 scope.go:117] "RemoveContainer" containerID="53ab32307d4bba133b74665292857b1dad285f3ef223f0758c3752cf1355d4da" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.515139 4612 scope.go:117] "RemoveContainer" containerID="f782a058f108da24717810364f65cf60ee65d23c78dbaba7cd8939c96eef7a6e" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.542503 4612 scope.go:117] "RemoveContainer" containerID="afbd64ee95f7e2f410dc87e9f75d523cc8c37d0567edc382c2d4bd06eaba651b" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.560281 4612 scope.go:117] "RemoveContainer" containerID="804ccd844cb7c9fe0448b798d55e5f40e8e4b949c76f7acaeba4be31bc805ea0" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.567197 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-hb4dm"] Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.575036 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-hb4dm"] Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.603143 4612 scope.go:117] "RemoveContainer" containerID="d6a8a1c2c4961b0cae8d98041227bad13852d7dab2d4043eb13ebfe7f9e983b8" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.622742 4612 scope.go:117] "RemoveContainer" containerID="3938aacd5fe03026f6a2c84286053cb2e237e9ebe6325bc6d4f1da844329e77e" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.697725 4612 scope.go:117] "RemoveContainer" containerID="272efda744ba67b324d44c1a9a0e60d9c2eede817ee274bdfa1a2adcef018517" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.709680 4612 scope.go:117] "RemoveContainer" containerID="3f34f9c07a0fd73cf62299b31f3851cf759b6f85e30aa577427a8cd07a956aaf" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.732556 4612 scope.go:117] "RemoveContainer" containerID="a920896c7db750f9e9ad9cd609c66d12c71145a4b4cc475ae07fbdc1d2694a7b" Feb 27 08:02:02 crc kubenswrapper[4612]: E0227 08:02:02.733256 4612 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a920896c7db750f9e9ad9cd609c66d12c71145a4b4cc475ae07fbdc1d2694a7b\": container with ID starting with a920896c7db750f9e9ad9cd609c66d12c71145a4b4cc475ae07fbdc1d2694a7b not found: ID does not exist" containerID="a920896c7db750f9e9ad9cd609c66d12c71145a4b4cc475ae07fbdc1d2694a7b" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.733406 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a920896c7db750f9e9ad9cd609c66d12c71145a4b4cc475ae07fbdc1d2694a7b"} err="failed to get container status \"a920896c7db750f9e9ad9cd609c66d12c71145a4b4cc475ae07fbdc1d2694a7b\": rpc error: code = NotFound desc = could not find container \"a920896c7db750f9e9ad9cd609c66d12c71145a4b4cc475ae07fbdc1d2694a7b\": container with ID starting with a920896c7db750f9e9ad9cd609c66d12c71145a4b4cc475ae07fbdc1d2694a7b not found: ID does not exist" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.733524 4612 scope.go:117] "RemoveContainer" containerID="53ab32307d4bba133b74665292857b1dad285f3ef223f0758c3752cf1355d4da" Feb 27 08:02:02 crc kubenswrapper[4612]: E0227 08:02:02.734039 4612 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"53ab32307d4bba133b74665292857b1dad285f3ef223f0758c3752cf1355d4da\": container with ID starting with 53ab32307d4bba133b74665292857b1dad285f3ef223f0758c3752cf1355d4da not found: ID does not exist" containerID="53ab32307d4bba133b74665292857b1dad285f3ef223f0758c3752cf1355d4da" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.734129 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"53ab32307d4bba133b74665292857b1dad285f3ef223f0758c3752cf1355d4da"} err="failed to get container status \"53ab32307d4bba133b74665292857b1dad285f3ef223f0758c3752cf1355d4da\": rpc error: code = NotFound desc = could not find container \"53ab32307d4bba133b74665292857b1dad285f3ef223f0758c3752cf1355d4da\": container with ID starting with 53ab32307d4bba133b74665292857b1dad285f3ef223f0758c3752cf1355d4da not found: ID does not exist" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.734178 4612 scope.go:117] "RemoveContainer" containerID="f782a058f108da24717810364f65cf60ee65d23c78dbaba7cd8939c96eef7a6e" Feb 27 08:02:02 crc kubenswrapper[4612]: E0227 08:02:02.734541 4612 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f782a058f108da24717810364f65cf60ee65d23c78dbaba7cd8939c96eef7a6e\": container with ID starting with f782a058f108da24717810364f65cf60ee65d23c78dbaba7cd8939c96eef7a6e not found: ID does not exist" containerID="f782a058f108da24717810364f65cf60ee65d23c78dbaba7cd8939c96eef7a6e" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.734586 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f782a058f108da24717810364f65cf60ee65d23c78dbaba7cd8939c96eef7a6e"} err="failed to get container status \"f782a058f108da24717810364f65cf60ee65d23c78dbaba7cd8939c96eef7a6e\": rpc error: code = NotFound desc = could not find container \"f782a058f108da24717810364f65cf60ee65d23c78dbaba7cd8939c96eef7a6e\": container with ID starting with f782a058f108da24717810364f65cf60ee65d23c78dbaba7cd8939c96eef7a6e not found: ID does not exist" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.734616 4612 scope.go:117] "RemoveContainer" containerID="afbd64ee95f7e2f410dc87e9f75d523cc8c37d0567edc382c2d4bd06eaba651b" Feb 27 08:02:02 crc kubenswrapper[4612]: E0227 08:02:02.734991 4612 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"afbd64ee95f7e2f410dc87e9f75d523cc8c37d0567edc382c2d4bd06eaba651b\": container with ID starting with afbd64ee95f7e2f410dc87e9f75d523cc8c37d0567edc382c2d4bd06eaba651b not found: ID does not exist" containerID="afbd64ee95f7e2f410dc87e9f75d523cc8c37d0567edc382c2d4bd06eaba651b" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.735031 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"afbd64ee95f7e2f410dc87e9f75d523cc8c37d0567edc382c2d4bd06eaba651b"} err="failed to get container status \"afbd64ee95f7e2f410dc87e9f75d523cc8c37d0567edc382c2d4bd06eaba651b\": rpc error: code = NotFound desc = could not find container \"afbd64ee95f7e2f410dc87e9f75d523cc8c37d0567edc382c2d4bd06eaba651b\": container with ID starting with afbd64ee95f7e2f410dc87e9f75d523cc8c37d0567edc382c2d4bd06eaba651b not found: ID does not exist" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.735062 4612 scope.go:117] "RemoveContainer" containerID="804ccd844cb7c9fe0448b798d55e5f40e8e4b949c76f7acaeba4be31bc805ea0" Feb 27 08:02:02 crc kubenswrapper[4612]: E0227 08:02:02.735295 4612 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"804ccd844cb7c9fe0448b798d55e5f40e8e4b949c76f7acaeba4be31bc805ea0\": container with ID starting with 804ccd844cb7c9fe0448b798d55e5f40e8e4b949c76f7acaeba4be31bc805ea0 not found: ID does not exist" containerID="804ccd844cb7c9fe0448b798d55e5f40e8e4b949c76f7acaeba4be31bc805ea0" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.735342 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"804ccd844cb7c9fe0448b798d55e5f40e8e4b949c76f7acaeba4be31bc805ea0"} err="failed to get container status \"804ccd844cb7c9fe0448b798d55e5f40e8e4b949c76f7acaeba4be31bc805ea0\": rpc error: code = NotFound desc = could not find container \"804ccd844cb7c9fe0448b798d55e5f40e8e4b949c76f7acaeba4be31bc805ea0\": container with ID starting with 804ccd844cb7c9fe0448b798d55e5f40e8e4b949c76f7acaeba4be31bc805ea0 not found: ID does not exist" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.735371 4612 scope.go:117] "RemoveContainer" containerID="d6a8a1c2c4961b0cae8d98041227bad13852d7dab2d4043eb13ebfe7f9e983b8" Feb 27 08:02:02 crc kubenswrapper[4612]: E0227 08:02:02.735842 4612 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d6a8a1c2c4961b0cae8d98041227bad13852d7dab2d4043eb13ebfe7f9e983b8\": container with ID starting with d6a8a1c2c4961b0cae8d98041227bad13852d7dab2d4043eb13ebfe7f9e983b8 not found: ID does not exist" containerID="d6a8a1c2c4961b0cae8d98041227bad13852d7dab2d4043eb13ebfe7f9e983b8" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.735966 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d6a8a1c2c4961b0cae8d98041227bad13852d7dab2d4043eb13ebfe7f9e983b8"} err="failed to get container status \"d6a8a1c2c4961b0cae8d98041227bad13852d7dab2d4043eb13ebfe7f9e983b8\": rpc error: code = NotFound desc = could not find container \"d6a8a1c2c4961b0cae8d98041227bad13852d7dab2d4043eb13ebfe7f9e983b8\": container with ID starting with d6a8a1c2c4961b0cae8d98041227bad13852d7dab2d4043eb13ebfe7f9e983b8 not found: ID does not exist" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.736068 4612 scope.go:117] "RemoveContainer" containerID="3938aacd5fe03026f6a2c84286053cb2e237e9ebe6325bc6d4f1da844329e77e" Feb 27 08:02:02 crc kubenswrapper[4612]: E0227 08:02:02.736391 4612 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3938aacd5fe03026f6a2c84286053cb2e237e9ebe6325bc6d4f1da844329e77e\": container with ID starting with 3938aacd5fe03026f6a2c84286053cb2e237e9ebe6325bc6d4f1da844329e77e not found: ID does not exist" containerID="3938aacd5fe03026f6a2c84286053cb2e237e9ebe6325bc6d4f1da844329e77e" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.736423 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3938aacd5fe03026f6a2c84286053cb2e237e9ebe6325bc6d4f1da844329e77e"} err="failed to get container status \"3938aacd5fe03026f6a2c84286053cb2e237e9ebe6325bc6d4f1da844329e77e\": rpc error: code = NotFound desc = could not find container \"3938aacd5fe03026f6a2c84286053cb2e237e9ebe6325bc6d4f1da844329e77e\": container with ID starting with 3938aacd5fe03026f6a2c84286053cb2e237e9ebe6325bc6d4f1da844329e77e not found: ID does not exist" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.736443 4612 scope.go:117] "RemoveContainer" containerID="272efda744ba67b324d44c1a9a0e60d9c2eede817ee274bdfa1a2adcef018517" Feb 27 08:02:02 crc kubenswrapper[4612]: E0227 08:02:02.736669 4612 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"272efda744ba67b324d44c1a9a0e60d9c2eede817ee274bdfa1a2adcef018517\": container with ID starting with 272efda744ba67b324d44c1a9a0e60d9c2eede817ee274bdfa1a2adcef018517 not found: ID does not exist" containerID="272efda744ba67b324d44c1a9a0e60d9c2eede817ee274bdfa1a2adcef018517" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.736791 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"272efda744ba67b324d44c1a9a0e60d9c2eede817ee274bdfa1a2adcef018517"} err="failed to get container status \"272efda744ba67b324d44c1a9a0e60d9c2eede817ee274bdfa1a2adcef018517\": rpc error: code = NotFound desc = could not find container \"272efda744ba67b324d44c1a9a0e60d9c2eede817ee274bdfa1a2adcef018517\": container with ID starting with 272efda744ba67b324d44c1a9a0e60d9c2eede817ee274bdfa1a2adcef018517 not found: ID does not exist" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.736888 4612 scope.go:117] "RemoveContainer" containerID="3f34f9c07a0fd73cf62299b31f3851cf759b6f85e30aa577427a8cd07a956aaf" Feb 27 08:02:02 crc kubenswrapper[4612]: E0227 08:02:02.737360 4612 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3f34f9c07a0fd73cf62299b31f3851cf759b6f85e30aa577427a8cd07a956aaf\": container with ID starting with 3f34f9c07a0fd73cf62299b31f3851cf759b6f85e30aa577427a8cd07a956aaf not found: ID does not exist" containerID="3f34f9c07a0fd73cf62299b31f3851cf759b6f85e30aa577427a8cd07a956aaf" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.737470 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3f34f9c07a0fd73cf62299b31f3851cf759b6f85e30aa577427a8cd07a956aaf"} err="failed to get container status \"3f34f9c07a0fd73cf62299b31f3851cf759b6f85e30aa577427a8cd07a956aaf\": rpc error: code = NotFound desc = could not find container \"3f34f9c07a0fd73cf62299b31f3851cf759b6f85e30aa577427a8cd07a956aaf\": container with ID starting with 3f34f9c07a0fd73cf62299b31f3851cf759b6f85e30aa577427a8cd07a956aaf not found: ID does not exist" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.737582 4612 scope.go:117] "RemoveContainer" containerID="a920896c7db750f9e9ad9cd609c66d12c71145a4b4cc475ae07fbdc1d2694a7b" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.738034 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a920896c7db750f9e9ad9cd609c66d12c71145a4b4cc475ae07fbdc1d2694a7b"} err="failed to get container status \"a920896c7db750f9e9ad9cd609c66d12c71145a4b4cc475ae07fbdc1d2694a7b\": rpc error: code = NotFound desc = could not find container \"a920896c7db750f9e9ad9cd609c66d12c71145a4b4cc475ae07fbdc1d2694a7b\": container with ID starting with a920896c7db750f9e9ad9cd609c66d12c71145a4b4cc475ae07fbdc1d2694a7b not found: ID does not exist" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.738058 4612 scope.go:117] "RemoveContainer" containerID="53ab32307d4bba133b74665292857b1dad285f3ef223f0758c3752cf1355d4da" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.738334 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"53ab32307d4bba133b74665292857b1dad285f3ef223f0758c3752cf1355d4da"} err="failed to get container status \"53ab32307d4bba133b74665292857b1dad285f3ef223f0758c3752cf1355d4da\": rpc error: code = NotFound desc = could not find container \"53ab32307d4bba133b74665292857b1dad285f3ef223f0758c3752cf1355d4da\": container with ID starting with 53ab32307d4bba133b74665292857b1dad285f3ef223f0758c3752cf1355d4da not found: ID does not exist" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.738524 4612 scope.go:117] "RemoveContainer" containerID="f782a058f108da24717810364f65cf60ee65d23c78dbaba7cd8939c96eef7a6e" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.738917 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f782a058f108da24717810364f65cf60ee65d23c78dbaba7cd8939c96eef7a6e"} err="failed to get container status \"f782a058f108da24717810364f65cf60ee65d23c78dbaba7cd8939c96eef7a6e\": rpc error: code = NotFound desc = could not find container \"f782a058f108da24717810364f65cf60ee65d23c78dbaba7cd8939c96eef7a6e\": container with ID starting with f782a058f108da24717810364f65cf60ee65d23c78dbaba7cd8939c96eef7a6e not found: ID does not exist" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.738938 4612 scope.go:117] "RemoveContainer" containerID="afbd64ee95f7e2f410dc87e9f75d523cc8c37d0567edc382c2d4bd06eaba651b" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.739226 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"afbd64ee95f7e2f410dc87e9f75d523cc8c37d0567edc382c2d4bd06eaba651b"} err="failed to get container status \"afbd64ee95f7e2f410dc87e9f75d523cc8c37d0567edc382c2d4bd06eaba651b\": rpc error: code = NotFound desc = could not find container \"afbd64ee95f7e2f410dc87e9f75d523cc8c37d0567edc382c2d4bd06eaba651b\": container with ID starting with afbd64ee95f7e2f410dc87e9f75d523cc8c37d0567edc382c2d4bd06eaba651b not found: ID does not exist" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.739274 4612 scope.go:117] "RemoveContainer" containerID="804ccd844cb7c9fe0448b798d55e5f40e8e4b949c76f7acaeba4be31bc805ea0" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.739592 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"804ccd844cb7c9fe0448b798d55e5f40e8e4b949c76f7acaeba4be31bc805ea0"} err="failed to get container status \"804ccd844cb7c9fe0448b798d55e5f40e8e4b949c76f7acaeba4be31bc805ea0\": rpc error: code = NotFound desc = could not find container \"804ccd844cb7c9fe0448b798d55e5f40e8e4b949c76f7acaeba4be31bc805ea0\": container with ID starting with 804ccd844cb7c9fe0448b798d55e5f40e8e4b949c76f7acaeba4be31bc805ea0 not found: ID does not exist" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.739612 4612 scope.go:117] "RemoveContainer" containerID="d6a8a1c2c4961b0cae8d98041227bad13852d7dab2d4043eb13ebfe7f9e983b8" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.739839 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d6a8a1c2c4961b0cae8d98041227bad13852d7dab2d4043eb13ebfe7f9e983b8"} err="failed to get container status \"d6a8a1c2c4961b0cae8d98041227bad13852d7dab2d4043eb13ebfe7f9e983b8\": rpc error: code = NotFound desc = could not find container \"d6a8a1c2c4961b0cae8d98041227bad13852d7dab2d4043eb13ebfe7f9e983b8\": container with ID starting with d6a8a1c2c4961b0cae8d98041227bad13852d7dab2d4043eb13ebfe7f9e983b8 not found: ID does not exist" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.739864 4612 scope.go:117] "RemoveContainer" containerID="3938aacd5fe03026f6a2c84286053cb2e237e9ebe6325bc6d4f1da844329e77e" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.740047 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3938aacd5fe03026f6a2c84286053cb2e237e9ebe6325bc6d4f1da844329e77e"} err="failed to get container status \"3938aacd5fe03026f6a2c84286053cb2e237e9ebe6325bc6d4f1da844329e77e\": rpc error: code = NotFound desc = could not find container \"3938aacd5fe03026f6a2c84286053cb2e237e9ebe6325bc6d4f1da844329e77e\": container with ID starting with 3938aacd5fe03026f6a2c84286053cb2e237e9ebe6325bc6d4f1da844329e77e not found: ID does not exist" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.740066 4612 scope.go:117] "RemoveContainer" containerID="272efda744ba67b324d44c1a9a0e60d9c2eede817ee274bdfa1a2adcef018517" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.740242 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"272efda744ba67b324d44c1a9a0e60d9c2eede817ee274bdfa1a2adcef018517"} err="failed to get container status \"272efda744ba67b324d44c1a9a0e60d9c2eede817ee274bdfa1a2adcef018517\": rpc error: code = NotFound desc = could not find container \"272efda744ba67b324d44c1a9a0e60d9c2eede817ee274bdfa1a2adcef018517\": container with ID starting with 272efda744ba67b324d44c1a9a0e60d9c2eede817ee274bdfa1a2adcef018517 not found: ID does not exist" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.740258 4612 scope.go:117] "RemoveContainer" containerID="3f34f9c07a0fd73cf62299b31f3851cf759b6f85e30aa577427a8cd07a956aaf" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.740442 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3f34f9c07a0fd73cf62299b31f3851cf759b6f85e30aa577427a8cd07a956aaf"} err="failed to get container status \"3f34f9c07a0fd73cf62299b31f3851cf759b6f85e30aa577427a8cd07a956aaf\": rpc error: code = NotFound desc = could not find container \"3f34f9c07a0fd73cf62299b31f3851cf759b6f85e30aa577427a8cd07a956aaf\": container with ID starting with 3f34f9c07a0fd73cf62299b31f3851cf759b6f85e30aa577427a8cd07a956aaf not found: ID does not exist" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.740797 4612 scope.go:117] "RemoveContainer" containerID="a920896c7db750f9e9ad9cd609c66d12c71145a4b4cc475ae07fbdc1d2694a7b" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.740995 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a920896c7db750f9e9ad9cd609c66d12c71145a4b4cc475ae07fbdc1d2694a7b"} err="failed to get container status \"a920896c7db750f9e9ad9cd609c66d12c71145a4b4cc475ae07fbdc1d2694a7b\": rpc error: code = NotFound desc = could not find container \"a920896c7db750f9e9ad9cd609c66d12c71145a4b4cc475ae07fbdc1d2694a7b\": container with ID starting with a920896c7db750f9e9ad9cd609c66d12c71145a4b4cc475ae07fbdc1d2694a7b not found: ID does not exist" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.741019 4612 scope.go:117] "RemoveContainer" containerID="53ab32307d4bba133b74665292857b1dad285f3ef223f0758c3752cf1355d4da" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.741914 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"53ab32307d4bba133b74665292857b1dad285f3ef223f0758c3752cf1355d4da"} err="failed to get container status \"53ab32307d4bba133b74665292857b1dad285f3ef223f0758c3752cf1355d4da\": rpc error: code = NotFound desc = could not find container \"53ab32307d4bba133b74665292857b1dad285f3ef223f0758c3752cf1355d4da\": container with ID starting with 53ab32307d4bba133b74665292857b1dad285f3ef223f0758c3752cf1355d4da not found: ID does not exist" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.741954 4612 scope.go:117] "RemoveContainer" containerID="f782a058f108da24717810364f65cf60ee65d23c78dbaba7cd8939c96eef7a6e" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.742257 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f782a058f108da24717810364f65cf60ee65d23c78dbaba7cd8939c96eef7a6e"} err="failed to get container status \"f782a058f108da24717810364f65cf60ee65d23c78dbaba7cd8939c96eef7a6e\": rpc error: code = NotFound desc = could not find container \"f782a058f108da24717810364f65cf60ee65d23c78dbaba7cd8939c96eef7a6e\": container with ID starting with f782a058f108da24717810364f65cf60ee65d23c78dbaba7cd8939c96eef7a6e not found: ID does not exist" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.742283 4612 scope.go:117] "RemoveContainer" containerID="afbd64ee95f7e2f410dc87e9f75d523cc8c37d0567edc382c2d4bd06eaba651b" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.742481 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"afbd64ee95f7e2f410dc87e9f75d523cc8c37d0567edc382c2d4bd06eaba651b"} err="failed to get container status \"afbd64ee95f7e2f410dc87e9f75d523cc8c37d0567edc382c2d4bd06eaba651b\": rpc error: code = NotFound desc = could not find container \"afbd64ee95f7e2f410dc87e9f75d523cc8c37d0567edc382c2d4bd06eaba651b\": container with ID starting with afbd64ee95f7e2f410dc87e9f75d523cc8c37d0567edc382c2d4bd06eaba651b not found: ID does not exist" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.742500 4612 scope.go:117] "RemoveContainer" containerID="804ccd844cb7c9fe0448b798d55e5f40e8e4b949c76f7acaeba4be31bc805ea0" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.742729 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"804ccd844cb7c9fe0448b798d55e5f40e8e4b949c76f7acaeba4be31bc805ea0"} err="failed to get container status \"804ccd844cb7c9fe0448b798d55e5f40e8e4b949c76f7acaeba4be31bc805ea0\": rpc error: code = NotFound desc = could not find container \"804ccd844cb7c9fe0448b798d55e5f40e8e4b949c76f7acaeba4be31bc805ea0\": container with ID starting with 804ccd844cb7c9fe0448b798d55e5f40e8e4b949c76f7acaeba4be31bc805ea0 not found: ID does not exist" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.742754 4612 scope.go:117] "RemoveContainer" containerID="d6a8a1c2c4961b0cae8d98041227bad13852d7dab2d4043eb13ebfe7f9e983b8" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.742960 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d6a8a1c2c4961b0cae8d98041227bad13852d7dab2d4043eb13ebfe7f9e983b8"} err="failed to get container status \"d6a8a1c2c4961b0cae8d98041227bad13852d7dab2d4043eb13ebfe7f9e983b8\": rpc error: code = NotFound desc = could not find container \"d6a8a1c2c4961b0cae8d98041227bad13852d7dab2d4043eb13ebfe7f9e983b8\": container with ID starting with d6a8a1c2c4961b0cae8d98041227bad13852d7dab2d4043eb13ebfe7f9e983b8 not found: ID does not exist" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.742983 4612 scope.go:117] "RemoveContainer" containerID="3938aacd5fe03026f6a2c84286053cb2e237e9ebe6325bc6d4f1da844329e77e" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.743150 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3938aacd5fe03026f6a2c84286053cb2e237e9ebe6325bc6d4f1da844329e77e"} err="failed to get container status \"3938aacd5fe03026f6a2c84286053cb2e237e9ebe6325bc6d4f1da844329e77e\": rpc error: code = NotFound desc = could not find container \"3938aacd5fe03026f6a2c84286053cb2e237e9ebe6325bc6d4f1da844329e77e\": container with ID starting with 3938aacd5fe03026f6a2c84286053cb2e237e9ebe6325bc6d4f1da844329e77e not found: ID does not exist" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.743171 4612 scope.go:117] "RemoveContainer" containerID="272efda744ba67b324d44c1a9a0e60d9c2eede817ee274bdfa1a2adcef018517" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.743350 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"272efda744ba67b324d44c1a9a0e60d9c2eede817ee274bdfa1a2adcef018517"} err="failed to get container status \"272efda744ba67b324d44c1a9a0e60d9c2eede817ee274bdfa1a2adcef018517\": rpc error: code = NotFound desc = could not find container \"272efda744ba67b324d44c1a9a0e60d9c2eede817ee274bdfa1a2adcef018517\": container with ID starting with 272efda744ba67b324d44c1a9a0e60d9c2eede817ee274bdfa1a2adcef018517 not found: ID does not exist" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.743379 4612 scope.go:117] "RemoveContainer" containerID="3f34f9c07a0fd73cf62299b31f3851cf759b6f85e30aa577427a8cd07a956aaf" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.747794 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3f34f9c07a0fd73cf62299b31f3851cf759b6f85e30aa577427a8cd07a956aaf"} err="failed to get container status \"3f34f9c07a0fd73cf62299b31f3851cf759b6f85e30aa577427a8cd07a956aaf\": rpc error: code = NotFound desc = could not find container \"3f34f9c07a0fd73cf62299b31f3851cf759b6f85e30aa577427a8cd07a956aaf\": container with ID starting with 3f34f9c07a0fd73cf62299b31f3851cf759b6f85e30aa577427a8cd07a956aaf not found: ID does not exist" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.747822 4612 scope.go:117] "RemoveContainer" containerID="a920896c7db750f9e9ad9cd609c66d12c71145a4b4cc475ae07fbdc1d2694a7b" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.748317 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a920896c7db750f9e9ad9cd609c66d12c71145a4b4cc475ae07fbdc1d2694a7b"} err="failed to get container status \"a920896c7db750f9e9ad9cd609c66d12c71145a4b4cc475ae07fbdc1d2694a7b\": rpc error: code = NotFound desc = could not find container \"a920896c7db750f9e9ad9cd609c66d12c71145a4b4cc475ae07fbdc1d2694a7b\": container with ID starting with a920896c7db750f9e9ad9cd609c66d12c71145a4b4cc475ae07fbdc1d2694a7b not found: ID does not exist" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.748340 4612 scope.go:117] "RemoveContainer" containerID="53ab32307d4bba133b74665292857b1dad285f3ef223f0758c3752cf1355d4da" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.748536 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"53ab32307d4bba133b74665292857b1dad285f3ef223f0758c3752cf1355d4da"} err="failed to get container status \"53ab32307d4bba133b74665292857b1dad285f3ef223f0758c3752cf1355d4da\": rpc error: code = NotFound desc = could not find container \"53ab32307d4bba133b74665292857b1dad285f3ef223f0758c3752cf1355d4da\": container with ID starting with 53ab32307d4bba133b74665292857b1dad285f3ef223f0758c3752cf1355d4da not found: ID does not exist" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.748554 4612 scope.go:117] "RemoveContainer" containerID="f782a058f108da24717810364f65cf60ee65d23c78dbaba7cd8939c96eef7a6e" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.749761 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f782a058f108da24717810364f65cf60ee65d23c78dbaba7cd8939c96eef7a6e"} err="failed to get container status \"f782a058f108da24717810364f65cf60ee65d23c78dbaba7cd8939c96eef7a6e\": rpc error: code = NotFound desc = could not find container \"f782a058f108da24717810364f65cf60ee65d23c78dbaba7cd8939c96eef7a6e\": container with ID starting with f782a058f108da24717810364f65cf60ee65d23c78dbaba7cd8939c96eef7a6e not found: ID does not exist" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.749782 4612 scope.go:117] "RemoveContainer" containerID="afbd64ee95f7e2f410dc87e9f75d523cc8c37d0567edc382c2d4bd06eaba651b" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.750027 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"afbd64ee95f7e2f410dc87e9f75d523cc8c37d0567edc382c2d4bd06eaba651b"} err="failed to get container status \"afbd64ee95f7e2f410dc87e9f75d523cc8c37d0567edc382c2d4bd06eaba651b\": rpc error: code = NotFound desc = could not find container \"afbd64ee95f7e2f410dc87e9f75d523cc8c37d0567edc382c2d4bd06eaba651b\": container with ID starting with afbd64ee95f7e2f410dc87e9f75d523cc8c37d0567edc382c2d4bd06eaba651b not found: ID does not exist" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.750051 4612 scope.go:117] "RemoveContainer" containerID="804ccd844cb7c9fe0448b798d55e5f40e8e4b949c76f7acaeba4be31bc805ea0" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.750223 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"804ccd844cb7c9fe0448b798d55e5f40e8e4b949c76f7acaeba4be31bc805ea0"} err="failed to get container status \"804ccd844cb7c9fe0448b798d55e5f40e8e4b949c76f7acaeba4be31bc805ea0\": rpc error: code = NotFound desc = could not find container \"804ccd844cb7c9fe0448b798d55e5f40e8e4b949c76f7acaeba4be31bc805ea0\": container with ID starting with 804ccd844cb7c9fe0448b798d55e5f40e8e4b949c76f7acaeba4be31bc805ea0 not found: ID does not exist" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.750241 4612 scope.go:117] "RemoveContainer" containerID="d6a8a1c2c4961b0cae8d98041227bad13852d7dab2d4043eb13ebfe7f9e983b8" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.750392 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d6a8a1c2c4961b0cae8d98041227bad13852d7dab2d4043eb13ebfe7f9e983b8"} err="failed to get container status \"d6a8a1c2c4961b0cae8d98041227bad13852d7dab2d4043eb13ebfe7f9e983b8\": rpc error: code = NotFound desc = could not find container \"d6a8a1c2c4961b0cae8d98041227bad13852d7dab2d4043eb13ebfe7f9e983b8\": container with ID starting with d6a8a1c2c4961b0cae8d98041227bad13852d7dab2d4043eb13ebfe7f9e983b8 not found: ID does not exist" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.750409 4612 scope.go:117] "RemoveContainer" containerID="3938aacd5fe03026f6a2c84286053cb2e237e9ebe6325bc6d4f1da844329e77e" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.750661 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3938aacd5fe03026f6a2c84286053cb2e237e9ebe6325bc6d4f1da844329e77e"} err="failed to get container status \"3938aacd5fe03026f6a2c84286053cb2e237e9ebe6325bc6d4f1da844329e77e\": rpc error: code = NotFound desc = could not find container \"3938aacd5fe03026f6a2c84286053cb2e237e9ebe6325bc6d4f1da844329e77e\": container with ID starting with 3938aacd5fe03026f6a2c84286053cb2e237e9ebe6325bc6d4f1da844329e77e not found: ID does not exist" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.750833 4612 scope.go:117] "RemoveContainer" containerID="272efda744ba67b324d44c1a9a0e60d9c2eede817ee274bdfa1a2adcef018517" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.751184 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"272efda744ba67b324d44c1a9a0e60d9c2eede817ee274bdfa1a2adcef018517"} err="failed to get container status \"272efda744ba67b324d44c1a9a0e60d9c2eede817ee274bdfa1a2adcef018517\": rpc error: code = NotFound desc = could not find container \"272efda744ba67b324d44c1a9a0e60d9c2eede817ee274bdfa1a2adcef018517\": container with ID starting with 272efda744ba67b324d44c1a9a0e60d9c2eede817ee274bdfa1a2adcef018517 not found: ID does not exist" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.751203 4612 scope.go:117] "RemoveContainer" containerID="3f34f9c07a0fd73cf62299b31f3851cf759b6f85e30aa577427a8cd07a956aaf" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.751451 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3f34f9c07a0fd73cf62299b31f3851cf759b6f85e30aa577427a8cd07a956aaf"} err="failed to get container status \"3f34f9c07a0fd73cf62299b31f3851cf759b6f85e30aa577427a8cd07a956aaf\": rpc error: code = NotFound desc = could not find container \"3f34f9c07a0fd73cf62299b31f3851cf759b6f85e30aa577427a8cd07a956aaf\": container with ID starting with 3f34f9c07a0fd73cf62299b31f3851cf759b6f85e30aa577427a8cd07a956aaf not found: ID does not exist" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.751560 4612 scope.go:117] "RemoveContainer" containerID="a920896c7db750f9e9ad9cd609c66d12c71145a4b4cc475ae07fbdc1d2694a7b" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.751943 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a920896c7db750f9e9ad9cd609c66d12c71145a4b4cc475ae07fbdc1d2694a7b"} err="failed to get container status \"a920896c7db750f9e9ad9cd609c66d12c71145a4b4cc475ae07fbdc1d2694a7b\": rpc error: code = NotFound desc = could not find container \"a920896c7db750f9e9ad9cd609c66d12c71145a4b4cc475ae07fbdc1d2694a7b\": container with ID starting with a920896c7db750f9e9ad9cd609c66d12c71145a4b4cc475ae07fbdc1d2694a7b not found: ID does not exist" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.751971 4612 scope.go:117] "RemoveContainer" containerID="53ab32307d4bba133b74665292857b1dad285f3ef223f0758c3752cf1355d4da" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.752225 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"53ab32307d4bba133b74665292857b1dad285f3ef223f0758c3752cf1355d4da"} err="failed to get container status \"53ab32307d4bba133b74665292857b1dad285f3ef223f0758c3752cf1355d4da\": rpc error: code = NotFound desc = could not find container \"53ab32307d4bba133b74665292857b1dad285f3ef223f0758c3752cf1355d4da\": container with ID starting with 53ab32307d4bba133b74665292857b1dad285f3ef223f0758c3752cf1355d4da not found: ID does not exist" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.752351 4612 scope.go:117] "RemoveContainer" containerID="f782a058f108da24717810364f65cf60ee65d23c78dbaba7cd8939c96eef7a6e" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.752668 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f782a058f108da24717810364f65cf60ee65d23c78dbaba7cd8939c96eef7a6e"} err="failed to get container status \"f782a058f108da24717810364f65cf60ee65d23c78dbaba7cd8939c96eef7a6e\": rpc error: code = NotFound desc = could not find container \"f782a058f108da24717810364f65cf60ee65d23c78dbaba7cd8939c96eef7a6e\": container with ID starting with f782a058f108da24717810364f65cf60ee65d23c78dbaba7cd8939c96eef7a6e not found: ID does not exist" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.752697 4612 scope.go:117] "RemoveContainer" containerID="afbd64ee95f7e2f410dc87e9f75d523cc8c37d0567edc382c2d4bd06eaba651b" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.752866 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"afbd64ee95f7e2f410dc87e9f75d523cc8c37d0567edc382c2d4bd06eaba651b"} err="failed to get container status \"afbd64ee95f7e2f410dc87e9f75d523cc8c37d0567edc382c2d4bd06eaba651b\": rpc error: code = NotFound desc = could not find container \"afbd64ee95f7e2f410dc87e9f75d523cc8c37d0567edc382c2d4bd06eaba651b\": container with ID starting with afbd64ee95f7e2f410dc87e9f75d523cc8c37d0567edc382c2d4bd06eaba651b not found: ID does not exist" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.752883 4612 scope.go:117] "RemoveContainer" containerID="804ccd844cb7c9fe0448b798d55e5f40e8e4b949c76f7acaeba4be31bc805ea0" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.753108 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"804ccd844cb7c9fe0448b798d55e5f40e8e4b949c76f7acaeba4be31bc805ea0"} err="failed to get container status \"804ccd844cb7c9fe0448b798d55e5f40e8e4b949c76f7acaeba4be31bc805ea0\": rpc error: code = NotFound desc = could not find container \"804ccd844cb7c9fe0448b798d55e5f40e8e4b949c76f7acaeba4be31bc805ea0\": container with ID starting with 804ccd844cb7c9fe0448b798d55e5f40e8e4b949c76f7acaeba4be31bc805ea0 not found: ID does not exist" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.753213 4612 scope.go:117] "RemoveContainer" containerID="d6a8a1c2c4961b0cae8d98041227bad13852d7dab2d4043eb13ebfe7f9e983b8" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.753526 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d6a8a1c2c4961b0cae8d98041227bad13852d7dab2d4043eb13ebfe7f9e983b8"} err="failed to get container status \"d6a8a1c2c4961b0cae8d98041227bad13852d7dab2d4043eb13ebfe7f9e983b8\": rpc error: code = NotFound desc = could not find container \"d6a8a1c2c4961b0cae8d98041227bad13852d7dab2d4043eb13ebfe7f9e983b8\": container with ID starting with d6a8a1c2c4961b0cae8d98041227bad13852d7dab2d4043eb13ebfe7f9e983b8 not found: ID does not exist" Feb 27 08:02:02 crc kubenswrapper[4612]: I0227 08:02:02.880394 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="457f0d99-2681-403b-abc2-92af86fa76e5" path="/var/lib/kubelet/pods/457f0d99-2681-403b-abc2-92af86fa76e5/volumes" Feb 27 08:02:03 crc kubenswrapper[4612]: I0227 08:02:03.469206 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-4hcsd_5ec92260-2494-43dc-8d39-0c554e8e161b/kube-multus/0.log" Feb 27 08:02:03 crc kubenswrapper[4612]: I0227 08:02:03.469835 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-4hcsd" event={"ID":"5ec92260-2494-43dc-8d39-0c554e8e161b","Type":"ContainerStarted","Data":"14ba58adfc50959d3452acf10bcb6d03ffd6b50847dcbc4b04f4fc8099a927f8"} Feb 27 08:02:03 crc kubenswrapper[4612]: I0227 08:02:03.477635 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ckhjd" event={"ID":"9c746ad6-c797-40c9-8199-0a84a258fed3","Type":"ContainerStarted","Data":"50055ca560322fe608ee4d7438dffe1d202b4896609d67be8b63b54e4e7a58b1"} Feb 27 08:02:03 crc kubenswrapper[4612]: I0227 08:02:03.477720 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ckhjd" event={"ID":"9c746ad6-c797-40c9-8199-0a84a258fed3","Type":"ContainerStarted","Data":"ebdc8922a2575c06d3eab65ab4e07945b064fc3d9915f0c2ba5e81d57b117bcc"} Feb 27 08:02:03 crc kubenswrapper[4612]: I0227 08:02:03.477741 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ckhjd" event={"ID":"9c746ad6-c797-40c9-8199-0a84a258fed3","Type":"ContainerStarted","Data":"a1376465bd5f08290930f3397451517df29493705e809057234ac6841cc77669"} Feb 27 08:02:03 crc kubenswrapper[4612]: I0227 08:02:03.477759 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ckhjd" event={"ID":"9c746ad6-c797-40c9-8199-0a84a258fed3","Type":"ContainerStarted","Data":"9e4f737b899b8e80ca1c3c3c4141f922a3dab5b06dbb75f9901ee610942ad6bf"} Feb 27 08:02:03 crc kubenswrapper[4612]: I0227 08:02:03.477776 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ckhjd" event={"ID":"9c746ad6-c797-40c9-8199-0a84a258fed3","Type":"ContainerStarted","Data":"01abe0fd39fdbe522c214b0bb4df79424a7f43a9f237e988702f19d9b2cb6e41"} Feb 27 08:02:03 crc kubenswrapper[4612]: I0227 08:02:03.477792 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ckhjd" event={"ID":"9c746ad6-c797-40c9-8199-0a84a258fed3","Type":"ContainerStarted","Data":"c26a5c84bc11f8372d63e3af9d707412b6a36d6f6ead34ebdc483b9cfcd4a1a5"} Feb 27 08:02:03 crc kubenswrapper[4612]: I0227 08:02:03.517119 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536322-w8jkp" Feb 27 08:02:03 crc kubenswrapper[4612]: I0227 08:02:03.650809 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-llzzn\" (UniqueName: \"kubernetes.io/projected/2114ee13-c2af-4646-a086-8491370cb05d-kube-api-access-llzzn\") pod \"2114ee13-c2af-4646-a086-8491370cb05d\" (UID: \"2114ee13-c2af-4646-a086-8491370cb05d\") " Feb 27 08:02:03 crc kubenswrapper[4612]: I0227 08:02:03.656020 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2114ee13-c2af-4646-a086-8491370cb05d-kube-api-access-llzzn" (OuterVolumeSpecName: "kube-api-access-llzzn") pod "2114ee13-c2af-4646-a086-8491370cb05d" (UID: "2114ee13-c2af-4646-a086-8491370cb05d"). InnerVolumeSpecName "kube-api-access-llzzn". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 08:02:03 crc kubenswrapper[4612]: I0227 08:02:03.752007 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-llzzn\" (UniqueName: \"kubernetes.io/projected/2114ee13-c2af-4646-a086-8491370cb05d-kube-api-access-llzzn\") on node \"crc\" DevicePath \"\"" Feb 27 08:02:04 crc kubenswrapper[4612]: I0227 08:02:04.484955 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536322-w8jkp" event={"ID":"2114ee13-c2af-4646-a086-8491370cb05d","Type":"ContainerDied","Data":"1baaa97756c08c8de8fb236c2812e4fd1485309cef2946ba1ceeb64fc4a4c5da"} Feb 27 08:02:04 crc kubenswrapper[4612]: I0227 08:02:04.484989 4612 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1baaa97756c08c8de8fb236c2812e4fd1485309cef2946ba1ceeb64fc4a4c5da" Feb 27 08:02:04 crc kubenswrapper[4612]: I0227 08:02:04.485037 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536322-w8jkp" Feb 27 08:02:04 crc kubenswrapper[4612]: I0227 08:02:04.583975 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29536316-5mt4v"] Feb 27 08:02:04 crc kubenswrapper[4612]: I0227 08:02:04.588821 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29536316-5mt4v"] Feb 27 08:02:04 crc kubenswrapper[4612]: I0227 08:02:04.866218 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="863bc847-68ff-4ddc-92c6-eb4f95a00347" path="/var/lib/kubelet/pods/863bc847-68ff-4ddc-92c6-eb4f95a00347/volumes" Feb 27 08:02:05 crc kubenswrapper[4612]: I0227 08:02:05.493580 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ckhjd" event={"ID":"9c746ad6-c797-40c9-8199-0a84a258fed3","Type":"ContainerStarted","Data":"c88c412aee8f12bd5093cbcdd62a8893171f4245857324e5a1a99cad671ee946"} Feb 27 08:02:08 crc kubenswrapper[4612]: I0227 08:02:08.530008 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-ckhjd" event={"ID":"9c746ad6-c797-40c9-8199-0a84a258fed3","Type":"ContainerStarted","Data":"9da969b242036201045dd8945012f10e49eaba6b6c19b250664a67b5ea4b5251"} Feb 27 08:02:08 crc kubenswrapper[4612]: I0227 08:02:08.530880 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-ckhjd" Feb 27 08:02:08 crc kubenswrapper[4612]: I0227 08:02:08.530917 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-ckhjd" Feb 27 08:02:08 crc kubenswrapper[4612]: I0227 08:02:08.530941 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-ckhjd" Feb 27 08:02:08 crc kubenswrapper[4612]: I0227 08:02:08.570576 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-ckhjd" Feb 27 08:02:08 crc kubenswrapper[4612]: I0227 08:02:08.574628 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-ckhjd" Feb 27 08:02:08 crc kubenswrapper[4612]: I0227 08:02:08.579522 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-ckhjd" podStartSLOduration=7.579505505 podStartE2EDuration="7.579505505s" podCreationTimestamp="2026-02-27 08:02:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 08:02:08.574661776 +0000 UTC m=+786.428591784" watchObservedRunningTime="2026-02-27 08:02:08.579505505 +0000 UTC m=+786.433435513" Feb 27 08:02:16 crc kubenswrapper[4612]: I0227 08:02:16.027285 4612 patch_prober.go:28] interesting pod/machine-config-daemon-924vb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 27 08:02:16 crc kubenswrapper[4612]: I0227 08:02:16.027967 4612 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 27 08:02:32 crc kubenswrapper[4612]: I0227 08:02:32.274338 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-ckhjd" Feb 27 08:02:39 crc kubenswrapper[4612]: I0227 08:02:39.374124 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82g84wq"] Feb 27 08:02:39 crc kubenswrapper[4612]: E0227 08:02:39.376133 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2114ee13-c2af-4646-a086-8491370cb05d" containerName="oc" Feb 27 08:02:39 crc kubenswrapper[4612]: I0227 08:02:39.376249 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="2114ee13-c2af-4646-a086-8491370cb05d" containerName="oc" Feb 27 08:02:39 crc kubenswrapper[4612]: I0227 08:02:39.376505 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="2114ee13-c2af-4646-a086-8491370cb05d" containerName="oc" Feb 27 08:02:39 crc kubenswrapper[4612]: I0227 08:02:39.377770 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82g84wq" Feb 27 08:02:39 crc kubenswrapper[4612]: I0227 08:02:39.380403 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Feb 27 08:02:39 crc kubenswrapper[4612]: I0227 08:02:39.384082 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82g84wq"] Feb 27 08:02:39 crc kubenswrapper[4612]: I0227 08:02:39.542858 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d7ee024d-7259-4702-bc99-fb938a71345f-util\") pod \"0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82g84wq\" (UID: \"d7ee024d-7259-4702-bc99-fb938a71345f\") " pod="openshift-marketplace/0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82g84wq" Feb 27 08:02:39 crc kubenswrapper[4612]: I0227 08:02:39.542985 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l6ctk\" (UniqueName: \"kubernetes.io/projected/d7ee024d-7259-4702-bc99-fb938a71345f-kube-api-access-l6ctk\") pod \"0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82g84wq\" (UID: \"d7ee024d-7259-4702-bc99-fb938a71345f\") " pod="openshift-marketplace/0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82g84wq" Feb 27 08:02:39 crc kubenswrapper[4612]: I0227 08:02:39.543125 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d7ee024d-7259-4702-bc99-fb938a71345f-bundle\") pod \"0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82g84wq\" (UID: \"d7ee024d-7259-4702-bc99-fb938a71345f\") " pod="openshift-marketplace/0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82g84wq" Feb 27 08:02:39 crc kubenswrapper[4612]: I0227 08:02:39.644313 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d7ee024d-7259-4702-bc99-fb938a71345f-util\") pod \"0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82g84wq\" (UID: \"d7ee024d-7259-4702-bc99-fb938a71345f\") " pod="openshift-marketplace/0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82g84wq" Feb 27 08:02:39 crc kubenswrapper[4612]: I0227 08:02:39.644778 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l6ctk\" (UniqueName: \"kubernetes.io/projected/d7ee024d-7259-4702-bc99-fb938a71345f-kube-api-access-l6ctk\") pod \"0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82g84wq\" (UID: \"d7ee024d-7259-4702-bc99-fb938a71345f\") " pod="openshift-marketplace/0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82g84wq" Feb 27 08:02:39 crc kubenswrapper[4612]: I0227 08:02:39.645001 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d7ee024d-7259-4702-bc99-fb938a71345f-util\") pod \"0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82g84wq\" (UID: \"d7ee024d-7259-4702-bc99-fb938a71345f\") " pod="openshift-marketplace/0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82g84wq" Feb 27 08:02:39 crc kubenswrapper[4612]: I0227 08:02:39.645215 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d7ee024d-7259-4702-bc99-fb938a71345f-bundle\") pod \"0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82g84wq\" (UID: \"d7ee024d-7259-4702-bc99-fb938a71345f\") " pod="openshift-marketplace/0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82g84wq" Feb 27 08:02:39 crc kubenswrapper[4612]: I0227 08:02:39.645420 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d7ee024d-7259-4702-bc99-fb938a71345f-bundle\") pod \"0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82g84wq\" (UID: \"d7ee024d-7259-4702-bc99-fb938a71345f\") " pod="openshift-marketplace/0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82g84wq" Feb 27 08:02:39 crc kubenswrapper[4612]: I0227 08:02:39.679193 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l6ctk\" (UniqueName: \"kubernetes.io/projected/d7ee024d-7259-4702-bc99-fb938a71345f-kube-api-access-l6ctk\") pod \"0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82g84wq\" (UID: \"d7ee024d-7259-4702-bc99-fb938a71345f\") " pod="openshift-marketplace/0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82g84wq" Feb 27 08:02:39 crc kubenswrapper[4612]: I0227 08:02:39.694254 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82g84wq" Feb 27 08:02:40 crc kubenswrapper[4612]: I0227 08:02:40.086629 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82g84wq"] Feb 27 08:02:40 crc kubenswrapper[4612]: I0227 08:02:40.764625 4612 generic.go:334] "Generic (PLEG): container finished" podID="d7ee024d-7259-4702-bc99-fb938a71345f" containerID="8e91c371c7c04cd06318ab249aa1a5280eb9d48128f2c67ec89588a077e4db22" exitCode=0 Feb 27 08:02:40 crc kubenswrapper[4612]: I0227 08:02:40.764757 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82g84wq" event={"ID":"d7ee024d-7259-4702-bc99-fb938a71345f","Type":"ContainerDied","Data":"8e91c371c7c04cd06318ab249aa1a5280eb9d48128f2c67ec89588a077e4db22"} Feb 27 08:02:40 crc kubenswrapper[4612]: I0227 08:02:40.765027 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82g84wq" event={"ID":"d7ee024d-7259-4702-bc99-fb938a71345f","Type":"ContainerStarted","Data":"215fe6826fa5e129022598f2c33937e9d48c93a37dc19be9097ce77b63630699"} Feb 27 08:02:41 crc kubenswrapper[4612]: I0227 08:02:41.573446 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-knw8s"] Feb 27 08:02:41 crc kubenswrapper[4612]: I0227 08:02:41.574929 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-knw8s" Feb 27 08:02:41 crc kubenswrapper[4612]: I0227 08:02:41.602724 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-knw8s"] Feb 27 08:02:41 crc kubenswrapper[4612]: I0227 08:02:41.672146 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/67c16138-cb89-4526-a7d8-604a710da7e2-catalog-content\") pod \"redhat-operators-knw8s\" (UID: \"67c16138-cb89-4526-a7d8-604a710da7e2\") " pod="openshift-marketplace/redhat-operators-knw8s" Feb 27 08:02:41 crc kubenswrapper[4612]: I0227 08:02:41.672210 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d2mg5\" (UniqueName: \"kubernetes.io/projected/67c16138-cb89-4526-a7d8-604a710da7e2-kube-api-access-d2mg5\") pod \"redhat-operators-knw8s\" (UID: \"67c16138-cb89-4526-a7d8-604a710da7e2\") " pod="openshift-marketplace/redhat-operators-knw8s" Feb 27 08:02:41 crc kubenswrapper[4612]: I0227 08:02:41.672230 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/67c16138-cb89-4526-a7d8-604a710da7e2-utilities\") pod \"redhat-operators-knw8s\" (UID: \"67c16138-cb89-4526-a7d8-604a710da7e2\") " pod="openshift-marketplace/redhat-operators-knw8s" Feb 27 08:02:41 crc kubenswrapper[4612]: I0227 08:02:41.773203 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/67c16138-cb89-4526-a7d8-604a710da7e2-catalog-content\") pod \"redhat-operators-knw8s\" (UID: \"67c16138-cb89-4526-a7d8-604a710da7e2\") " pod="openshift-marketplace/redhat-operators-knw8s" Feb 27 08:02:41 crc kubenswrapper[4612]: I0227 08:02:41.773271 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d2mg5\" (UniqueName: \"kubernetes.io/projected/67c16138-cb89-4526-a7d8-604a710da7e2-kube-api-access-d2mg5\") pod \"redhat-operators-knw8s\" (UID: \"67c16138-cb89-4526-a7d8-604a710da7e2\") " pod="openshift-marketplace/redhat-operators-knw8s" Feb 27 08:02:41 crc kubenswrapper[4612]: I0227 08:02:41.773291 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/67c16138-cb89-4526-a7d8-604a710da7e2-utilities\") pod \"redhat-operators-knw8s\" (UID: \"67c16138-cb89-4526-a7d8-604a710da7e2\") " pod="openshift-marketplace/redhat-operators-knw8s" Feb 27 08:02:41 crc kubenswrapper[4612]: I0227 08:02:41.773660 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/67c16138-cb89-4526-a7d8-604a710da7e2-catalog-content\") pod \"redhat-operators-knw8s\" (UID: \"67c16138-cb89-4526-a7d8-604a710da7e2\") " pod="openshift-marketplace/redhat-operators-knw8s" Feb 27 08:02:41 crc kubenswrapper[4612]: I0227 08:02:41.773707 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/67c16138-cb89-4526-a7d8-604a710da7e2-utilities\") pod \"redhat-operators-knw8s\" (UID: \"67c16138-cb89-4526-a7d8-604a710da7e2\") " pod="openshift-marketplace/redhat-operators-knw8s" Feb 27 08:02:41 crc kubenswrapper[4612]: I0227 08:02:41.792983 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d2mg5\" (UniqueName: \"kubernetes.io/projected/67c16138-cb89-4526-a7d8-604a710da7e2-kube-api-access-d2mg5\") pod \"redhat-operators-knw8s\" (UID: \"67c16138-cb89-4526-a7d8-604a710da7e2\") " pod="openshift-marketplace/redhat-operators-knw8s" Feb 27 08:02:41 crc kubenswrapper[4612]: I0227 08:02:41.896188 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-knw8s" Feb 27 08:02:42 crc kubenswrapper[4612]: I0227 08:02:42.133730 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-knw8s"] Feb 27 08:02:42 crc kubenswrapper[4612]: W0227 08:02:42.154065 4612 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod67c16138_cb89_4526_a7d8_604a710da7e2.slice/crio-9aa87c3e9b8b1922430a5f311e2b619d91f62bd4848e0724f5c44fada60887f4 WatchSource:0}: Error finding container 9aa87c3e9b8b1922430a5f311e2b619d91f62bd4848e0724f5c44fada60887f4: Status 404 returned error can't find the container with id 9aa87c3e9b8b1922430a5f311e2b619d91f62bd4848e0724f5c44fada60887f4 Feb 27 08:02:42 crc kubenswrapper[4612]: I0227 08:02:42.775895 4612 generic.go:334] "Generic (PLEG): container finished" podID="67c16138-cb89-4526-a7d8-604a710da7e2" containerID="dc77ad940eb3a00087300237d4a7397a0c806a6e0b563d041b4898f334240328" exitCode=0 Feb 27 08:02:42 crc kubenswrapper[4612]: I0227 08:02:42.776152 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-knw8s" event={"ID":"67c16138-cb89-4526-a7d8-604a710da7e2","Type":"ContainerDied","Data":"dc77ad940eb3a00087300237d4a7397a0c806a6e0b563d041b4898f334240328"} Feb 27 08:02:42 crc kubenswrapper[4612]: I0227 08:02:42.776177 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-knw8s" event={"ID":"67c16138-cb89-4526-a7d8-604a710da7e2","Type":"ContainerStarted","Data":"9aa87c3e9b8b1922430a5f311e2b619d91f62bd4848e0724f5c44fada60887f4"} Feb 27 08:02:42 crc kubenswrapper[4612]: I0227 08:02:42.779077 4612 generic.go:334] "Generic (PLEG): container finished" podID="d7ee024d-7259-4702-bc99-fb938a71345f" containerID="d03b71ce8197d9d47980a3f5c4f24b2812efe0fde486587b8bb156506458d6af" exitCode=0 Feb 27 08:02:42 crc kubenswrapper[4612]: I0227 08:02:42.779125 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82g84wq" event={"ID":"d7ee024d-7259-4702-bc99-fb938a71345f","Type":"ContainerDied","Data":"d03b71ce8197d9d47980a3f5c4f24b2812efe0fde486587b8bb156506458d6af"} Feb 27 08:02:43 crc kubenswrapper[4612]: I0227 08:02:43.790801 4612 generic.go:334] "Generic (PLEG): container finished" podID="d7ee024d-7259-4702-bc99-fb938a71345f" containerID="087927dfb9d697ec5bd8c145c92e36088793dc68379eed66b28c18fa1c709799" exitCode=0 Feb 27 08:02:43 crc kubenswrapper[4612]: I0227 08:02:43.791283 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82g84wq" event={"ID":"d7ee024d-7259-4702-bc99-fb938a71345f","Type":"ContainerDied","Data":"087927dfb9d697ec5bd8c145c92e36088793dc68379eed66b28c18fa1c709799"} Feb 27 08:02:44 crc kubenswrapper[4612]: I0227 08:02:44.799976 4612 generic.go:334] "Generic (PLEG): container finished" podID="67c16138-cb89-4526-a7d8-604a710da7e2" containerID="b01dacb0be9ca7368b260f13d2396841bc91268067a56ab34354ef04fa6767ce" exitCode=0 Feb 27 08:02:44 crc kubenswrapper[4612]: I0227 08:02:44.800809 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-knw8s" event={"ID":"67c16138-cb89-4526-a7d8-604a710da7e2","Type":"ContainerDied","Data":"b01dacb0be9ca7368b260f13d2396841bc91268067a56ab34354ef04fa6767ce"} Feb 27 08:02:45 crc kubenswrapper[4612]: I0227 08:02:45.112727 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82g84wq" Feb 27 08:02:45 crc kubenswrapper[4612]: I0227 08:02:45.214661 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d7ee024d-7259-4702-bc99-fb938a71345f-bundle\") pod \"d7ee024d-7259-4702-bc99-fb938a71345f\" (UID: \"d7ee024d-7259-4702-bc99-fb938a71345f\") " Feb 27 08:02:45 crc kubenswrapper[4612]: I0227 08:02:45.214794 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l6ctk\" (UniqueName: \"kubernetes.io/projected/d7ee024d-7259-4702-bc99-fb938a71345f-kube-api-access-l6ctk\") pod \"d7ee024d-7259-4702-bc99-fb938a71345f\" (UID: \"d7ee024d-7259-4702-bc99-fb938a71345f\") " Feb 27 08:02:45 crc kubenswrapper[4612]: I0227 08:02:45.214878 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d7ee024d-7259-4702-bc99-fb938a71345f-util\") pod \"d7ee024d-7259-4702-bc99-fb938a71345f\" (UID: \"d7ee024d-7259-4702-bc99-fb938a71345f\") " Feb 27 08:02:45 crc kubenswrapper[4612]: I0227 08:02:45.215565 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d7ee024d-7259-4702-bc99-fb938a71345f-bundle" (OuterVolumeSpecName: "bundle") pod "d7ee024d-7259-4702-bc99-fb938a71345f" (UID: "d7ee024d-7259-4702-bc99-fb938a71345f"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 08:02:45 crc kubenswrapper[4612]: I0227 08:02:45.220914 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d7ee024d-7259-4702-bc99-fb938a71345f-kube-api-access-l6ctk" (OuterVolumeSpecName: "kube-api-access-l6ctk") pod "d7ee024d-7259-4702-bc99-fb938a71345f" (UID: "d7ee024d-7259-4702-bc99-fb938a71345f"). InnerVolumeSpecName "kube-api-access-l6ctk". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 08:02:45 crc kubenswrapper[4612]: I0227 08:02:45.244409 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d7ee024d-7259-4702-bc99-fb938a71345f-util" (OuterVolumeSpecName: "util") pod "d7ee024d-7259-4702-bc99-fb938a71345f" (UID: "d7ee024d-7259-4702-bc99-fb938a71345f"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 08:02:45 crc kubenswrapper[4612]: I0227 08:02:45.316652 4612 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d7ee024d-7259-4702-bc99-fb938a71345f-util\") on node \"crc\" DevicePath \"\"" Feb 27 08:02:45 crc kubenswrapper[4612]: I0227 08:02:45.316743 4612 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d7ee024d-7259-4702-bc99-fb938a71345f-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 08:02:45 crc kubenswrapper[4612]: I0227 08:02:45.316769 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l6ctk\" (UniqueName: \"kubernetes.io/projected/d7ee024d-7259-4702-bc99-fb938a71345f-kube-api-access-l6ctk\") on node \"crc\" DevicePath \"\"" Feb 27 08:02:45 crc kubenswrapper[4612]: I0227 08:02:45.811443 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-knw8s" event={"ID":"67c16138-cb89-4526-a7d8-604a710da7e2","Type":"ContainerStarted","Data":"3d0b3af6e88a68a1b77fc307e48d4b83663c2bfd1c7e7534ad0b08f73841d2c9"} Feb 27 08:02:45 crc kubenswrapper[4612]: I0227 08:02:45.815602 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82g84wq" event={"ID":"d7ee024d-7259-4702-bc99-fb938a71345f","Type":"ContainerDied","Data":"215fe6826fa5e129022598f2c33937e9d48c93a37dc19be9097ce77b63630699"} Feb 27 08:02:45 crc kubenswrapper[4612]: I0227 08:02:45.815635 4612 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="215fe6826fa5e129022598f2c33937e9d48c93a37dc19be9097ce77b63630699" Feb 27 08:02:45 crc kubenswrapper[4612]: I0227 08:02:45.815715 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82g84wq" Feb 27 08:02:45 crc kubenswrapper[4612]: I0227 08:02:45.836497 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-knw8s" podStartSLOduration=2.31618697 podStartE2EDuration="4.836479161s" podCreationTimestamp="2026-02-27 08:02:41 +0000 UTC" firstStartedPulling="2026-02-27 08:02:42.777661142 +0000 UTC m=+820.631591140" lastFinishedPulling="2026-02-27 08:02:45.297953333 +0000 UTC m=+823.151883331" observedRunningTime="2026-02-27 08:02:45.832498107 +0000 UTC m=+823.686428135" watchObservedRunningTime="2026-02-27 08:02:45.836479161 +0000 UTC m=+823.690409159" Feb 27 08:02:46 crc kubenswrapper[4612]: I0227 08:02:46.027071 4612 patch_prober.go:28] interesting pod/machine-config-daemon-924vb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 27 08:02:46 crc kubenswrapper[4612]: I0227 08:02:46.027160 4612 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 27 08:02:46 crc kubenswrapper[4612]: I0227 08:02:46.027217 4612 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-924vb" Feb 27 08:02:46 crc kubenswrapper[4612]: I0227 08:02:46.027806 4612 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"b0982c79ab30408b51bb931dbdd8a46865a7a758c57125e452728efbaccdd1ec"} pod="openshift-machine-config-operator/machine-config-daemon-924vb" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 27 08:02:46 crc kubenswrapper[4612]: I0227 08:02:46.027890 4612 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" containerName="machine-config-daemon" containerID="cri-o://b0982c79ab30408b51bb931dbdd8a46865a7a758c57125e452728efbaccdd1ec" gracePeriod=600 Feb 27 08:02:46 crc kubenswrapper[4612]: I0227 08:02:46.823580 4612 generic.go:334] "Generic (PLEG): container finished" podID="24599373-7adc-4a1b-8bb4-797bf726f43d" containerID="b0982c79ab30408b51bb931dbdd8a46865a7a758c57125e452728efbaccdd1ec" exitCode=0 Feb 27 08:02:46 crc kubenswrapper[4612]: I0227 08:02:46.823655 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-924vb" event={"ID":"24599373-7adc-4a1b-8bb4-797bf726f43d","Type":"ContainerDied","Data":"b0982c79ab30408b51bb931dbdd8a46865a7a758c57125e452728efbaccdd1ec"} Feb 27 08:02:46 crc kubenswrapper[4612]: I0227 08:02:46.824245 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-924vb" event={"ID":"24599373-7adc-4a1b-8bb4-797bf726f43d","Type":"ContainerStarted","Data":"3ee7350e6fe9f892458f31227d045e092d0ee2f4312ea6cc953bb5da78fe0fa7"} Feb 27 08:02:46 crc kubenswrapper[4612]: I0227 08:02:46.824274 4612 scope.go:117] "RemoveContainer" containerID="778719935b065f111a6a6df70937016b6e803a5b231a05b52794a3dcd1909559" Feb 27 08:02:46 crc kubenswrapper[4612]: I0227 08:02:46.886232 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-75c5dccd6c-p6nbk"] Feb 27 08:02:46 crc kubenswrapper[4612]: E0227 08:02:46.886429 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d7ee024d-7259-4702-bc99-fb938a71345f" containerName="extract" Feb 27 08:02:46 crc kubenswrapper[4612]: I0227 08:02:46.886440 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="d7ee024d-7259-4702-bc99-fb938a71345f" containerName="extract" Feb 27 08:02:46 crc kubenswrapper[4612]: E0227 08:02:46.886457 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d7ee024d-7259-4702-bc99-fb938a71345f" containerName="util" Feb 27 08:02:46 crc kubenswrapper[4612]: I0227 08:02:46.886463 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="d7ee024d-7259-4702-bc99-fb938a71345f" containerName="util" Feb 27 08:02:46 crc kubenswrapper[4612]: E0227 08:02:46.886474 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d7ee024d-7259-4702-bc99-fb938a71345f" containerName="pull" Feb 27 08:02:46 crc kubenswrapper[4612]: I0227 08:02:46.886480 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="d7ee024d-7259-4702-bc99-fb938a71345f" containerName="pull" Feb 27 08:02:46 crc kubenswrapper[4612]: I0227 08:02:46.886576 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="d7ee024d-7259-4702-bc99-fb938a71345f" containerName="extract" Feb 27 08:02:46 crc kubenswrapper[4612]: I0227 08:02:46.886919 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-75c5dccd6c-p6nbk" Feb 27 08:02:46 crc kubenswrapper[4612]: I0227 08:02:46.889241 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Feb 27 08:02:46 crc kubenswrapper[4612]: I0227 08:02:46.889300 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-n5d4f" Feb 27 08:02:46 crc kubenswrapper[4612]: I0227 08:02:46.889409 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Feb 27 08:02:46 crc kubenswrapper[4612]: I0227 08:02:46.899471 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-75c5dccd6c-p6nbk"] Feb 27 08:02:47 crc kubenswrapper[4612]: I0227 08:02:47.037818 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hh5hb\" (UniqueName: \"kubernetes.io/projected/49fa2d59-6289-4edc-a6d8-2c3f7733a536-kube-api-access-hh5hb\") pod \"nmstate-operator-75c5dccd6c-p6nbk\" (UID: \"49fa2d59-6289-4edc-a6d8-2c3f7733a536\") " pod="openshift-nmstate/nmstate-operator-75c5dccd6c-p6nbk" Feb 27 08:02:47 crc kubenswrapper[4612]: I0227 08:02:47.138929 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hh5hb\" (UniqueName: \"kubernetes.io/projected/49fa2d59-6289-4edc-a6d8-2c3f7733a536-kube-api-access-hh5hb\") pod \"nmstate-operator-75c5dccd6c-p6nbk\" (UID: \"49fa2d59-6289-4edc-a6d8-2c3f7733a536\") " pod="openshift-nmstate/nmstate-operator-75c5dccd6c-p6nbk" Feb 27 08:02:47 crc kubenswrapper[4612]: I0227 08:02:47.177341 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hh5hb\" (UniqueName: \"kubernetes.io/projected/49fa2d59-6289-4edc-a6d8-2c3f7733a536-kube-api-access-hh5hb\") pod \"nmstate-operator-75c5dccd6c-p6nbk\" (UID: \"49fa2d59-6289-4edc-a6d8-2c3f7733a536\") " pod="openshift-nmstate/nmstate-operator-75c5dccd6c-p6nbk" Feb 27 08:02:47 crc kubenswrapper[4612]: I0227 08:02:47.202764 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-75c5dccd6c-p6nbk" Feb 27 08:02:47 crc kubenswrapper[4612]: I0227 08:02:47.461481 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-75c5dccd6c-p6nbk"] Feb 27 08:02:47 crc kubenswrapper[4612]: I0227 08:02:47.830661 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-75c5dccd6c-p6nbk" event={"ID":"49fa2d59-6289-4edc-a6d8-2c3f7733a536","Type":"ContainerStarted","Data":"3a0d90358e8a86c8567bac05fa3d500b44887cc438ee322aa366848c3d30ef2f"} Feb 27 08:02:50 crc kubenswrapper[4612]: I0227 08:02:50.864667 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-75c5dccd6c-p6nbk" event={"ID":"49fa2d59-6289-4edc-a6d8-2c3f7733a536","Type":"ContainerStarted","Data":"ebdbfc5c8030b309e26223c531b91e27b66e26824a9c2ffae1a080ca0de8b57b"} Feb 27 08:02:50 crc kubenswrapper[4612]: I0227 08:02:50.889483 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-75c5dccd6c-p6nbk" podStartSLOduration=1.826058999 podStartE2EDuration="4.889463189s" podCreationTimestamp="2026-02-27 08:02:46 +0000 UTC" firstStartedPulling="2026-02-27 08:02:47.477601796 +0000 UTC m=+825.331531794" lastFinishedPulling="2026-02-27 08:02:50.541005986 +0000 UTC m=+828.394935984" observedRunningTime="2026-02-27 08:02:50.88705514 +0000 UTC m=+828.740985148" watchObservedRunningTime="2026-02-27 08:02:50.889463189 +0000 UTC m=+828.743393187" Feb 27 08:02:51 crc kubenswrapper[4612]: I0227 08:02:51.897200 4612 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-knw8s" Feb 27 08:02:51 crc kubenswrapper[4612]: I0227 08:02:51.900421 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-knw8s" Feb 27 08:02:51 crc kubenswrapper[4612]: I0227 08:02:51.930832 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-69594cc75-czssc"] Feb 27 08:02:51 crc kubenswrapper[4612]: I0227 08:02:51.932499 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-69594cc75-czssc" Feb 27 08:02:51 crc kubenswrapper[4612]: I0227 08:02:51.940165 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-zcx4s" Feb 27 08:02:51 crc kubenswrapper[4612]: I0227 08:02:51.960683 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-69594cc75-czssc"] Feb 27 08:02:51 crc kubenswrapper[4612]: I0227 08:02:51.965516 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-786f45cff4-m8llg"] Feb 27 08:02:51 crc kubenswrapper[4612]: I0227 08:02:51.966518 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-786f45cff4-m8llg" Feb 27 08:02:51 crc kubenswrapper[4612]: I0227 08:02:51.969852 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Feb 27 08:02:52 crc kubenswrapper[4612]: I0227 08:02:52.004400 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-26zsp"] Feb 27 08:02:52 crc kubenswrapper[4612]: I0227 08:02:52.005439 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-26zsp" Feb 27 08:02:52 crc kubenswrapper[4612]: I0227 08:02:52.018148 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-786f45cff4-m8llg"] Feb 27 08:02:52 crc kubenswrapper[4612]: I0227 08:02:52.038326 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mp9sb\" (UniqueName: \"kubernetes.io/projected/d058174c-69c3-49cf-b846-af074b709972-kube-api-access-mp9sb\") pod \"nmstate-metrics-69594cc75-czssc\" (UID: \"d058174c-69c3-49cf-b846-af074b709972\") " pod="openshift-nmstate/nmstate-metrics-69594cc75-czssc" Feb 27 08:02:52 crc kubenswrapper[4612]: I0227 08:02:52.113630 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-5dcbbd79cf-4cfq5"] Feb 27 08:02:52 crc kubenswrapper[4612]: I0227 08:02:52.114738 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-5dcbbd79cf-4cfq5" Feb 27 08:02:52 crc kubenswrapper[4612]: I0227 08:02:52.116353 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Feb 27 08:02:52 crc kubenswrapper[4612]: I0227 08:02:52.116751 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-7zmlx" Feb 27 08:02:52 crc kubenswrapper[4612]: I0227 08:02:52.120326 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Feb 27 08:02:52 crc kubenswrapper[4612]: I0227 08:02:52.124913 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-5dcbbd79cf-4cfq5"] Feb 27 08:02:52 crc kubenswrapper[4612]: I0227 08:02:52.139404 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sdchb\" (UniqueName: \"kubernetes.io/projected/a82cc418-4f77-4788-addb-25341d42d9fa-kube-api-access-sdchb\") pod \"nmstate-console-plugin-5dcbbd79cf-4cfq5\" (UID: \"a82cc418-4f77-4788-addb-25341d42d9fa\") " pod="openshift-nmstate/nmstate-console-plugin-5dcbbd79cf-4cfq5" Feb 27 08:02:52 crc kubenswrapper[4612]: I0227 08:02:52.139706 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qztsg\" (UniqueName: \"kubernetes.io/projected/b141dc9e-912e-4096-97fc-5df6e0ae245c-kube-api-access-qztsg\") pod \"nmstate-webhook-786f45cff4-m8llg\" (UID: \"b141dc9e-912e-4096-97fc-5df6e0ae245c\") " pod="openshift-nmstate/nmstate-webhook-786f45cff4-m8llg" Feb 27 08:02:52 crc kubenswrapper[4612]: I0227 08:02:52.139739 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/b141dc9e-912e-4096-97fc-5df6e0ae245c-tls-key-pair\") pod \"nmstate-webhook-786f45cff4-m8llg\" (UID: \"b141dc9e-912e-4096-97fc-5df6e0ae245c\") " pod="openshift-nmstate/nmstate-webhook-786f45cff4-m8llg" Feb 27 08:02:52 crc kubenswrapper[4612]: I0227 08:02:52.139758 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/5553e6c7-b45f-40c9-99ce-abd46767d31a-ovs-socket\") pod \"nmstate-handler-26zsp\" (UID: \"5553e6c7-b45f-40c9-99ce-abd46767d31a\") " pod="openshift-nmstate/nmstate-handler-26zsp" Feb 27 08:02:52 crc kubenswrapper[4612]: I0227 08:02:52.139783 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/5553e6c7-b45f-40c9-99ce-abd46767d31a-nmstate-lock\") pod \"nmstate-handler-26zsp\" (UID: \"5553e6c7-b45f-40c9-99ce-abd46767d31a\") " pod="openshift-nmstate/nmstate-handler-26zsp" Feb 27 08:02:52 crc kubenswrapper[4612]: I0227 08:02:52.139826 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mp9sb\" (UniqueName: \"kubernetes.io/projected/d058174c-69c3-49cf-b846-af074b709972-kube-api-access-mp9sb\") pod \"nmstate-metrics-69594cc75-czssc\" (UID: \"d058174c-69c3-49cf-b846-af074b709972\") " pod="openshift-nmstate/nmstate-metrics-69594cc75-czssc" Feb 27 08:02:52 crc kubenswrapper[4612]: I0227 08:02:52.139897 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-28m5k\" (UniqueName: \"kubernetes.io/projected/5553e6c7-b45f-40c9-99ce-abd46767d31a-kube-api-access-28m5k\") pod \"nmstate-handler-26zsp\" (UID: \"5553e6c7-b45f-40c9-99ce-abd46767d31a\") " pod="openshift-nmstate/nmstate-handler-26zsp" Feb 27 08:02:52 crc kubenswrapper[4612]: I0227 08:02:52.139925 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/a82cc418-4f77-4788-addb-25341d42d9fa-nginx-conf\") pod \"nmstate-console-plugin-5dcbbd79cf-4cfq5\" (UID: \"a82cc418-4f77-4788-addb-25341d42d9fa\") " pod="openshift-nmstate/nmstate-console-plugin-5dcbbd79cf-4cfq5" Feb 27 08:02:52 crc kubenswrapper[4612]: I0227 08:02:52.139980 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/5553e6c7-b45f-40c9-99ce-abd46767d31a-dbus-socket\") pod \"nmstate-handler-26zsp\" (UID: \"5553e6c7-b45f-40c9-99ce-abd46767d31a\") " pod="openshift-nmstate/nmstate-handler-26zsp" Feb 27 08:02:52 crc kubenswrapper[4612]: I0227 08:02:52.139997 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/a82cc418-4f77-4788-addb-25341d42d9fa-plugin-serving-cert\") pod \"nmstate-console-plugin-5dcbbd79cf-4cfq5\" (UID: \"a82cc418-4f77-4788-addb-25341d42d9fa\") " pod="openshift-nmstate/nmstate-console-plugin-5dcbbd79cf-4cfq5" Feb 27 08:02:52 crc kubenswrapper[4612]: I0227 08:02:52.158364 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mp9sb\" (UniqueName: \"kubernetes.io/projected/d058174c-69c3-49cf-b846-af074b709972-kube-api-access-mp9sb\") pod \"nmstate-metrics-69594cc75-czssc\" (UID: \"d058174c-69c3-49cf-b846-af074b709972\") " pod="openshift-nmstate/nmstate-metrics-69594cc75-czssc" Feb 27 08:02:52 crc kubenswrapper[4612]: I0227 08:02:52.240593 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/5553e6c7-b45f-40c9-99ce-abd46767d31a-ovs-socket\") pod \"nmstate-handler-26zsp\" (UID: \"5553e6c7-b45f-40c9-99ce-abd46767d31a\") " pod="openshift-nmstate/nmstate-handler-26zsp" Feb 27 08:02:52 crc kubenswrapper[4612]: I0227 08:02:52.240660 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/5553e6c7-b45f-40c9-99ce-abd46767d31a-nmstate-lock\") pod \"nmstate-handler-26zsp\" (UID: \"5553e6c7-b45f-40c9-99ce-abd46767d31a\") " pod="openshift-nmstate/nmstate-handler-26zsp" Feb 27 08:02:52 crc kubenswrapper[4612]: I0227 08:02:52.240716 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-28m5k\" (UniqueName: \"kubernetes.io/projected/5553e6c7-b45f-40c9-99ce-abd46767d31a-kube-api-access-28m5k\") pod \"nmstate-handler-26zsp\" (UID: \"5553e6c7-b45f-40c9-99ce-abd46767d31a\") " pod="openshift-nmstate/nmstate-handler-26zsp" Feb 27 08:02:52 crc kubenswrapper[4612]: I0227 08:02:52.240756 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/a82cc418-4f77-4788-addb-25341d42d9fa-nginx-conf\") pod \"nmstate-console-plugin-5dcbbd79cf-4cfq5\" (UID: \"a82cc418-4f77-4788-addb-25341d42d9fa\") " pod="openshift-nmstate/nmstate-console-plugin-5dcbbd79cf-4cfq5" Feb 27 08:02:52 crc kubenswrapper[4612]: I0227 08:02:52.240785 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/5553e6c7-b45f-40c9-99ce-abd46767d31a-ovs-socket\") pod \"nmstate-handler-26zsp\" (UID: \"5553e6c7-b45f-40c9-99ce-abd46767d31a\") " pod="openshift-nmstate/nmstate-handler-26zsp" Feb 27 08:02:52 crc kubenswrapper[4612]: I0227 08:02:52.240805 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/5553e6c7-b45f-40c9-99ce-abd46767d31a-dbus-socket\") pod \"nmstate-handler-26zsp\" (UID: \"5553e6c7-b45f-40c9-99ce-abd46767d31a\") " pod="openshift-nmstate/nmstate-handler-26zsp" Feb 27 08:02:52 crc kubenswrapper[4612]: I0227 08:02:52.240857 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/5553e6c7-b45f-40c9-99ce-abd46767d31a-nmstate-lock\") pod \"nmstate-handler-26zsp\" (UID: \"5553e6c7-b45f-40c9-99ce-abd46767d31a\") " pod="openshift-nmstate/nmstate-handler-26zsp" Feb 27 08:02:52 crc kubenswrapper[4612]: I0227 08:02:52.241064 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/a82cc418-4f77-4788-addb-25341d42d9fa-plugin-serving-cert\") pod \"nmstate-console-plugin-5dcbbd79cf-4cfq5\" (UID: \"a82cc418-4f77-4788-addb-25341d42d9fa\") " pod="openshift-nmstate/nmstate-console-plugin-5dcbbd79cf-4cfq5" Feb 27 08:02:52 crc kubenswrapper[4612]: I0227 08:02:52.241113 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sdchb\" (UniqueName: \"kubernetes.io/projected/a82cc418-4f77-4788-addb-25341d42d9fa-kube-api-access-sdchb\") pod \"nmstate-console-plugin-5dcbbd79cf-4cfq5\" (UID: \"a82cc418-4f77-4788-addb-25341d42d9fa\") " pod="openshift-nmstate/nmstate-console-plugin-5dcbbd79cf-4cfq5" Feb 27 08:02:52 crc kubenswrapper[4612]: I0227 08:02:52.241139 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qztsg\" (UniqueName: \"kubernetes.io/projected/b141dc9e-912e-4096-97fc-5df6e0ae245c-kube-api-access-qztsg\") pod \"nmstate-webhook-786f45cff4-m8llg\" (UID: \"b141dc9e-912e-4096-97fc-5df6e0ae245c\") " pod="openshift-nmstate/nmstate-webhook-786f45cff4-m8llg" Feb 27 08:02:52 crc kubenswrapper[4612]: I0227 08:02:52.241170 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/b141dc9e-912e-4096-97fc-5df6e0ae245c-tls-key-pair\") pod \"nmstate-webhook-786f45cff4-m8llg\" (UID: \"b141dc9e-912e-4096-97fc-5df6e0ae245c\") " pod="openshift-nmstate/nmstate-webhook-786f45cff4-m8llg" Feb 27 08:02:52 crc kubenswrapper[4612]: E0227 08:02:52.241390 4612 secret.go:188] Couldn't get secret openshift-nmstate/plugin-serving-cert: secret "plugin-serving-cert" not found Feb 27 08:02:52 crc kubenswrapper[4612]: I0227 08:02:52.241443 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/5553e6c7-b45f-40c9-99ce-abd46767d31a-dbus-socket\") pod \"nmstate-handler-26zsp\" (UID: \"5553e6c7-b45f-40c9-99ce-abd46767d31a\") " pod="openshift-nmstate/nmstate-handler-26zsp" Feb 27 08:02:52 crc kubenswrapper[4612]: E0227 08:02:52.241485 4612 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a82cc418-4f77-4788-addb-25341d42d9fa-plugin-serving-cert podName:a82cc418-4f77-4788-addb-25341d42d9fa nodeName:}" failed. No retries permitted until 2026-02-27 08:02:52.741462981 +0000 UTC m=+830.595393079 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "plugin-serving-cert" (UniqueName: "kubernetes.io/secret/a82cc418-4f77-4788-addb-25341d42d9fa-plugin-serving-cert") pod "nmstate-console-plugin-5dcbbd79cf-4cfq5" (UID: "a82cc418-4f77-4788-addb-25341d42d9fa") : secret "plugin-serving-cert" not found Feb 27 08:02:52 crc kubenswrapper[4612]: I0227 08:02:52.241614 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/a82cc418-4f77-4788-addb-25341d42d9fa-nginx-conf\") pod \"nmstate-console-plugin-5dcbbd79cf-4cfq5\" (UID: \"a82cc418-4f77-4788-addb-25341d42d9fa\") " pod="openshift-nmstate/nmstate-console-plugin-5dcbbd79cf-4cfq5" Feb 27 08:02:52 crc kubenswrapper[4612]: I0227 08:02:52.245031 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/b141dc9e-912e-4096-97fc-5df6e0ae245c-tls-key-pair\") pod \"nmstate-webhook-786f45cff4-m8llg\" (UID: \"b141dc9e-912e-4096-97fc-5df6e0ae245c\") " pod="openshift-nmstate/nmstate-webhook-786f45cff4-m8llg" Feb 27 08:02:52 crc kubenswrapper[4612]: I0227 08:02:52.254071 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-69594cc75-czssc" Feb 27 08:02:52 crc kubenswrapper[4612]: I0227 08:02:52.266546 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sdchb\" (UniqueName: \"kubernetes.io/projected/a82cc418-4f77-4788-addb-25341d42d9fa-kube-api-access-sdchb\") pod \"nmstate-console-plugin-5dcbbd79cf-4cfq5\" (UID: \"a82cc418-4f77-4788-addb-25341d42d9fa\") " pod="openshift-nmstate/nmstate-console-plugin-5dcbbd79cf-4cfq5" Feb 27 08:02:52 crc kubenswrapper[4612]: I0227 08:02:52.276528 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qztsg\" (UniqueName: \"kubernetes.io/projected/b141dc9e-912e-4096-97fc-5df6e0ae245c-kube-api-access-qztsg\") pod \"nmstate-webhook-786f45cff4-m8llg\" (UID: \"b141dc9e-912e-4096-97fc-5df6e0ae245c\") " pod="openshift-nmstate/nmstate-webhook-786f45cff4-m8llg" Feb 27 08:02:52 crc kubenswrapper[4612]: I0227 08:02:52.280158 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-28m5k\" (UniqueName: \"kubernetes.io/projected/5553e6c7-b45f-40c9-99ce-abd46767d31a-kube-api-access-28m5k\") pod \"nmstate-handler-26zsp\" (UID: \"5553e6c7-b45f-40c9-99ce-abd46767d31a\") " pod="openshift-nmstate/nmstate-handler-26zsp" Feb 27 08:02:52 crc kubenswrapper[4612]: I0227 08:02:52.291842 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-786f45cff4-m8llg" Feb 27 08:02:52 crc kubenswrapper[4612]: I0227 08:02:52.319507 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-26zsp" Feb 27 08:02:52 crc kubenswrapper[4612]: I0227 08:02:52.378280 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-7c8f999976-kdvq5"] Feb 27 08:02:52 crc kubenswrapper[4612]: I0227 08:02:52.379867 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-7c8f999976-kdvq5" Feb 27 08:02:52 crc kubenswrapper[4612]: I0227 08:02:52.418675 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-7c8f999976-kdvq5"] Feb 27 08:02:52 crc kubenswrapper[4612]: I0227 08:02:52.446467 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/eb7933c3-204a-46b8-870f-4f776d70a03c-console-oauth-config\") pod \"console-7c8f999976-kdvq5\" (UID: \"eb7933c3-204a-46b8-870f-4f776d70a03c\") " pod="openshift-console/console-7c8f999976-kdvq5" Feb 27 08:02:52 crc kubenswrapper[4612]: I0227 08:02:52.446505 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6l6pj\" (UniqueName: \"kubernetes.io/projected/eb7933c3-204a-46b8-870f-4f776d70a03c-kube-api-access-6l6pj\") pod \"console-7c8f999976-kdvq5\" (UID: \"eb7933c3-204a-46b8-870f-4f776d70a03c\") " pod="openshift-console/console-7c8f999976-kdvq5" Feb 27 08:02:52 crc kubenswrapper[4612]: I0227 08:02:52.446637 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/eb7933c3-204a-46b8-870f-4f776d70a03c-oauth-serving-cert\") pod \"console-7c8f999976-kdvq5\" (UID: \"eb7933c3-204a-46b8-870f-4f776d70a03c\") " pod="openshift-console/console-7c8f999976-kdvq5" Feb 27 08:02:52 crc kubenswrapper[4612]: I0227 08:02:52.446658 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/eb7933c3-204a-46b8-870f-4f776d70a03c-console-serving-cert\") pod \"console-7c8f999976-kdvq5\" (UID: \"eb7933c3-204a-46b8-870f-4f776d70a03c\") " pod="openshift-console/console-7c8f999976-kdvq5" Feb 27 08:02:52 crc kubenswrapper[4612]: I0227 08:02:52.446736 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/eb7933c3-204a-46b8-870f-4f776d70a03c-service-ca\") pod \"console-7c8f999976-kdvq5\" (UID: \"eb7933c3-204a-46b8-870f-4f776d70a03c\") " pod="openshift-console/console-7c8f999976-kdvq5" Feb 27 08:02:52 crc kubenswrapper[4612]: I0227 08:02:52.446762 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/eb7933c3-204a-46b8-870f-4f776d70a03c-console-config\") pod \"console-7c8f999976-kdvq5\" (UID: \"eb7933c3-204a-46b8-870f-4f776d70a03c\") " pod="openshift-console/console-7c8f999976-kdvq5" Feb 27 08:02:52 crc kubenswrapper[4612]: I0227 08:02:52.446886 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/eb7933c3-204a-46b8-870f-4f776d70a03c-trusted-ca-bundle\") pod \"console-7c8f999976-kdvq5\" (UID: \"eb7933c3-204a-46b8-870f-4f776d70a03c\") " pod="openshift-console/console-7c8f999976-kdvq5" Feb 27 08:02:52 crc kubenswrapper[4612]: I0227 08:02:52.547655 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/eb7933c3-204a-46b8-870f-4f776d70a03c-console-config\") pod \"console-7c8f999976-kdvq5\" (UID: \"eb7933c3-204a-46b8-870f-4f776d70a03c\") " pod="openshift-console/console-7c8f999976-kdvq5" Feb 27 08:02:52 crc kubenswrapper[4612]: I0227 08:02:52.547746 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/eb7933c3-204a-46b8-870f-4f776d70a03c-trusted-ca-bundle\") pod \"console-7c8f999976-kdvq5\" (UID: \"eb7933c3-204a-46b8-870f-4f776d70a03c\") " pod="openshift-console/console-7c8f999976-kdvq5" Feb 27 08:02:52 crc kubenswrapper[4612]: I0227 08:02:52.547795 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/eb7933c3-204a-46b8-870f-4f776d70a03c-console-oauth-config\") pod \"console-7c8f999976-kdvq5\" (UID: \"eb7933c3-204a-46b8-870f-4f776d70a03c\") " pod="openshift-console/console-7c8f999976-kdvq5" Feb 27 08:02:52 crc kubenswrapper[4612]: I0227 08:02:52.547820 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6l6pj\" (UniqueName: \"kubernetes.io/projected/eb7933c3-204a-46b8-870f-4f776d70a03c-kube-api-access-6l6pj\") pod \"console-7c8f999976-kdvq5\" (UID: \"eb7933c3-204a-46b8-870f-4f776d70a03c\") " pod="openshift-console/console-7c8f999976-kdvq5" Feb 27 08:02:52 crc kubenswrapper[4612]: I0227 08:02:52.547841 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/eb7933c3-204a-46b8-870f-4f776d70a03c-oauth-serving-cert\") pod \"console-7c8f999976-kdvq5\" (UID: \"eb7933c3-204a-46b8-870f-4f776d70a03c\") " pod="openshift-console/console-7c8f999976-kdvq5" Feb 27 08:02:52 crc kubenswrapper[4612]: I0227 08:02:52.548177 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/eb7933c3-204a-46b8-870f-4f776d70a03c-console-serving-cert\") pod \"console-7c8f999976-kdvq5\" (UID: \"eb7933c3-204a-46b8-870f-4f776d70a03c\") " pod="openshift-console/console-7c8f999976-kdvq5" Feb 27 08:02:52 crc kubenswrapper[4612]: I0227 08:02:52.548409 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/eb7933c3-204a-46b8-870f-4f776d70a03c-service-ca\") pod \"console-7c8f999976-kdvq5\" (UID: \"eb7933c3-204a-46b8-870f-4f776d70a03c\") " pod="openshift-console/console-7c8f999976-kdvq5" Feb 27 08:02:52 crc kubenswrapper[4612]: I0227 08:02:52.548847 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/eb7933c3-204a-46b8-870f-4f776d70a03c-oauth-serving-cert\") pod \"console-7c8f999976-kdvq5\" (UID: \"eb7933c3-204a-46b8-870f-4f776d70a03c\") " pod="openshift-console/console-7c8f999976-kdvq5" Feb 27 08:02:52 crc kubenswrapper[4612]: I0227 08:02:52.549310 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/eb7933c3-204a-46b8-870f-4f776d70a03c-service-ca\") pod \"console-7c8f999976-kdvq5\" (UID: \"eb7933c3-204a-46b8-870f-4f776d70a03c\") " pod="openshift-console/console-7c8f999976-kdvq5" Feb 27 08:02:52 crc kubenswrapper[4612]: I0227 08:02:52.550272 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/eb7933c3-204a-46b8-870f-4f776d70a03c-trusted-ca-bundle\") pod \"console-7c8f999976-kdvq5\" (UID: \"eb7933c3-204a-46b8-870f-4f776d70a03c\") " pod="openshift-console/console-7c8f999976-kdvq5" Feb 27 08:02:52 crc kubenswrapper[4612]: I0227 08:02:52.550294 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/eb7933c3-204a-46b8-870f-4f776d70a03c-console-config\") pod \"console-7c8f999976-kdvq5\" (UID: \"eb7933c3-204a-46b8-870f-4f776d70a03c\") " pod="openshift-console/console-7c8f999976-kdvq5" Feb 27 08:02:52 crc kubenswrapper[4612]: I0227 08:02:52.557066 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/eb7933c3-204a-46b8-870f-4f776d70a03c-console-oauth-config\") pod \"console-7c8f999976-kdvq5\" (UID: \"eb7933c3-204a-46b8-870f-4f776d70a03c\") " pod="openshift-console/console-7c8f999976-kdvq5" Feb 27 08:02:52 crc kubenswrapper[4612]: I0227 08:02:52.558916 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/eb7933c3-204a-46b8-870f-4f776d70a03c-console-serving-cert\") pod \"console-7c8f999976-kdvq5\" (UID: \"eb7933c3-204a-46b8-870f-4f776d70a03c\") " pod="openshift-console/console-7c8f999976-kdvq5" Feb 27 08:02:52 crc kubenswrapper[4612]: I0227 08:02:52.563946 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6l6pj\" (UniqueName: \"kubernetes.io/projected/eb7933c3-204a-46b8-870f-4f776d70a03c-kube-api-access-6l6pj\") pod \"console-7c8f999976-kdvq5\" (UID: \"eb7933c3-204a-46b8-870f-4f776d70a03c\") " pod="openshift-console/console-7c8f999976-kdvq5" Feb 27 08:02:52 crc kubenswrapper[4612]: I0227 08:02:52.588305 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-69594cc75-czssc"] Feb 27 08:02:52 crc kubenswrapper[4612]: I0227 08:02:52.750288 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/a82cc418-4f77-4788-addb-25341d42d9fa-plugin-serving-cert\") pod \"nmstate-console-plugin-5dcbbd79cf-4cfq5\" (UID: \"a82cc418-4f77-4788-addb-25341d42d9fa\") " pod="openshift-nmstate/nmstate-console-plugin-5dcbbd79cf-4cfq5" Feb 27 08:02:52 crc kubenswrapper[4612]: I0227 08:02:52.752092 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-7c8f999976-kdvq5" Feb 27 08:02:52 crc kubenswrapper[4612]: I0227 08:02:52.754378 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/a82cc418-4f77-4788-addb-25341d42d9fa-plugin-serving-cert\") pod \"nmstate-console-plugin-5dcbbd79cf-4cfq5\" (UID: \"a82cc418-4f77-4788-addb-25341d42d9fa\") " pod="openshift-nmstate/nmstate-console-plugin-5dcbbd79cf-4cfq5" Feb 27 08:02:52 crc kubenswrapper[4612]: I0227 08:02:52.844338 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-786f45cff4-m8llg"] Feb 27 08:02:52 crc kubenswrapper[4612]: W0227 08:02:52.845627 4612 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb141dc9e_912e_4096_97fc_5df6e0ae245c.slice/crio-dd0ad76633268fe255b2365625e0cd4b8fa16602a76247cc309b10e37cd77091 WatchSource:0}: Error finding container dd0ad76633268fe255b2365625e0cd4b8fa16602a76247cc309b10e37cd77091: Status 404 returned error can't find the container with id dd0ad76633268fe255b2365625e0cd4b8fa16602a76247cc309b10e37cd77091 Feb 27 08:02:52 crc kubenswrapper[4612]: I0227 08:02:52.901391 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-26zsp" event={"ID":"5553e6c7-b45f-40c9-99ce-abd46767d31a","Type":"ContainerStarted","Data":"99b237fb5ee274b71416f4c01045ef6e7e23e03c91fdd38403c69def8309f17f"} Feb 27 08:02:52 crc kubenswrapper[4612]: I0227 08:02:52.908726 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-69594cc75-czssc" event={"ID":"d058174c-69c3-49cf-b846-af074b709972","Type":"ContainerStarted","Data":"5d86a2b7563e8901aff79fa0389f7dec21ff2549521a5452e67afd0185e54866"} Feb 27 08:02:52 crc kubenswrapper[4612]: I0227 08:02:52.910861 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-786f45cff4-m8llg" event={"ID":"b141dc9e-912e-4096-97fc-5df6e0ae245c","Type":"ContainerStarted","Data":"dd0ad76633268fe255b2365625e0cd4b8fa16602a76247cc309b10e37cd77091"} Feb 27 08:02:52 crc kubenswrapper[4612]: I0227 08:02:52.972938 4612 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-knw8s" podUID="67c16138-cb89-4526-a7d8-604a710da7e2" containerName="registry-server" probeResult="failure" output=< Feb 27 08:02:52 crc kubenswrapper[4612]: timeout: failed to connect service ":50051" within 1s Feb 27 08:02:52 crc kubenswrapper[4612]: > Feb 27 08:02:53 crc kubenswrapper[4612]: I0227 08:02:53.037246 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-5dcbbd79cf-4cfq5" Feb 27 08:02:53 crc kubenswrapper[4612]: I0227 08:02:53.233048 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-7c8f999976-kdvq5"] Feb 27 08:02:53 crc kubenswrapper[4612]: W0227 08:02:53.575147 4612 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podeb7933c3_204a_46b8_870f_4f776d70a03c.slice/crio-1a5de4e24c2154792e2e984e21308c587dcd9279e709d35c425559d6b0f7e837 WatchSource:0}: Error finding container 1a5de4e24c2154792e2e984e21308c587dcd9279e709d35c425559d6b0f7e837: Status 404 returned error can't find the container with id 1a5de4e24c2154792e2e984e21308c587dcd9279e709d35c425559d6b0f7e837 Feb 27 08:02:53 crc kubenswrapper[4612]: I0227 08:02:53.871173 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-5dcbbd79cf-4cfq5"] Feb 27 08:02:53 crc kubenswrapper[4612]: I0227 08:02:53.917011 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-5dcbbd79cf-4cfq5" event={"ID":"a82cc418-4f77-4788-addb-25341d42d9fa","Type":"ContainerStarted","Data":"cb98775be970314fee2a4d0e7dc3964fa9ae2d6f5dfe813977e6d2fb7cb853c2"} Feb 27 08:02:53 crc kubenswrapper[4612]: I0227 08:02:53.919083 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-7c8f999976-kdvq5" event={"ID":"eb7933c3-204a-46b8-870f-4f776d70a03c","Type":"ContainerStarted","Data":"e4696b6b378b2dd101a246766c75736367e8d575e052960d5bd2f7eee28701c2"} Feb 27 08:02:53 crc kubenswrapper[4612]: I0227 08:02:53.919131 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-7c8f999976-kdvq5" event={"ID":"eb7933c3-204a-46b8-870f-4f776d70a03c","Type":"ContainerStarted","Data":"1a5de4e24c2154792e2e984e21308c587dcd9279e709d35c425559d6b0f7e837"} Feb 27 08:02:53 crc kubenswrapper[4612]: I0227 08:02:53.943416 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-7c8f999976-kdvq5" podStartSLOduration=1.943393288 podStartE2EDuration="1.943393288s" podCreationTimestamp="2026-02-27 08:02:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 08:02:53.937662084 +0000 UTC m=+831.791592092" watchObservedRunningTime="2026-02-27 08:02:53.943393288 +0000 UTC m=+831.797323286" Feb 27 08:02:55 crc kubenswrapper[4612]: I0227 08:02:55.938701 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-786f45cff4-m8llg" event={"ID":"b141dc9e-912e-4096-97fc-5df6e0ae245c","Type":"ContainerStarted","Data":"160e7e8a0995e2a28c262bf550cfbc69a0903ca9eeda3e148b7f83c4f168eb38"} Feb 27 08:02:55 crc kubenswrapper[4612]: I0227 08:02:55.939394 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-786f45cff4-m8llg" Feb 27 08:02:55 crc kubenswrapper[4612]: I0227 08:02:55.946118 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-69594cc75-czssc" event={"ID":"d058174c-69c3-49cf-b846-af074b709972","Type":"ContainerStarted","Data":"87c26913fb350b249ed52b9993cb1cc61367c6fceb2beaf844ff5c7442e6e8b4"} Feb 27 08:02:55 crc kubenswrapper[4612]: I0227 08:02:55.963243 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-786f45cff4-m8llg" podStartSLOduration=2.365304647 podStartE2EDuration="4.963221432s" podCreationTimestamp="2026-02-27 08:02:51 +0000 UTC" firstStartedPulling="2026-02-27 08:02:52.849282074 +0000 UTC m=+830.703212072" lastFinishedPulling="2026-02-27 08:02:55.447198859 +0000 UTC m=+833.301128857" observedRunningTime="2026-02-27 08:02:55.956376126 +0000 UTC m=+833.810306124" watchObservedRunningTime="2026-02-27 08:02:55.963221432 +0000 UTC m=+833.817151430" Feb 27 08:02:56 crc kubenswrapper[4612]: I0227 08:02:56.955675 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-5dcbbd79cf-4cfq5" event={"ID":"a82cc418-4f77-4788-addb-25341d42d9fa","Type":"ContainerStarted","Data":"4ef2f09854bc78ef59129e84ec4e90b5236a1642c04b5ec2d138f1c8bcbb36fd"} Feb 27 08:02:56 crc kubenswrapper[4612]: I0227 08:02:56.957251 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-26zsp" event={"ID":"5553e6c7-b45f-40c9-99ce-abd46767d31a","Type":"ContainerStarted","Data":"29183025cba40113a39bf93187e42ebc01e2d93bbf341f25cc1d49bf425f99cc"} Feb 27 08:02:56 crc kubenswrapper[4612]: I0227 08:02:56.957383 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-26zsp" Feb 27 08:02:56 crc kubenswrapper[4612]: I0227 08:02:56.975133 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-5dcbbd79cf-4cfq5" podStartSLOduration=2.060969976 podStartE2EDuration="4.975115101s" podCreationTimestamp="2026-02-27 08:02:52 +0000 UTC" firstStartedPulling="2026-02-27 08:02:53.877457889 +0000 UTC m=+831.731387897" lastFinishedPulling="2026-02-27 08:02:56.791603024 +0000 UTC m=+834.645533022" observedRunningTime="2026-02-27 08:02:56.971484157 +0000 UTC m=+834.825414155" watchObservedRunningTime="2026-02-27 08:02:56.975115101 +0000 UTC m=+834.829045099" Feb 27 08:02:56 crc kubenswrapper[4612]: I0227 08:02:56.989215 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-26zsp" podStartSLOduration=2.986635926 podStartE2EDuration="5.989194124s" podCreationTimestamp="2026-02-27 08:02:51 +0000 UTC" firstStartedPulling="2026-02-27 08:02:52.41737688 +0000 UTC m=+830.271306878" lastFinishedPulling="2026-02-27 08:02:55.419935078 +0000 UTC m=+833.273865076" observedRunningTime="2026-02-27 08:02:56.987168526 +0000 UTC m=+834.841098524" watchObservedRunningTime="2026-02-27 08:02:56.989194124 +0000 UTC m=+834.843124132" Feb 27 08:02:58 crc kubenswrapper[4612]: I0227 08:02:58.973745 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-69594cc75-czssc" event={"ID":"d058174c-69c3-49cf-b846-af074b709972","Type":"ContainerStarted","Data":"585a6d2a94113566c02b3f97920efd984476aea58c0948dd643c3fa3074c935b"} Feb 27 08:02:59 crc kubenswrapper[4612]: I0227 08:02:59.001797 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-69594cc75-czssc" podStartSLOduration=2.454484461 podStartE2EDuration="8.001771222s" podCreationTimestamp="2026-02-27 08:02:51 +0000 UTC" firstStartedPulling="2026-02-27 08:02:52.594177565 +0000 UTC m=+830.448107553" lastFinishedPulling="2026-02-27 08:02:58.141464316 +0000 UTC m=+835.995394314" observedRunningTime="2026-02-27 08:02:58.99649847 +0000 UTC m=+836.850428508" watchObservedRunningTime="2026-02-27 08:02:59.001771222 +0000 UTC m=+836.855701280" Feb 27 08:03:00 crc kubenswrapper[4612]: I0227 08:03:00.803554 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-rrpnq"] Feb 27 08:03:00 crc kubenswrapper[4612]: I0227 08:03:00.804767 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rrpnq" Feb 27 08:03:00 crc kubenswrapper[4612]: I0227 08:03:00.830331 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-rrpnq"] Feb 27 08:03:00 crc kubenswrapper[4612]: I0227 08:03:00.966035 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/869d3ce9-173d-4aa6-8d95-d559dd273810-utilities\") pod \"community-operators-rrpnq\" (UID: \"869d3ce9-173d-4aa6-8d95-d559dd273810\") " pod="openshift-marketplace/community-operators-rrpnq" Feb 27 08:03:00 crc kubenswrapper[4612]: I0227 08:03:00.966124 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/869d3ce9-173d-4aa6-8d95-d559dd273810-catalog-content\") pod \"community-operators-rrpnq\" (UID: \"869d3ce9-173d-4aa6-8d95-d559dd273810\") " pod="openshift-marketplace/community-operators-rrpnq" Feb 27 08:03:00 crc kubenswrapper[4612]: I0227 08:03:00.966158 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-csrhg\" (UniqueName: \"kubernetes.io/projected/869d3ce9-173d-4aa6-8d95-d559dd273810-kube-api-access-csrhg\") pod \"community-operators-rrpnq\" (UID: \"869d3ce9-173d-4aa6-8d95-d559dd273810\") " pod="openshift-marketplace/community-operators-rrpnq" Feb 27 08:03:01 crc kubenswrapper[4612]: I0227 08:03:01.066972 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-csrhg\" (UniqueName: \"kubernetes.io/projected/869d3ce9-173d-4aa6-8d95-d559dd273810-kube-api-access-csrhg\") pod \"community-operators-rrpnq\" (UID: \"869d3ce9-173d-4aa6-8d95-d559dd273810\") " pod="openshift-marketplace/community-operators-rrpnq" Feb 27 08:03:01 crc kubenswrapper[4612]: I0227 08:03:01.067109 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/869d3ce9-173d-4aa6-8d95-d559dd273810-utilities\") pod \"community-operators-rrpnq\" (UID: \"869d3ce9-173d-4aa6-8d95-d559dd273810\") " pod="openshift-marketplace/community-operators-rrpnq" Feb 27 08:03:01 crc kubenswrapper[4612]: I0227 08:03:01.067174 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/869d3ce9-173d-4aa6-8d95-d559dd273810-catalog-content\") pod \"community-operators-rrpnq\" (UID: \"869d3ce9-173d-4aa6-8d95-d559dd273810\") " pod="openshift-marketplace/community-operators-rrpnq" Feb 27 08:03:01 crc kubenswrapper[4612]: I0227 08:03:01.067811 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/869d3ce9-173d-4aa6-8d95-d559dd273810-catalog-content\") pod \"community-operators-rrpnq\" (UID: \"869d3ce9-173d-4aa6-8d95-d559dd273810\") " pod="openshift-marketplace/community-operators-rrpnq" Feb 27 08:03:01 crc kubenswrapper[4612]: I0227 08:03:01.067972 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/869d3ce9-173d-4aa6-8d95-d559dd273810-utilities\") pod \"community-operators-rrpnq\" (UID: \"869d3ce9-173d-4aa6-8d95-d559dd273810\") " pod="openshift-marketplace/community-operators-rrpnq" Feb 27 08:03:01 crc kubenswrapper[4612]: I0227 08:03:01.095706 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-csrhg\" (UniqueName: \"kubernetes.io/projected/869d3ce9-173d-4aa6-8d95-d559dd273810-kube-api-access-csrhg\") pod \"community-operators-rrpnq\" (UID: \"869d3ce9-173d-4aa6-8d95-d559dd273810\") " pod="openshift-marketplace/community-operators-rrpnq" Feb 27 08:03:01 crc kubenswrapper[4612]: I0227 08:03:01.121937 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rrpnq" Feb 27 08:03:01 crc kubenswrapper[4612]: I0227 08:03:01.369418 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-rrpnq"] Feb 27 08:03:01 crc kubenswrapper[4612]: I0227 08:03:01.973418 4612 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-knw8s" Feb 27 08:03:02 crc kubenswrapper[4612]: I0227 08:03:02.004037 4612 generic.go:334] "Generic (PLEG): container finished" podID="869d3ce9-173d-4aa6-8d95-d559dd273810" containerID="bfb3a7784d3412e7c7931d4fa418d0ad7f73015bf8a37e0c5a5031707a83a83a" exitCode=0 Feb 27 08:03:02 crc kubenswrapper[4612]: I0227 08:03:02.004093 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rrpnq" event={"ID":"869d3ce9-173d-4aa6-8d95-d559dd273810","Type":"ContainerDied","Data":"bfb3a7784d3412e7c7931d4fa418d0ad7f73015bf8a37e0c5a5031707a83a83a"} Feb 27 08:03:02 crc kubenswrapper[4612]: I0227 08:03:02.004142 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rrpnq" event={"ID":"869d3ce9-173d-4aa6-8d95-d559dd273810","Type":"ContainerStarted","Data":"5055c8b430710ff513b25eceddcf410c3e39bd0f565ffd2877f7bf91124480b8"} Feb 27 08:03:02 crc kubenswrapper[4612]: I0227 08:03:02.008110 4612 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 27 08:03:02 crc kubenswrapper[4612]: I0227 08:03:02.047147 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-knw8s" Feb 27 08:03:02 crc kubenswrapper[4612]: I0227 08:03:02.353060 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-26zsp" Feb 27 08:03:02 crc kubenswrapper[4612]: I0227 08:03:02.752770 4612 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-7c8f999976-kdvq5" Feb 27 08:03:02 crc kubenswrapper[4612]: I0227 08:03:02.752841 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-7c8f999976-kdvq5" Feb 27 08:03:02 crc kubenswrapper[4612]: I0227 08:03:02.760316 4612 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-7c8f999976-kdvq5" Feb 27 08:03:03 crc kubenswrapper[4612]: I0227 08:03:03.015522 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rrpnq" event={"ID":"869d3ce9-173d-4aa6-8d95-d559dd273810","Type":"ContainerStarted","Data":"c52026743e61b3bcd600e4e47e97dc3eed2027f803c1388a68f1c99c84f09e46"} Feb 27 08:03:03 crc kubenswrapper[4612]: I0227 08:03:03.020867 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-7c8f999976-kdvq5" Feb 27 08:03:03 crc kubenswrapper[4612]: I0227 08:03:03.097701 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-8wk54"] Feb 27 08:03:03 crc kubenswrapper[4612]: I0227 08:03:03.611108 4612 scope.go:117] "RemoveContainer" containerID="6f038fa34ec7b8bedc6054ae35fc4e63cb940f189520a2151c51ca811dee4347" Feb 27 08:03:04 crc kubenswrapper[4612]: I0227 08:03:04.023257 4612 generic.go:334] "Generic (PLEG): container finished" podID="869d3ce9-173d-4aa6-8d95-d559dd273810" containerID="c52026743e61b3bcd600e4e47e97dc3eed2027f803c1388a68f1c99c84f09e46" exitCode=0 Feb 27 08:03:04 crc kubenswrapper[4612]: I0227 08:03:04.023423 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rrpnq" event={"ID":"869d3ce9-173d-4aa6-8d95-d559dd273810","Type":"ContainerDied","Data":"c52026743e61b3bcd600e4e47e97dc3eed2027f803c1388a68f1c99c84f09e46"} Feb 27 08:03:04 crc kubenswrapper[4612]: I0227 08:03:04.388350 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-knw8s"] Feb 27 08:03:04 crc kubenswrapper[4612]: I0227 08:03:04.389144 4612 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-knw8s" podUID="67c16138-cb89-4526-a7d8-604a710da7e2" containerName="registry-server" containerID="cri-o://3d0b3af6e88a68a1b77fc307e48d4b83663c2bfd1c7e7534ad0b08f73841d2c9" gracePeriod=2 Feb 27 08:03:04 crc kubenswrapper[4612]: I0227 08:03:04.809967 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-knw8s" Feb 27 08:03:04 crc kubenswrapper[4612]: I0227 08:03:04.822886 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d2mg5\" (UniqueName: \"kubernetes.io/projected/67c16138-cb89-4526-a7d8-604a710da7e2-kube-api-access-d2mg5\") pod \"67c16138-cb89-4526-a7d8-604a710da7e2\" (UID: \"67c16138-cb89-4526-a7d8-604a710da7e2\") " Feb 27 08:03:04 crc kubenswrapper[4612]: I0227 08:03:04.822941 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/67c16138-cb89-4526-a7d8-604a710da7e2-catalog-content\") pod \"67c16138-cb89-4526-a7d8-604a710da7e2\" (UID: \"67c16138-cb89-4526-a7d8-604a710da7e2\") " Feb 27 08:03:04 crc kubenswrapper[4612]: I0227 08:03:04.822971 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/67c16138-cb89-4526-a7d8-604a710da7e2-utilities\") pod \"67c16138-cb89-4526-a7d8-604a710da7e2\" (UID: \"67c16138-cb89-4526-a7d8-604a710da7e2\") " Feb 27 08:03:04 crc kubenswrapper[4612]: I0227 08:03:04.823924 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/67c16138-cb89-4526-a7d8-604a710da7e2-utilities" (OuterVolumeSpecName: "utilities") pod "67c16138-cb89-4526-a7d8-604a710da7e2" (UID: "67c16138-cb89-4526-a7d8-604a710da7e2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 08:03:04 crc kubenswrapper[4612]: I0227 08:03:04.844122 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/67c16138-cb89-4526-a7d8-604a710da7e2-kube-api-access-d2mg5" (OuterVolumeSpecName: "kube-api-access-d2mg5") pod "67c16138-cb89-4526-a7d8-604a710da7e2" (UID: "67c16138-cb89-4526-a7d8-604a710da7e2"). InnerVolumeSpecName "kube-api-access-d2mg5". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 08:03:04 crc kubenswrapper[4612]: I0227 08:03:04.924726 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d2mg5\" (UniqueName: \"kubernetes.io/projected/67c16138-cb89-4526-a7d8-604a710da7e2-kube-api-access-d2mg5\") on node \"crc\" DevicePath \"\"" Feb 27 08:03:04 crc kubenswrapper[4612]: I0227 08:03:04.924757 4612 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/67c16138-cb89-4526-a7d8-604a710da7e2-utilities\") on node \"crc\" DevicePath \"\"" Feb 27 08:03:04 crc kubenswrapper[4612]: I0227 08:03:04.957668 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/67c16138-cb89-4526-a7d8-604a710da7e2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "67c16138-cb89-4526-a7d8-604a710da7e2" (UID: "67c16138-cb89-4526-a7d8-604a710da7e2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 08:03:05 crc kubenswrapper[4612]: I0227 08:03:05.025949 4612 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/67c16138-cb89-4526-a7d8-604a710da7e2-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 27 08:03:05 crc kubenswrapper[4612]: I0227 08:03:05.033285 4612 generic.go:334] "Generic (PLEG): container finished" podID="67c16138-cb89-4526-a7d8-604a710da7e2" containerID="3d0b3af6e88a68a1b77fc307e48d4b83663c2bfd1c7e7534ad0b08f73841d2c9" exitCode=0 Feb 27 08:03:05 crc kubenswrapper[4612]: I0227 08:03:05.033343 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-knw8s" Feb 27 08:03:05 crc kubenswrapper[4612]: I0227 08:03:05.033352 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-knw8s" event={"ID":"67c16138-cb89-4526-a7d8-604a710da7e2","Type":"ContainerDied","Data":"3d0b3af6e88a68a1b77fc307e48d4b83663c2bfd1c7e7534ad0b08f73841d2c9"} Feb 27 08:03:05 crc kubenswrapper[4612]: I0227 08:03:05.033402 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-knw8s" event={"ID":"67c16138-cb89-4526-a7d8-604a710da7e2","Type":"ContainerDied","Data":"9aa87c3e9b8b1922430a5f311e2b619d91f62bd4848e0724f5c44fada60887f4"} Feb 27 08:03:05 crc kubenswrapper[4612]: I0227 08:03:05.033420 4612 scope.go:117] "RemoveContainer" containerID="3d0b3af6e88a68a1b77fc307e48d4b83663c2bfd1c7e7534ad0b08f73841d2c9" Feb 27 08:03:05 crc kubenswrapper[4612]: I0227 08:03:05.040113 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rrpnq" event={"ID":"869d3ce9-173d-4aa6-8d95-d559dd273810","Type":"ContainerStarted","Data":"a97bec672c157e41227b965b0e707b487b1820d8d489c686e2f61641f444c7af"} Feb 27 08:03:05 crc kubenswrapper[4612]: I0227 08:03:05.055865 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-rrpnq" podStartSLOduration=2.629056696 podStartE2EDuration="5.0558484s" podCreationTimestamp="2026-02-27 08:03:00 +0000 UTC" firstStartedPulling="2026-02-27 08:03:02.007751247 +0000 UTC m=+839.861681285" lastFinishedPulling="2026-02-27 08:03:04.434542931 +0000 UTC m=+842.288472989" observedRunningTime="2026-02-27 08:03:05.052580997 +0000 UTC m=+842.906510995" watchObservedRunningTime="2026-02-27 08:03:05.0558484 +0000 UTC m=+842.909778408" Feb 27 08:03:05 crc kubenswrapper[4612]: I0227 08:03:05.057658 4612 scope.go:117] "RemoveContainer" containerID="b01dacb0be9ca7368b260f13d2396841bc91268067a56ab34354ef04fa6767ce" Feb 27 08:03:05 crc kubenswrapper[4612]: I0227 08:03:05.084879 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-knw8s"] Feb 27 08:03:05 crc kubenswrapper[4612]: I0227 08:03:05.092866 4612 scope.go:117] "RemoveContainer" containerID="dc77ad940eb3a00087300237d4a7397a0c806a6e0b563d041b4898f334240328" Feb 27 08:03:05 crc kubenswrapper[4612]: I0227 08:03:05.092930 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-knw8s"] Feb 27 08:03:05 crc kubenswrapper[4612]: I0227 08:03:05.111257 4612 scope.go:117] "RemoveContainer" containerID="3d0b3af6e88a68a1b77fc307e48d4b83663c2bfd1c7e7534ad0b08f73841d2c9" Feb 27 08:03:05 crc kubenswrapper[4612]: E0227 08:03:05.111756 4612 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3d0b3af6e88a68a1b77fc307e48d4b83663c2bfd1c7e7534ad0b08f73841d2c9\": container with ID starting with 3d0b3af6e88a68a1b77fc307e48d4b83663c2bfd1c7e7534ad0b08f73841d2c9 not found: ID does not exist" containerID="3d0b3af6e88a68a1b77fc307e48d4b83663c2bfd1c7e7534ad0b08f73841d2c9" Feb 27 08:03:05 crc kubenswrapper[4612]: I0227 08:03:05.111789 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3d0b3af6e88a68a1b77fc307e48d4b83663c2bfd1c7e7534ad0b08f73841d2c9"} err="failed to get container status \"3d0b3af6e88a68a1b77fc307e48d4b83663c2bfd1c7e7534ad0b08f73841d2c9\": rpc error: code = NotFound desc = could not find container \"3d0b3af6e88a68a1b77fc307e48d4b83663c2bfd1c7e7534ad0b08f73841d2c9\": container with ID starting with 3d0b3af6e88a68a1b77fc307e48d4b83663c2bfd1c7e7534ad0b08f73841d2c9 not found: ID does not exist" Feb 27 08:03:05 crc kubenswrapper[4612]: I0227 08:03:05.111811 4612 scope.go:117] "RemoveContainer" containerID="b01dacb0be9ca7368b260f13d2396841bc91268067a56ab34354ef04fa6767ce" Feb 27 08:03:05 crc kubenswrapper[4612]: E0227 08:03:05.112027 4612 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b01dacb0be9ca7368b260f13d2396841bc91268067a56ab34354ef04fa6767ce\": container with ID starting with b01dacb0be9ca7368b260f13d2396841bc91268067a56ab34354ef04fa6767ce not found: ID does not exist" containerID="b01dacb0be9ca7368b260f13d2396841bc91268067a56ab34354ef04fa6767ce" Feb 27 08:03:05 crc kubenswrapper[4612]: I0227 08:03:05.112049 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b01dacb0be9ca7368b260f13d2396841bc91268067a56ab34354ef04fa6767ce"} err="failed to get container status \"b01dacb0be9ca7368b260f13d2396841bc91268067a56ab34354ef04fa6767ce\": rpc error: code = NotFound desc = could not find container \"b01dacb0be9ca7368b260f13d2396841bc91268067a56ab34354ef04fa6767ce\": container with ID starting with b01dacb0be9ca7368b260f13d2396841bc91268067a56ab34354ef04fa6767ce not found: ID does not exist" Feb 27 08:03:05 crc kubenswrapper[4612]: I0227 08:03:05.112065 4612 scope.go:117] "RemoveContainer" containerID="dc77ad940eb3a00087300237d4a7397a0c806a6e0b563d041b4898f334240328" Feb 27 08:03:05 crc kubenswrapper[4612]: E0227 08:03:05.112312 4612 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dc77ad940eb3a00087300237d4a7397a0c806a6e0b563d041b4898f334240328\": container with ID starting with dc77ad940eb3a00087300237d4a7397a0c806a6e0b563d041b4898f334240328 not found: ID does not exist" containerID="dc77ad940eb3a00087300237d4a7397a0c806a6e0b563d041b4898f334240328" Feb 27 08:03:05 crc kubenswrapper[4612]: I0227 08:03:05.112332 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dc77ad940eb3a00087300237d4a7397a0c806a6e0b563d041b4898f334240328"} err="failed to get container status \"dc77ad940eb3a00087300237d4a7397a0c806a6e0b563d041b4898f334240328\": rpc error: code = NotFound desc = could not find container \"dc77ad940eb3a00087300237d4a7397a0c806a6e0b563d041b4898f334240328\": container with ID starting with dc77ad940eb3a00087300237d4a7397a0c806a6e0b563d041b4898f334240328 not found: ID does not exist" Feb 27 08:03:06 crc kubenswrapper[4612]: I0227 08:03:06.863091 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="67c16138-cb89-4526-a7d8-604a710da7e2" path="/var/lib/kubelet/pods/67c16138-cb89-4526-a7d8-604a710da7e2/volumes" Feb 27 08:03:09 crc kubenswrapper[4612]: I0227 08:03:09.408675 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-rlnv8"] Feb 27 08:03:09 crc kubenswrapper[4612]: E0227 08:03:09.410280 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="67c16138-cb89-4526-a7d8-604a710da7e2" containerName="extract-content" Feb 27 08:03:09 crc kubenswrapper[4612]: I0227 08:03:09.410406 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="67c16138-cb89-4526-a7d8-604a710da7e2" containerName="extract-content" Feb 27 08:03:09 crc kubenswrapper[4612]: E0227 08:03:09.410544 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="67c16138-cb89-4526-a7d8-604a710da7e2" containerName="registry-server" Feb 27 08:03:09 crc kubenswrapper[4612]: I0227 08:03:09.410643 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="67c16138-cb89-4526-a7d8-604a710da7e2" containerName="registry-server" Feb 27 08:03:09 crc kubenswrapper[4612]: E0227 08:03:09.410781 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="67c16138-cb89-4526-a7d8-604a710da7e2" containerName="extract-utilities" Feb 27 08:03:09 crc kubenswrapper[4612]: I0227 08:03:09.410884 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="67c16138-cb89-4526-a7d8-604a710da7e2" containerName="extract-utilities" Feb 27 08:03:09 crc kubenswrapper[4612]: I0227 08:03:09.411137 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="67c16138-cb89-4526-a7d8-604a710da7e2" containerName="registry-server" Feb 27 08:03:09 crc kubenswrapper[4612]: I0227 08:03:09.412660 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rlnv8" Feb 27 08:03:09 crc kubenswrapper[4612]: I0227 08:03:09.415176 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-rlnv8"] Feb 27 08:03:09 crc kubenswrapper[4612]: I0227 08:03:09.586676 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5825c2bd-486a-4502-b612-0635c30aed60-catalog-content\") pod \"redhat-marketplace-rlnv8\" (UID: \"5825c2bd-486a-4502-b612-0635c30aed60\") " pod="openshift-marketplace/redhat-marketplace-rlnv8" Feb 27 08:03:09 crc kubenswrapper[4612]: I0227 08:03:09.586836 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5825c2bd-486a-4502-b612-0635c30aed60-utilities\") pod \"redhat-marketplace-rlnv8\" (UID: \"5825c2bd-486a-4502-b612-0635c30aed60\") " pod="openshift-marketplace/redhat-marketplace-rlnv8" Feb 27 08:03:09 crc kubenswrapper[4612]: I0227 08:03:09.586865 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hlshk\" (UniqueName: \"kubernetes.io/projected/5825c2bd-486a-4502-b612-0635c30aed60-kube-api-access-hlshk\") pod \"redhat-marketplace-rlnv8\" (UID: \"5825c2bd-486a-4502-b612-0635c30aed60\") " pod="openshift-marketplace/redhat-marketplace-rlnv8" Feb 27 08:03:09 crc kubenswrapper[4612]: I0227 08:03:09.687956 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5825c2bd-486a-4502-b612-0635c30aed60-utilities\") pod \"redhat-marketplace-rlnv8\" (UID: \"5825c2bd-486a-4502-b612-0635c30aed60\") " pod="openshift-marketplace/redhat-marketplace-rlnv8" Feb 27 08:03:09 crc kubenswrapper[4612]: I0227 08:03:09.688264 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hlshk\" (UniqueName: \"kubernetes.io/projected/5825c2bd-486a-4502-b612-0635c30aed60-kube-api-access-hlshk\") pod \"redhat-marketplace-rlnv8\" (UID: \"5825c2bd-486a-4502-b612-0635c30aed60\") " pod="openshift-marketplace/redhat-marketplace-rlnv8" Feb 27 08:03:09 crc kubenswrapper[4612]: I0227 08:03:09.688371 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5825c2bd-486a-4502-b612-0635c30aed60-utilities\") pod \"redhat-marketplace-rlnv8\" (UID: \"5825c2bd-486a-4502-b612-0635c30aed60\") " pod="openshift-marketplace/redhat-marketplace-rlnv8" Feb 27 08:03:09 crc kubenswrapper[4612]: I0227 08:03:09.688369 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5825c2bd-486a-4502-b612-0635c30aed60-catalog-content\") pod \"redhat-marketplace-rlnv8\" (UID: \"5825c2bd-486a-4502-b612-0635c30aed60\") " pod="openshift-marketplace/redhat-marketplace-rlnv8" Feb 27 08:03:09 crc kubenswrapper[4612]: I0227 08:03:09.688680 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5825c2bd-486a-4502-b612-0635c30aed60-catalog-content\") pod \"redhat-marketplace-rlnv8\" (UID: \"5825c2bd-486a-4502-b612-0635c30aed60\") " pod="openshift-marketplace/redhat-marketplace-rlnv8" Feb 27 08:03:09 crc kubenswrapper[4612]: I0227 08:03:09.706539 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hlshk\" (UniqueName: \"kubernetes.io/projected/5825c2bd-486a-4502-b612-0635c30aed60-kube-api-access-hlshk\") pod \"redhat-marketplace-rlnv8\" (UID: \"5825c2bd-486a-4502-b612-0635c30aed60\") " pod="openshift-marketplace/redhat-marketplace-rlnv8" Feb 27 08:03:09 crc kubenswrapper[4612]: I0227 08:03:09.756067 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rlnv8" Feb 27 08:03:10 crc kubenswrapper[4612]: I0227 08:03:10.175186 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-rlnv8"] Feb 27 08:03:11 crc kubenswrapper[4612]: I0227 08:03:11.095228 4612 generic.go:334] "Generic (PLEG): container finished" podID="5825c2bd-486a-4502-b612-0635c30aed60" containerID="58e388a89e5476c2aebdfd45168b9cee651d55a20cfb2d3e30093d2d595b565c" exitCode=0 Feb 27 08:03:11 crc kubenswrapper[4612]: I0227 08:03:11.095296 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rlnv8" event={"ID":"5825c2bd-486a-4502-b612-0635c30aed60","Type":"ContainerDied","Data":"58e388a89e5476c2aebdfd45168b9cee651d55a20cfb2d3e30093d2d595b565c"} Feb 27 08:03:11 crc kubenswrapper[4612]: I0227 08:03:11.095364 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rlnv8" event={"ID":"5825c2bd-486a-4502-b612-0635c30aed60","Type":"ContainerStarted","Data":"8f9c4e51a6f2c2cb203b905dd32449ad1d58011cf96e9d031ff34b384d728ff1"} Feb 27 08:03:11 crc kubenswrapper[4612]: I0227 08:03:11.125277 4612 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-rrpnq" Feb 27 08:03:11 crc kubenswrapper[4612]: I0227 08:03:11.125336 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-rrpnq" Feb 27 08:03:11 crc kubenswrapper[4612]: I0227 08:03:11.185020 4612 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-rrpnq" Feb 27 08:03:12 crc kubenswrapper[4612]: I0227 08:03:12.137677 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-rrpnq" Feb 27 08:03:12 crc kubenswrapper[4612]: I0227 08:03:12.299668 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-786f45cff4-m8llg" Feb 27 08:03:13 crc kubenswrapper[4612]: I0227 08:03:13.123031 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rlnv8" event={"ID":"5825c2bd-486a-4502-b612-0635c30aed60","Type":"ContainerStarted","Data":"439dfac276d54694cd44ae8df83d2411e5a621f9467be5ea9d13a1064d89bb91"} Feb 27 08:03:13 crc kubenswrapper[4612]: I0227 08:03:13.579123 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-rrpnq"] Feb 27 08:03:14 crc kubenswrapper[4612]: I0227 08:03:14.129742 4612 generic.go:334] "Generic (PLEG): container finished" podID="5825c2bd-486a-4502-b612-0635c30aed60" containerID="439dfac276d54694cd44ae8df83d2411e5a621f9467be5ea9d13a1064d89bb91" exitCode=0 Feb 27 08:03:14 crc kubenswrapper[4612]: I0227 08:03:14.130164 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rlnv8" event={"ID":"5825c2bd-486a-4502-b612-0635c30aed60","Type":"ContainerDied","Data":"439dfac276d54694cd44ae8df83d2411e5a621f9467be5ea9d13a1064d89bb91"} Feb 27 08:03:15 crc kubenswrapper[4612]: I0227 08:03:15.139779 4612 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-rrpnq" podUID="869d3ce9-173d-4aa6-8d95-d559dd273810" containerName="registry-server" containerID="cri-o://a97bec672c157e41227b965b0e707b487b1820d8d489c686e2f61641f444c7af" gracePeriod=2 Feb 27 08:03:15 crc kubenswrapper[4612]: I0227 08:03:15.139819 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rlnv8" event={"ID":"5825c2bd-486a-4502-b612-0635c30aed60","Type":"ContainerStarted","Data":"44e0636151d4ad48c49f13c1e2ffb564817fc2cd9bfff5956094ebd83073f2d6"} Feb 27 08:03:15 crc kubenswrapper[4612]: I0227 08:03:15.526418 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rrpnq" Feb 27 08:03:15 crc kubenswrapper[4612]: I0227 08:03:15.551720 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-rlnv8" podStartSLOduration=3.06569831 podStartE2EDuration="6.551674477s" podCreationTimestamp="2026-02-27 08:03:09 +0000 UTC" firstStartedPulling="2026-02-27 08:03:11.096997878 +0000 UTC m=+848.950927916" lastFinishedPulling="2026-02-27 08:03:14.582974085 +0000 UTC m=+852.436904083" observedRunningTime="2026-02-27 08:03:15.170340692 +0000 UTC m=+853.024270700" watchObservedRunningTime="2026-02-27 08:03:15.551674477 +0000 UTC m=+853.405604485" Feb 27 08:03:15 crc kubenswrapper[4612]: I0227 08:03:15.663570 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-csrhg\" (UniqueName: \"kubernetes.io/projected/869d3ce9-173d-4aa6-8d95-d559dd273810-kube-api-access-csrhg\") pod \"869d3ce9-173d-4aa6-8d95-d559dd273810\" (UID: \"869d3ce9-173d-4aa6-8d95-d559dd273810\") " Feb 27 08:03:15 crc kubenswrapper[4612]: I0227 08:03:15.663683 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/869d3ce9-173d-4aa6-8d95-d559dd273810-utilities\") pod \"869d3ce9-173d-4aa6-8d95-d559dd273810\" (UID: \"869d3ce9-173d-4aa6-8d95-d559dd273810\") " Feb 27 08:03:15 crc kubenswrapper[4612]: I0227 08:03:15.663759 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/869d3ce9-173d-4aa6-8d95-d559dd273810-catalog-content\") pod \"869d3ce9-173d-4aa6-8d95-d559dd273810\" (UID: \"869d3ce9-173d-4aa6-8d95-d559dd273810\") " Feb 27 08:03:15 crc kubenswrapper[4612]: I0227 08:03:15.664942 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/869d3ce9-173d-4aa6-8d95-d559dd273810-utilities" (OuterVolumeSpecName: "utilities") pod "869d3ce9-173d-4aa6-8d95-d559dd273810" (UID: "869d3ce9-173d-4aa6-8d95-d559dd273810"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 08:03:15 crc kubenswrapper[4612]: I0227 08:03:15.669377 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/869d3ce9-173d-4aa6-8d95-d559dd273810-kube-api-access-csrhg" (OuterVolumeSpecName: "kube-api-access-csrhg") pod "869d3ce9-173d-4aa6-8d95-d559dd273810" (UID: "869d3ce9-173d-4aa6-8d95-d559dd273810"). InnerVolumeSpecName "kube-api-access-csrhg". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 08:03:15 crc kubenswrapper[4612]: I0227 08:03:15.726763 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/869d3ce9-173d-4aa6-8d95-d559dd273810-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "869d3ce9-173d-4aa6-8d95-d559dd273810" (UID: "869d3ce9-173d-4aa6-8d95-d559dd273810"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 08:03:15 crc kubenswrapper[4612]: I0227 08:03:15.765365 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-csrhg\" (UniqueName: \"kubernetes.io/projected/869d3ce9-173d-4aa6-8d95-d559dd273810-kube-api-access-csrhg\") on node \"crc\" DevicePath \"\"" Feb 27 08:03:15 crc kubenswrapper[4612]: I0227 08:03:15.765397 4612 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/869d3ce9-173d-4aa6-8d95-d559dd273810-utilities\") on node \"crc\" DevicePath \"\"" Feb 27 08:03:15 crc kubenswrapper[4612]: I0227 08:03:15.765407 4612 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/869d3ce9-173d-4aa6-8d95-d559dd273810-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 27 08:03:16 crc kubenswrapper[4612]: I0227 08:03:16.147388 4612 generic.go:334] "Generic (PLEG): container finished" podID="869d3ce9-173d-4aa6-8d95-d559dd273810" containerID="a97bec672c157e41227b965b0e707b487b1820d8d489c686e2f61641f444c7af" exitCode=0 Feb 27 08:03:16 crc kubenswrapper[4612]: I0227 08:03:16.147472 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rrpnq" Feb 27 08:03:16 crc kubenswrapper[4612]: I0227 08:03:16.147530 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rrpnq" event={"ID":"869d3ce9-173d-4aa6-8d95-d559dd273810","Type":"ContainerDied","Data":"a97bec672c157e41227b965b0e707b487b1820d8d489c686e2f61641f444c7af"} Feb 27 08:03:16 crc kubenswrapper[4612]: I0227 08:03:16.147574 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rrpnq" event={"ID":"869d3ce9-173d-4aa6-8d95-d559dd273810","Type":"ContainerDied","Data":"5055c8b430710ff513b25eceddcf410c3e39bd0f565ffd2877f7bf91124480b8"} Feb 27 08:03:16 crc kubenswrapper[4612]: I0227 08:03:16.147597 4612 scope.go:117] "RemoveContainer" containerID="a97bec672c157e41227b965b0e707b487b1820d8d489c686e2f61641f444c7af" Feb 27 08:03:16 crc kubenswrapper[4612]: I0227 08:03:16.165101 4612 scope.go:117] "RemoveContainer" containerID="c52026743e61b3bcd600e4e47e97dc3eed2027f803c1388a68f1c99c84f09e46" Feb 27 08:03:16 crc kubenswrapper[4612]: I0227 08:03:16.204102 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-rrpnq"] Feb 27 08:03:16 crc kubenswrapper[4612]: I0227 08:03:16.205729 4612 scope.go:117] "RemoveContainer" containerID="bfb3a7784d3412e7c7931d4fa418d0ad7f73015bf8a37e0c5a5031707a83a83a" Feb 27 08:03:16 crc kubenswrapper[4612]: I0227 08:03:16.208855 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-rrpnq"] Feb 27 08:03:16 crc kubenswrapper[4612]: I0227 08:03:16.223804 4612 scope.go:117] "RemoveContainer" containerID="a97bec672c157e41227b965b0e707b487b1820d8d489c686e2f61641f444c7af" Feb 27 08:03:16 crc kubenswrapper[4612]: E0227 08:03:16.224190 4612 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a97bec672c157e41227b965b0e707b487b1820d8d489c686e2f61641f444c7af\": container with ID starting with a97bec672c157e41227b965b0e707b487b1820d8d489c686e2f61641f444c7af not found: ID does not exist" containerID="a97bec672c157e41227b965b0e707b487b1820d8d489c686e2f61641f444c7af" Feb 27 08:03:16 crc kubenswrapper[4612]: I0227 08:03:16.224225 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a97bec672c157e41227b965b0e707b487b1820d8d489c686e2f61641f444c7af"} err="failed to get container status \"a97bec672c157e41227b965b0e707b487b1820d8d489c686e2f61641f444c7af\": rpc error: code = NotFound desc = could not find container \"a97bec672c157e41227b965b0e707b487b1820d8d489c686e2f61641f444c7af\": container with ID starting with a97bec672c157e41227b965b0e707b487b1820d8d489c686e2f61641f444c7af not found: ID does not exist" Feb 27 08:03:16 crc kubenswrapper[4612]: I0227 08:03:16.224254 4612 scope.go:117] "RemoveContainer" containerID="c52026743e61b3bcd600e4e47e97dc3eed2027f803c1388a68f1c99c84f09e46" Feb 27 08:03:16 crc kubenswrapper[4612]: E0227 08:03:16.224762 4612 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c52026743e61b3bcd600e4e47e97dc3eed2027f803c1388a68f1c99c84f09e46\": container with ID starting with c52026743e61b3bcd600e4e47e97dc3eed2027f803c1388a68f1c99c84f09e46 not found: ID does not exist" containerID="c52026743e61b3bcd600e4e47e97dc3eed2027f803c1388a68f1c99c84f09e46" Feb 27 08:03:16 crc kubenswrapper[4612]: I0227 08:03:16.224799 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c52026743e61b3bcd600e4e47e97dc3eed2027f803c1388a68f1c99c84f09e46"} err="failed to get container status \"c52026743e61b3bcd600e4e47e97dc3eed2027f803c1388a68f1c99c84f09e46\": rpc error: code = NotFound desc = could not find container \"c52026743e61b3bcd600e4e47e97dc3eed2027f803c1388a68f1c99c84f09e46\": container with ID starting with c52026743e61b3bcd600e4e47e97dc3eed2027f803c1388a68f1c99c84f09e46 not found: ID does not exist" Feb 27 08:03:16 crc kubenswrapper[4612]: I0227 08:03:16.224815 4612 scope.go:117] "RemoveContainer" containerID="bfb3a7784d3412e7c7931d4fa418d0ad7f73015bf8a37e0c5a5031707a83a83a" Feb 27 08:03:16 crc kubenswrapper[4612]: E0227 08:03:16.225242 4612 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bfb3a7784d3412e7c7931d4fa418d0ad7f73015bf8a37e0c5a5031707a83a83a\": container with ID starting with bfb3a7784d3412e7c7931d4fa418d0ad7f73015bf8a37e0c5a5031707a83a83a not found: ID does not exist" containerID="bfb3a7784d3412e7c7931d4fa418d0ad7f73015bf8a37e0c5a5031707a83a83a" Feb 27 08:03:16 crc kubenswrapper[4612]: I0227 08:03:16.225266 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bfb3a7784d3412e7c7931d4fa418d0ad7f73015bf8a37e0c5a5031707a83a83a"} err="failed to get container status \"bfb3a7784d3412e7c7931d4fa418d0ad7f73015bf8a37e0c5a5031707a83a83a\": rpc error: code = NotFound desc = could not find container \"bfb3a7784d3412e7c7931d4fa418d0ad7f73015bf8a37e0c5a5031707a83a83a\": container with ID starting with bfb3a7784d3412e7c7931d4fa418d0ad7f73015bf8a37e0c5a5031707a83a83a not found: ID does not exist" Feb 27 08:03:16 crc kubenswrapper[4612]: I0227 08:03:16.864837 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="869d3ce9-173d-4aa6-8d95-d559dd273810" path="/var/lib/kubelet/pods/869d3ce9-173d-4aa6-8d95-d559dd273810/volumes" Feb 27 08:03:19 crc kubenswrapper[4612]: I0227 08:03:19.760464 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-rlnv8" Feb 27 08:03:19 crc kubenswrapper[4612]: I0227 08:03:19.762813 4612 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-rlnv8" Feb 27 08:03:19 crc kubenswrapper[4612]: I0227 08:03:19.814128 4612 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-rlnv8" Feb 27 08:03:20 crc kubenswrapper[4612]: I0227 08:03:20.214941 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-rlnv8" Feb 27 08:03:20 crc kubenswrapper[4612]: I0227 08:03:20.778101 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-rlnv8"] Feb 27 08:03:22 crc kubenswrapper[4612]: I0227 08:03:22.184036 4612 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-rlnv8" podUID="5825c2bd-486a-4502-b612-0635c30aed60" containerName="registry-server" containerID="cri-o://44e0636151d4ad48c49f13c1e2ffb564817fc2cd9bfff5956094ebd83073f2d6" gracePeriod=2 Feb 27 08:03:22 crc kubenswrapper[4612]: I0227 08:03:22.596653 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rlnv8" Feb 27 08:03:22 crc kubenswrapper[4612]: I0227 08:03:22.761595 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5825c2bd-486a-4502-b612-0635c30aed60-utilities\") pod \"5825c2bd-486a-4502-b612-0635c30aed60\" (UID: \"5825c2bd-486a-4502-b612-0635c30aed60\") " Feb 27 08:03:22 crc kubenswrapper[4612]: I0227 08:03:22.761687 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5825c2bd-486a-4502-b612-0635c30aed60-catalog-content\") pod \"5825c2bd-486a-4502-b612-0635c30aed60\" (UID: \"5825c2bd-486a-4502-b612-0635c30aed60\") " Feb 27 08:03:22 crc kubenswrapper[4612]: I0227 08:03:22.761909 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hlshk\" (UniqueName: \"kubernetes.io/projected/5825c2bd-486a-4502-b612-0635c30aed60-kube-api-access-hlshk\") pod \"5825c2bd-486a-4502-b612-0635c30aed60\" (UID: \"5825c2bd-486a-4502-b612-0635c30aed60\") " Feb 27 08:03:22 crc kubenswrapper[4612]: I0227 08:03:22.762909 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5825c2bd-486a-4502-b612-0635c30aed60-utilities" (OuterVolumeSpecName: "utilities") pod "5825c2bd-486a-4502-b612-0635c30aed60" (UID: "5825c2bd-486a-4502-b612-0635c30aed60"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 08:03:22 crc kubenswrapper[4612]: I0227 08:03:22.770169 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5825c2bd-486a-4502-b612-0635c30aed60-kube-api-access-hlshk" (OuterVolumeSpecName: "kube-api-access-hlshk") pod "5825c2bd-486a-4502-b612-0635c30aed60" (UID: "5825c2bd-486a-4502-b612-0635c30aed60"). InnerVolumeSpecName "kube-api-access-hlshk". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 08:03:22 crc kubenswrapper[4612]: I0227 08:03:22.789440 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5825c2bd-486a-4502-b612-0635c30aed60-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5825c2bd-486a-4502-b612-0635c30aed60" (UID: "5825c2bd-486a-4502-b612-0635c30aed60"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 08:03:22 crc kubenswrapper[4612]: I0227 08:03:22.867300 4612 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5825c2bd-486a-4502-b612-0635c30aed60-utilities\") on node \"crc\" DevicePath \"\"" Feb 27 08:03:22 crc kubenswrapper[4612]: I0227 08:03:22.867486 4612 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5825c2bd-486a-4502-b612-0635c30aed60-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 27 08:03:22 crc kubenswrapper[4612]: I0227 08:03:22.867599 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hlshk\" (UniqueName: \"kubernetes.io/projected/5825c2bd-486a-4502-b612-0635c30aed60-kube-api-access-hlshk\") on node \"crc\" DevicePath \"\"" Feb 27 08:03:23 crc kubenswrapper[4612]: I0227 08:03:23.196728 4612 generic.go:334] "Generic (PLEG): container finished" podID="5825c2bd-486a-4502-b612-0635c30aed60" containerID="44e0636151d4ad48c49f13c1e2ffb564817fc2cd9bfff5956094ebd83073f2d6" exitCode=0 Feb 27 08:03:23 crc kubenswrapper[4612]: I0227 08:03:23.196885 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rlnv8" Feb 27 08:03:23 crc kubenswrapper[4612]: I0227 08:03:23.196886 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rlnv8" event={"ID":"5825c2bd-486a-4502-b612-0635c30aed60","Type":"ContainerDied","Data":"44e0636151d4ad48c49f13c1e2ffb564817fc2cd9bfff5956094ebd83073f2d6"} Feb 27 08:03:23 crc kubenswrapper[4612]: I0227 08:03:23.197338 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rlnv8" event={"ID":"5825c2bd-486a-4502-b612-0635c30aed60","Type":"ContainerDied","Data":"8f9c4e51a6f2c2cb203b905dd32449ad1d58011cf96e9d031ff34b384d728ff1"} Feb 27 08:03:23 crc kubenswrapper[4612]: I0227 08:03:23.197400 4612 scope.go:117] "RemoveContainer" containerID="44e0636151d4ad48c49f13c1e2ffb564817fc2cd9bfff5956094ebd83073f2d6" Feb 27 08:03:23 crc kubenswrapper[4612]: I0227 08:03:23.227438 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-rlnv8"] Feb 27 08:03:23 crc kubenswrapper[4612]: I0227 08:03:23.233251 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-rlnv8"] Feb 27 08:03:23 crc kubenswrapper[4612]: I0227 08:03:23.248614 4612 scope.go:117] "RemoveContainer" containerID="439dfac276d54694cd44ae8df83d2411e5a621f9467be5ea9d13a1064d89bb91" Feb 27 08:03:23 crc kubenswrapper[4612]: I0227 08:03:23.266195 4612 scope.go:117] "RemoveContainer" containerID="58e388a89e5476c2aebdfd45168b9cee651d55a20cfb2d3e30093d2d595b565c" Feb 27 08:03:23 crc kubenswrapper[4612]: I0227 08:03:23.286043 4612 scope.go:117] "RemoveContainer" containerID="44e0636151d4ad48c49f13c1e2ffb564817fc2cd9bfff5956094ebd83073f2d6" Feb 27 08:03:23 crc kubenswrapper[4612]: E0227 08:03:23.286637 4612 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"44e0636151d4ad48c49f13c1e2ffb564817fc2cd9bfff5956094ebd83073f2d6\": container with ID starting with 44e0636151d4ad48c49f13c1e2ffb564817fc2cd9bfff5956094ebd83073f2d6 not found: ID does not exist" containerID="44e0636151d4ad48c49f13c1e2ffb564817fc2cd9bfff5956094ebd83073f2d6" Feb 27 08:03:23 crc kubenswrapper[4612]: I0227 08:03:23.286678 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"44e0636151d4ad48c49f13c1e2ffb564817fc2cd9bfff5956094ebd83073f2d6"} err="failed to get container status \"44e0636151d4ad48c49f13c1e2ffb564817fc2cd9bfff5956094ebd83073f2d6\": rpc error: code = NotFound desc = could not find container \"44e0636151d4ad48c49f13c1e2ffb564817fc2cd9bfff5956094ebd83073f2d6\": container with ID starting with 44e0636151d4ad48c49f13c1e2ffb564817fc2cd9bfff5956094ebd83073f2d6 not found: ID does not exist" Feb 27 08:03:23 crc kubenswrapper[4612]: I0227 08:03:23.286717 4612 scope.go:117] "RemoveContainer" containerID="439dfac276d54694cd44ae8df83d2411e5a621f9467be5ea9d13a1064d89bb91" Feb 27 08:03:23 crc kubenswrapper[4612]: E0227 08:03:23.287185 4612 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"439dfac276d54694cd44ae8df83d2411e5a621f9467be5ea9d13a1064d89bb91\": container with ID starting with 439dfac276d54694cd44ae8df83d2411e5a621f9467be5ea9d13a1064d89bb91 not found: ID does not exist" containerID="439dfac276d54694cd44ae8df83d2411e5a621f9467be5ea9d13a1064d89bb91" Feb 27 08:03:23 crc kubenswrapper[4612]: I0227 08:03:23.287233 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"439dfac276d54694cd44ae8df83d2411e5a621f9467be5ea9d13a1064d89bb91"} err="failed to get container status \"439dfac276d54694cd44ae8df83d2411e5a621f9467be5ea9d13a1064d89bb91\": rpc error: code = NotFound desc = could not find container \"439dfac276d54694cd44ae8df83d2411e5a621f9467be5ea9d13a1064d89bb91\": container with ID starting with 439dfac276d54694cd44ae8df83d2411e5a621f9467be5ea9d13a1064d89bb91 not found: ID does not exist" Feb 27 08:03:23 crc kubenswrapper[4612]: I0227 08:03:23.287271 4612 scope.go:117] "RemoveContainer" containerID="58e388a89e5476c2aebdfd45168b9cee651d55a20cfb2d3e30093d2d595b565c" Feb 27 08:03:23 crc kubenswrapper[4612]: E0227 08:03:23.287575 4612 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"58e388a89e5476c2aebdfd45168b9cee651d55a20cfb2d3e30093d2d595b565c\": container with ID starting with 58e388a89e5476c2aebdfd45168b9cee651d55a20cfb2d3e30093d2d595b565c not found: ID does not exist" containerID="58e388a89e5476c2aebdfd45168b9cee651d55a20cfb2d3e30093d2d595b565c" Feb 27 08:03:23 crc kubenswrapper[4612]: I0227 08:03:23.287597 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"58e388a89e5476c2aebdfd45168b9cee651d55a20cfb2d3e30093d2d595b565c"} err="failed to get container status \"58e388a89e5476c2aebdfd45168b9cee651d55a20cfb2d3e30093d2d595b565c\": rpc error: code = NotFound desc = could not find container \"58e388a89e5476c2aebdfd45168b9cee651d55a20cfb2d3e30093d2d595b565c\": container with ID starting with 58e388a89e5476c2aebdfd45168b9cee651d55a20cfb2d3e30093d2d595b565c not found: ID does not exist" Feb 27 08:03:24 crc kubenswrapper[4612]: I0227 08:03:24.865065 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5825c2bd-486a-4502-b612-0635c30aed60" path="/var/lib/kubelet/pods/5825c2bd-486a-4502-b612-0635c30aed60/volumes" Feb 27 08:03:25 crc kubenswrapper[4612]: I0227 08:03:25.614529 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4nq7w2"] Feb 27 08:03:25 crc kubenswrapper[4612]: E0227 08:03:25.614974 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="869d3ce9-173d-4aa6-8d95-d559dd273810" containerName="registry-server" Feb 27 08:03:25 crc kubenswrapper[4612]: I0227 08:03:25.615038 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="869d3ce9-173d-4aa6-8d95-d559dd273810" containerName="registry-server" Feb 27 08:03:25 crc kubenswrapper[4612]: E0227 08:03:25.615099 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="869d3ce9-173d-4aa6-8d95-d559dd273810" containerName="extract-utilities" Feb 27 08:03:25 crc kubenswrapper[4612]: I0227 08:03:25.615152 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="869d3ce9-173d-4aa6-8d95-d559dd273810" containerName="extract-utilities" Feb 27 08:03:25 crc kubenswrapper[4612]: E0227 08:03:25.615213 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5825c2bd-486a-4502-b612-0635c30aed60" containerName="extract-utilities" Feb 27 08:03:25 crc kubenswrapper[4612]: I0227 08:03:25.615265 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="5825c2bd-486a-4502-b612-0635c30aed60" containerName="extract-utilities" Feb 27 08:03:25 crc kubenswrapper[4612]: E0227 08:03:25.615324 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5825c2bd-486a-4502-b612-0635c30aed60" containerName="extract-content" Feb 27 08:03:25 crc kubenswrapper[4612]: I0227 08:03:25.615378 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="5825c2bd-486a-4502-b612-0635c30aed60" containerName="extract-content" Feb 27 08:03:25 crc kubenswrapper[4612]: E0227 08:03:25.615433 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="869d3ce9-173d-4aa6-8d95-d559dd273810" containerName="extract-content" Feb 27 08:03:25 crc kubenswrapper[4612]: I0227 08:03:25.615483 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="869d3ce9-173d-4aa6-8d95-d559dd273810" containerName="extract-content" Feb 27 08:03:25 crc kubenswrapper[4612]: E0227 08:03:25.615534 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5825c2bd-486a-4502-b612-0635c30aed60" containerName="registry-server" Feb 27 08:03:25 crc kubenswrapper[4612]: I0227 08:03:25.615586 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="5825c2bd-486a-4502-b612-0635c30aed60" containerName="registry-server" Feb 27 08:03:25 crc kubenswrapper[4612]: I0227 08:03:25.615760 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="869d3ce9-173d-4aa6-8d95-d559dd273810" containerName="registry-server" Feb 27 08:03:25 crc kubenswrapper[4612]: I0227 08:03:25.615830 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="5825c2bd-486a-4502-b612-0635c30aed60" containerName="registry-server" Feb 27 08:03:25 crc kubenswrapper[4612]: I0227 08:03:25.616576 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4nq7w2" Feb 27 08:03:25 crc kubenswrapper[4612]: I0227 08:03:25.618285 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Feb 27 08:03:25 crc kubenswrapper[4612]: I0227 08:03:25.627379 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4nq7w2"] Feb 27 08:03:25 crc kubenswrapper[4612]: I0227 08:03:25.703980 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6xds6\" (UniqueName: \"kubernetes.io/projected/f3841018-8eb3-4d1d-b78a-c023bbb050cd-kube-api-access-6xds6\") pod \"d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4nq7w2\" (UID: \"f3841018-8eb3-4d1d-b78a-c023bbb050cd\") " pod="openshift-marketplace/d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4nq7w2" Feb 27 08:03:25 crc kubenswrapper[4612]: I0227 08:03:25.704053 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/f3841018-8eb3-4d1d-b78a-c023bbb050cd-util\") pod \"d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4nq7w2\" (UID: \"f3841018-8eb3-4d1d-b78a-c023bbb050cd\") " pod="openshift-marketplace/d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4nq7w2" Feb 27 08:03:25 crc kubenswrapper[4612]: I0227 08:03:25.704120 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/f3841018-8eb3-4d1d-b78a-c023bbb050cd-bundle\") pod \"d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4nq7w2\" (UID: \"f3841018-8eb3-4d1d-b78a-c023bbb050cd\") " pod="openshift-marketplace/d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4nq7w2" Feb 27 08:03:25 crc kubenswrapper[4612]: I0227 08:03:25.805762 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/f3841018-8eb3-4d1d-b78a-c023bbb050cd-util\") pod \"d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4nq7w2\" (UID: \"f3841018-8eb3-4d1d-b78a-c023bbb050cd\") " pod="openshift-marketplace/d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4nq7w2" Feb 27 08:03:25 crc kubenswrapper[4612]: I0227 08:03:25.805861 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/f3841018-8eb3-4d1d-b78a-c023bbb050cd-bundle\") pod \"d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4nq7w2\" (UID: \"f3841018-8eb3-4d1d-b78a-c023bbb050cd\") " pod="openshift-marketplace/d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4nq7w2" Feb 27 08:03:25 crc kubenswrapper[4612]: I0227 08:03:25.805915 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6xds6\" (UniqueName: \"kubernetes.io/projected/f3841018-8eb3-4d1d-b78a-c023bbb050cd-kube-api-access-6xds6\") pod \"d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4nq7w2\" (UID: \"f3841018-8eb3-4d1d-b78a-c023bbb050cd\") " pod="openshift-marketplace/d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4nq7w2" Feb 27 08:03:25 crc kubenswrapper[4612]: I0227 08:03:25.806505 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/f3841018-8eb3-4d1d-b78a-c023bbb050cd-util\") pod \"d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4nq7w2\" (UID: \"f3841018-8eb3-4d1d-b78a-c023bbb050cd\") " pod="openshift-marketplace/d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4nq7w2" Feb 27 08:03:25 crc kubenswrapper[4612]: I0227 08:03:25.806548 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/f3841018-8eb3-4d1d-b78a-c023bbb050cd-bundle\") pod \"d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4nq7w2\" (UID: \"f3841018-8eb3-4d1d-b78a-c023bbb050cd\") " pod="openshift-marketplace/d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4nq7w2" Feb 27 08:03:25 crc kubenswrapper[4612]: I0227 08:03:25.826896 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6xds6\" (UniqueName: \"kubernetes.io/projected/f3841018-8eb3-4d1d-b78a-c023bbb050cd-kube-api-access-6xds6\") pod \"d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4nq7w2\" (UID: \"f3841018-8eb3-4d1d-b78a-c023bbb050cd\") " pod="openshift-marketplace/d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4nq7w2" Feb 27 08:03:25 crc kubenswrapper[4612]: I0227 08:03:25.931539 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4nq7w2" Feb 27 08:03:26 crc kubenswrapper[4612]: I0227 08:03:26.389354 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4nq7w2"] Feb 27 08:03:27 crc kubenswrapper[4612]: I0227 08:03:27.225253 4612 generic.go:334] "Generic (PLEG): container finished" podID="f3841018-8eb3-4d1d-b78a-c023bbb050cd" containerID="9ffdd445dc045e5193c85749bfaefc3f5f656ac3f11c68d31d7cd41c9e33e02a" exitCode=0 Feb 27 08:03:27 crc kubenswrapper[4612]: I0227 08:03:27.225325 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4nq7w2" event={"ID":"f3841018-8eb3-4d1d-b78a-c023bbb050cd","Type":"ContainerDied","Data":"9ffdd445dc045e5193c85749bfaefc3f5f656ac3f11c68d31d7cd41c9e33e02a"} Feb 27 08:03:27 crc kubenswrapper[4612]: I0227 08:03:27.225370 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4nq7w2" event={"ID":"f3841018-8eb3-4d1d-b78a-c023bbb050cd","Type":"ContainerStarted","Data":"a8e79b1729f62fcb637f679770866729ad6cefe986abcd622759aab9b9ed038b"} Feb 27 08:03:28 crc kubenswrapper[4612]: I0227 08:03:28.153484 4612 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-8wk54" podUID="da284312-b98a-48d6-8cff-f8071c334e0b" containerName="console" containerID="cri-o://f7fc77f7a6fb23f2d47be4c83b3939f6ce2f3662d142d7a4bd154ce263f0f715" gracePeriod=15 Feb 27 08:03:28 crc kubenswrapper[4612]: I0227 08:03:28.574723 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-8wk54_da284312-b98a-48d6-8cff-f8071c334e0b/console/0.log" Feb 27 08:03:28 crc kubenswrapper[4612]: I0227 08:03:28.574806 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-8wk54" Feb 27 08:03:28 crc kubenswrapper[4612]: I0227 08:03:28.747325 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/da284312-b98a-48d6-8cff-f8071c334e0b-console-config\") pod \"da284312-b98a-48d6-8cff-f8071c334e0b\" (UID: \"da284312-b98a-48d6-8cff-f8071c334e0b\") " Feb 27 08:03:28 crc kubenswrapper[4612]: I0227 08:03:28.747955 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/da284312-b98a-48d6-8cff-f8071c334e0b-trusted-ca-bundle\") pod \"da284312-b98a-48d6-8cff-f8071c334e0b\" (UID: \"da284312-b98a-48d6-8cff-f8071c334e0b\") " Feb 27 08:03:28 crc kubenswrapper[4612]: I0227 08:03:28.748001 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c876r\" (UniqueName: \"kubernetes.io/projected/da284312-b98a-48d6-8cff-f8071c334e0b-kube-api-access-c876r\") pod \"da284312-b98a-48d6-8cff-f8071c334e0b\" (UID: \"da284312-b98a-48d6-8cff-f8071c334e0b\") " Feb 27 08:03:28 crc kubenswrapper[4612]: I0227 08:03:28.748109 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/da284312-b98a-48d6-8cff-f8071c334e0b-oauth-serving-cert\") pod \"da284312-b98a-48d6-8cff-f8071c334e0b\" (UID: \"da284312-b98a-48d6-8cff-f8071c334e0b\") " Feb 27 08:03:28 crc kubenswrapper[4612]: I0227 08:03:28.748205 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/da284312-b98a-48d6-8cff-f8071c334e0b-console-serving-cert\") pod \"da284312-b98a-48d6-8cff-f8071c334e0b\" (UID: \"da284312-b98a-48d6-8cff-f8071c334e0b\") " Feb 27 08:03:28 crc kubenswrapper[4612]: I0227 08:03:28.749154 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/da284312-b98a-48d6-8cff-f8071c334e0b-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "da284312-b98a-48d6-8cff-f8071c334e0b" (UID: "da284312-b98a-48d6-8cff-f8071c334e0b"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 08:03:28 crc kubenswrapper[4612]: I0227 08:03:28.749240 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/da284312-b98a-48d6-8cff-f8071c334e0b-service-ca\") pod \"da284312-b98a-48d6-8cff-f8071c334e0b\" (UID: \"da284312-b98a-48d6-8cff-f8071c334e0b\") " Feb 27 08:03:28 crc kubenswrapper[4612]: I0227 08:03:28.750116 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/da284312-b98a-48d6-8cff-f8071c334e0b-service-ca" (OuterVolumeSpecName: "service-ca") pod "da284312-b98a-48d6-8cff-f8071c334e0b" (UID: "da284312-b98a-48d6-8cff-f8071c334e0b"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 08:03:28 crc kubenswrapper[4612]: I0227 08:03:28.749273 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/da284312-b98a-48d6-8cff-f8071c334e0b-console-oauth-config\") pod \"da284312-b98a-48d6-8cff-f8071c334e0b\" (UID: \"da284312-b98a-48d6-8cff-f8071c334e0b\") " Feb 27 08:03:28 crc kubenswrapper[4612]: I0227 08:03:28.750504 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/da284312-b98a-48d6-8cff-f8071c334e0b-console-config" (OuterVolumeSpecName: "console-config") pod "da284312-b98a-48d6-8cff-f8071c334e0b" (UID: "da284312-b98a-48d6-8cff-f8071c334e0b"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 08:03:28 crc kubenswrapper[4612]: I0227 08:03:28.750634 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/da284312-b98a-48d6-8cff-f8071c334e0b-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "da284312-b98a-48d6-8cff-f8071c334e0b" (UID: "da284312-b98a-48d6-8cff-f8071c334e0b"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 08:03:28 crc kubenswrapper[4612]: I0227 08:03:28.751336 4612 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/da284312-b98a-48d6-8cff-f8071c334e0b-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 27 08:03:28 crc kubenswrapper[4612]: I0227 08:03:28.751371 4612 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/da284312-b98a-48d6-8cff-f8071c334e0b-service-ca\") on node \"crc\" DevicePath \"\"" Feb 27 08:03:28 crc kubenswrapper[4612]: I0227 08:03:28.751384 4612 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/da284312-b98a-48d6-8cff-f8071c334e0b-console-config\") on node \"crc\" DevicePath \"\"" Feb 27 08:03:28 crc kubenswrapper[4612]: I0227 08:03:28.751397 4612 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/da284312-b98a-48d6-8cff-f8071c334e0b-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 08:03:28 crc kubenswrapper[4612]: I0227 08:03:28.756262 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/da284312-b98a-48d6-8cff-f8071c334e0b-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "da284312-b98a-48d6-8cff-f8071c334e0b" (UID: "da284312-b98a-48d6-8cff-f8071c334e0b"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:03:28 crc kubenswrapper[4612]: I0227 08:03:28.757106 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/da284312-b98a-48d6-8cff-f8071c334e0b-kube-api-access-c876r" (OuterVolumeSpecName: "kube-api-access-c876r") pod "da284312-b98a-48d6-8cff-f8071c334e0b" (UID: "da284312-b98a-48d6-8cff-f8071c334e0b"). InnerVolumeSpecName "kube-api-access-c876r". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 08:03:28 crc kubenswrapper[4612]: I0227 08:03:28.761947 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/da284312-b98a-48d6-8cff-f8071c334e0b-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "da284312-b98a-48d6-8cff-f8071c334e0b" (UID: "da284312-b98a-48d6-8cff-f8071c334e0b"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:03:28 crc kubenswrapper[4612]: I0227 08:03:28.852566 4612 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/da284312-b98a-48d6-8cff-f8071c334e0b-console-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 27 08:03:28 crc kubenswrapper[4612]: I0227 08:03:28.852617 4612 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/da284312-b98a-48d6-8cff-f8071c334e0b-console-oauth-config\") on node \"crc\" DevicePath \"\"" Feb 27 08:03:28 crc kubenswrapper[4612]: I0227 08:03:28.852638 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c876r\" (UniqueName: \"kubernetes.io/projected/da284312-b98a-48d6-8cff-f8071c334e0b-kube-api-access-c876r\") on node \"crc\" DevicePath \"\"" Feb 27 08:03:29 crc kubenswrapper[4612]: I0227 08:03:29.241347 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-8wk54_da284312-b98a-48d6-8cff-f8071c334e0b/console/0.log" Feb 27 08:03:29 crc kubenswrapper[4612]: I0227 08:03:29.241440 4612 generic.go:334] "Generic (PLEG): container finished" podID="da284312-b98a-48d6-8cff-f8071c334e0b" containerID="f7fc77f7a6fb23f2d47be4c83b3939f6ce2f3662d142d7a4bd154ce263f0f715" exitCode=2 Feb 27 08:03:29 crc kubenswrapper[4612]: I0227 08:03:29.241569 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-8wk54" Feb 27 08:03:29 crc kubenswrapper[4612]: I0227 08:03:29.241577 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-8wk54" event={"ID":"da284312-b98a-48d6-8cff-f8071c334e0b","Type":"ContainerDied","Data":"f7fc77f7a6fb23f2d47be4c83b3939f6ce2f3662d142d7a4bd154ce263f0f715"} Feb 27 08:03:29 crc kubenswrapper[4612]: I0227 08:03:29.241765 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-8wk54" event={"ID":"da284312-b98a-48d6-8cff-f8071c334e0b","Type":"ContainerDied","Data":"535ccf5b8854efb31e4b3373f23422349a74e17fc98ec63a67a191a8691cb543"} Feb 27 08:03:29 crc kubenswrapper[4612]: I0227 08:03:29.241815 4612 scope.go:117] "RemoveContainer" containerID="f7fc77f7a6fb23f2d47be4c83b3939f6ce2f3662d142d7a4bd154ce263f0f715" Feb 27 08:03:29 crc kubenswrapper[4612]: I0227 08:03:29.244798 4612 generic.go:334] "Generic (PLEG): container finished" podID="f3841018-8eb3-4d1d-b78a-c023bbb050cd" containerID="698e519c42fb188379198b2e985ba528ce39ad5c5d2b4c03dd52b8d7c7502e97" exitCode=0 Feb 27 08:03:29 crc kubenswrapper[4612]: I0227 08:03:29.244958 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4nq7w2" event={"ID":"f3841018-8eb3-4d1d-b78a-c023bbb050cd","Type":"ContainerDied","Data":"698e519c42fb188379198b2e985ba528ce39ad5c5d2b4c03dd52b8d7c7502e97"} Feb 27 08:03:29 crc kubenswrapper[4612]: I0227 08:03:29.305730 4612 scope.go:117] "RemoveContainer" containerID="f7fc77f7a6fb23f2d47be4c83b3939f6ce2f3662d142d7a4bd154ce263f0f715" Feb 27 08:03:29 crc kubenswrapper[4612]: E0227 08:03:29.306512 4612 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f7fc77f7a6fb23f2d47be4c83b3939f6ce2f3662d142d7a4bd154ce263f0f715\": container with ID starting with f7fc77f7a6fb23f2d47be4c83b3939f6ce2f3662d142d7a4bd154ce263f0f715 not found: ID does not exist" containerID="f7fc77f7a6fb23f2d47be4c83b3939f6ce2f3662d142d7a4bd154ce263f0f715" Feb 27 08:03:29 crc kubenswrapper[4612]: I0227 08:03:29.306581 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f7fc77f7a6fb23f2d47be4c83b3939f6ce2f3662d142d7a4bd154ce263f0f715"} err="failed to get container status \"f7fc77f7a6fb23f2d47be4c83b3939f6ce2f3662d142d7a4bd154ce263f0f715\": rpc error: code = NotFound desc = could not find container \"f7fc77f7a6fb23f2d47be4c83b3939f6ce2f3662d142d7a4bd154ce263f0f715\": container with ID starting with f7fc77f7a6fb23f2d47be4c83b3939f6ce2f3662d142d7a4bd154ce263f0f715 not found: ID does not exist" Feb 27 08:03:29 crc kubenswrapper[4612]: I0227 08:03:29.319890 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-8wk54"] Feb 27 08:03:29 crc kubenswrapper[4612]: I0227 08:03:29.327361 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-8wk54"] Feb 27 08:03:30 crc kubenswrapper[4612]: I0227 08:03:30.254485 4612 generic.go:334] "Generic (PLEG): container finished" podID="f3841018-8eb3-4d1d-b78a-c023bbb050cd" containerID="66bc5562deb3384d72a1a6174e054e42ae0f6f7942e12de3c188a6916423eabf" exitCode=0 Feb 27 08:03:30 crc kubenswrapper[4612]: I0227 08:03:30.254534 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4nq7w2" event={"ID":"f3841018-8eb3-4d1d-b78a-c023bbb050cd","Type":"ContainerDied","Data":"66bc5562deb3384d72a1a6174e054e42ae0f6f7942e12de3c188a6916423eabf"} Feb 27 08:03:30 crc kubenswrapper[4612]: I0227 08:03:30.867340 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="da284312-b98a-48d6-8cff-f8071c334e0b" path="/var/lib/kubelet/pods/da284312-b98a-48d6-8cff-f8071c334e0b/volumes" Feb 27 08:03:31 crc kubenswrapper[4612]: I0227 08:03:31.558466 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4nq7w2" Feb 27 08:03:31 crc kubenswrapper[4612]: I0227 08:03:31.691793 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/f3841018-8eb3-4d1d-b78a-c023bbb050cd-util\") pod \"f3841018-8eb3-4d1d-b78a-c023bbb050cd\" (UID: \"f3841018-8eb3-4d1d-b78a-c023bbb050cd\") " Feb 27 08:03:31 crc kubenswrapper[4612]: I0227 08:03:31.691861 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6xds6\" (UniqueName: \"kubernetes.io/projected/f3841018-8eb3-4d1d-b78a-c023bbb050cd-kube-api-access-6xds6\") pod \"f3841018-8eb3-4d1d-b78a-c023bbb050cd\" (UID: \"f3841018-8eb3-4d1d-b78a-c023bbb050cd\") " Feb 27 08:03:31 crc kubenswrapper[4612]: I0227 08:03:31.691952 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/f3841018-8eb3-4d1d-b78a-c023bbb050cd-bundle\") pod \"f3841018-8eb3-4d1d-b78a-c023bbb050cd\" (UID: \"f3841018-8eb3-4d1d-b78a-c023bbb050cd\") " Feb 27 08:03:31 crc kubenswrapper[4612]: I0227 08:03:31.694171 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f3841018-8eb3-4d1d-b78a-c023bbb050cd-bundle" (OuterVolumeSpecName: "bundle") pod "f3841018-8eb3-4d1d-b78a-c023bbb050cd" (UID: "f3841018-8eb3-4d1d-b78a-c023bbb050cd"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 08:03:31 crc kubenswrapper[4612]: I0227 08:03:31.697516 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f3841018-8eb3-4d1d-b78a-c023bbb050cd-kube-api-access-6xds6" (OuterVolumeSpecName: "kube-api-access-6xds6") pod "f3841018-8eb3-4d1d-b78a-c023bbb050cd" (UID: "f3841018-8eb3-4d1d-b78a-c023bbb050cd"). InnerVolumeSpecName "kube-api-access-6xds6". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 08:03:31 crc kubenswrapper[4612]: I0227 08:03:31.708932 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f3841018-8eb3-4d1d-b78a-c023bbb050cd-util" (OuterVolumeSpecName: "util") pod "f3841018-8eb3-4d1d-b78a-c023bbb050cd" (UID: "f3841018-8eb3-4d1d-b78a-c023bbb050cd"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 08:03:31 crc kubenswrapper[4612]: I0227 08:03:31.793547 4612 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/f3841018-8eb3-4d1d-b78a-c023bbb050cd-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 08:03:31 crc kubenswrapper[4612]: I0227 08:03:31.793609 4612 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/f3841018-8eb3-4d1d-b78a-c023bbb050cd-util\") on node \"crc\" DevicePath \"\"" Feb 27 08:03:31 crc kubenswrapper[4612]: I0227 08:03:31.793619 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6xds6\" (UniqueName: \"kubernetes.io/projected/f3841018-8eb3-4d1d-b78a-c023bbb050cd-kube-api-access-6xds6\") on node \"crc\" DevicePath \"\"" Feb 27 08:03:32 crc kubenswrapper[4612]: I0227 08:03:32.279412 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4nq7w2" event={"ID":"f3841018-8eb3-4d1d-b78a-c023bbb050cd","Type":"ContainerDied","Data":"a8e79b1729f62fcb637f679770866729ad6cefe986abcd622759aab9b9ed038b"} Feb 27 08:03:32 crc kubenswrapper[4612]: I0227 08:03:32.279788 4612 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a8e79b1729f62fcb637f679770866729ad6cefe986abcd622759aab9b9ed038b" Feb 27 08:03:32 crc kubenswrapper[4612]: I0227 08:03:32.279488 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4nq7w2" Feb 27 08:03:35 crc kubenswrapper[4612]: I0227 08:03:35.392679 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-s5f86"] Feb 27 08:03:35 crc kubenswrapper[4612]: E0227 08:03:35.393121 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f3841018-8eb3-4d1d-b78a-c023bbb050cd" containerName="util" Feb 27 08:03:35 crc kubenswrapper[4612]: I0227 08:03:35.393132 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="f3841018-8eb3-4d1d-b78a-c023bbb050cd" containerName="util" Feb 27 08:03:35 crc kubenswrapper[4612]: E0227 08:03:35.393150 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f3841018-8eb3-4d1d-b78a-c023bbb050cd" containerName="pull" Feb 27 08:03:35 crc kubenswrapper[4612]: I0227 08:03:35.393155 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="f3841018-8eb3-4d1d-b78a-c023bbb050cd" containerName="pull" Feb 27 08:03:35 crc kubenswrapper[4612]: E0227 08:03:35.393168 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="da284312-b98a-48d6-8cff-f8071c334e0b" containerName="console" Feb 27 08:03:35 crc kubenswrapper[4612]: I0227 08:03:35.393174 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="da284312-b98a-48d6-8cff-f8071c334e0b" containerName="console" Feb 27 08:03:35 crc kubenswrapper[4612]: E0227 08:03:35.393180 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f3841018-8eb3-4d1d-b78a-c023bbb050cd" containerName="extract" Feb 27 08:03:35 crc kubenswrapper[4612]: I0227 08:03:35.393186 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="f3841018-8eb3-4d1d-b78a-c023bbb050cd" containerName="extract" Feb 27 08:03:35 crc kubenswrapper[4612]: I0227 08:03:35.393304 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="f3841018-8eb3-4d1d-b78a-c023bbb050cd" containerName="extract" Feb 27 08:03:35 crc kubenswrapper[4612]: I0227 08:03:35.393322 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="da284312-b98a-48d6-8cff-f8071c334e0b" containerName="console" Feb 27 08:03:35 crc kubenswrapper[4612]: I0227 08:03:35.394168 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-s5f86" Feb 27 08:03:35 crc kubenswrapper[4612]: I0227 08:03:35.402340 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-s5f86"] Feb 27 08:03:35 crc kubenswrapper[4612]: I0227 08:03:35.542813 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r56bl\" (UniqueName: \"kubernetes.io/projected/5d964ece-5965-43d6-a48d-064f989f03d8-kube-api-access-r56bl\") pod \"certified-operators-s5f86\" (UID: \"5d964ece-5965-43d6-a48d-064f989f03d8\") " pod="openshift-marketplace/certified-operators-s5f86" Feb 27 08:03:35 crc kubenswrapper[4612]: I0227 08:03:35.542870 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5d964ece-5965-43d6-a48d-064f989f03d8-utilities\") pod \"certified-operators-s5f86\" (UID: \"5d964ece-5965-43d6-a48d-064f989f03d8\") " pod="openshift-marketplace/certified-operators-s5f86" Feb 27 08:03:35 crc kubenswrapper[4612]: I0227 08:03:35.542907 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5d964ece-5965-43d6-a48d-064f989f03d8-catalog-content\") pod \"certified-operators-s5f86\" (UID: \"5d964ece-5965-43d6-a48d-064f989f03d8\") " pod="openshift-marketplace/certified-operators-s5f86" Feb 27 08:03:35 crc kubenswrapper[4612]: I0227 08:03:35.645270 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r56bl\" (UniqueName: \"kubernetes.io/projected/5d964ece-5965-43d6-a48d-064f989f03d8-kube-api-access-r56bl\") pod \"certified-operators-s5f86\" (UID: \"5d964ece-5965-43d6-a48d-064f989f03d8\") " pod="openshift-marketplace/certified-operators-s5f86" Feb 27 08:03:35 crc kubenswrapper[4612]: I0227 08:03:35.645343 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5d964ece-5965-43d6-a48d-064f989f03d8-utilities\") pod \"certified-operators-s5f86\" (UID: \"5d964ece-5965-43d6-a48d-064f989f03d8\") " pod="openshift-marketplace/certified-operators-s5f86" Feb 27 08:03:35 crc kubenswrapper[4612]: I0227 08:03:35.645389 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5d964ece-5965-43d6-a48d-064f989f03d8-catalog-content\") pod \"certified-operators-s5f86\" (UID: \"5d964ece-5965-43d6-a48d-064f989f03d8\") " pod="openshift-marketplace/certified-operators-s5f86" Feb 27 08:03:35 crc kubenswrapper[4612]: I0227 08:03:35.645991 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5d964ece-5965-43d6-a48d-064f989f03d8-catalog-content\") pod \"certified-operators-s5f86\" (UID: \"5d964ece-5965-43d6-a48d-064f989f03d8\") " pod="openshift-marketplace/certified-operators-s5f86" Feb 27 08:03:35 crc kubenswrapper[4612]: I0227 08:03:35.646036 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5d964ece-5965-43d6-a48d-064f989f03d8-utilities\") pod \"certified-operators-s5f86\" (UID: \"5d964ece-5965-43d6-a48d-064f989f03d8\") " pod="openshift-marketplace/certified-operators-s5f86" Feb 27 08:03:35 crc kubenswrapper[4612]: I0227 08:03:35.682552 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r56bl\" (UniqueName: \"kubernetes.io/projected/5d964ece-5965-43d6-a48d-064f989f03d8-kube-api-access-r56bl\") pod \"certified-operators-s5f86\" (UID: \"5d964ece-5965-43d6-a48d-064f989f03d8\") " pod="openshift-marketplace/certified-operators-s5f86" Feb 27 08:03:35 crc kubenswrapper[4612]: I0227 08:03:35.712494 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-s5f86" Feb 27 08:03:36 crc kubenswrapper[4612]: I0227 08:03:36.081016 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-s5f86"] Feb 27 08:03:36 crc kubenswrapper[4612]: I0227 08:03:36.315946 4612 generic.go:334] "Generic (PLEG): container finished" podID="5d964ece-5965-43d6-a48d-064f989f03d8" containerID="dbfbf57122d28a8f255e17140c3112baaadef3f59177b1c3d90991a1fae49d29" exitCode=0 Feb 27 08:03:36 crc kubenswrapper[4612]: I0227 08:03:36.315990 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-s5f86" event={"ID":"5d964ece-5965-43d6-a48d-064f989f03d8","Type":"ContainerDied","Data":"dbfbf57122d28a8f255e17140c3112baaadef3f59177b1c3d90991a1fae49d29"} Feb 27 08:03:36 crc kubenswrapper[4612]: I0227 08:03:36.316041 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-s5f86" event={"ID":"5d964ece-5965-43d6-a48d-064f989f03d8","Type":"ContainerStarted","Data":"526bf0004bdcc11c7dc9042680a6f05051c6aed0ef6827ade12e594c36c9f45b"} Feb 27 08:03:37 crc kubenswrapper[4612]: I0227 08:03:37.322516 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-s5f86" event={"ID":"5d964ece-5965-43d6-a48d-064f989f03d8","Type":"ContainerStarted","Data":"0224e64708afb96be780e9fe1212b17a47e500dbbb7c7995c029cf9af7ce7d4b"} Feb 27 08:03:38 crc kubenswrapper[4612]: I0227 08:03:38.329515 4612 generic.go:334] "Generic (PLEG): container finished" podID="5d964ece-5965-43d6-a48d-064f989f03d8" containerID="0224e64708afb96be780e9fe1212b17a47e500dbbb7c7995c029cf9af7ce7d4b" exitCode=0 Feb 27 08:03:38 crc kubenswrapper[4612]: I0227 08:03:38.329593 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-s5f86" event={"ID":"5d964ece-5965-43d6-a48d-064f989f03d8","Type":"ContainerDied","Data":"0224e64708afb96be780e9fe1212b17a47e500dbbb7c7995c029cf9af7ce7d4b"} Feb 27 08:03:39 crc kubenswrapper[4612]: I0227 08:03:39.337090 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-s5f86" event={"ID":"5d964ece-5965-43d6-a48d-064f989f03d8","Type":"ContainerStarted","Data":"d5d77c07bca43580e509009658b455e44b1e69e64894e88afe40c9e1795648e0"} Feb 27 08:03:39 crc kubenswrapper[4612]: I0227 08:03:39.362650 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-s5f86" podStartSLOduration=1.924719907 podStartE2EDuration="4.362624958s" podCreationTimestamp="2026-02-27 08:03:35 +0000 UTC" firstStartedPulling="2026-02-27 08:03:36.317554672 +0000 UTC m=+874.171484670" lastFinishedPulling="2026-02-27 08:03:38.755459713 +0000 UTC m=+876.609389721" observedRunningTime="2026-02-27 08:03:39.356040829 +0000 UTC m=+877.209970847" watchObservedRunningTime="2026-02-27 08:03:39.362624958 +0000 UTC m=+877.216554966" Feb 27 08:03:43 crc kubenswrapper[4612]: I0227 08:03:43.762485 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-78d5b849b-c2rb5"] Feb 27 08:03:43 crc kubenswrapper[4612]: I0227 08:03:43.763931 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-78d5b849b-c2rb5" Feb 27 08:03:43 crc kubenswrapper[4612]: I0227 08:03:43.766327 4612 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Feb 27 08:03:43 crc kubenswrapper[4612]: I0227 08:03:43.766368 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Feb 27 08:03:43 crc kubenswrapper[4612]: I0227 08:03:43.766462 4612 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Feb 27 08:03:43 crc kubenswrapper[4612]: I0227 08:03:43.766731 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Feb 27 08:03:43 crc kubenswrapper[4612]: I0227 08:03:43.768462 4612 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-fpr74" Feb 27 08:03:43 crc kubenswrapper[4612]: I0227 08:03:43.786972 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-78d5b849b-c2rb5"] Feb 27 08:03:43 crc kubenswrapper[4612]: I0227 08:03:43.859633 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qvblr\" (UniqueName: \"kubernetes.io/projected/29cfdfc3-d4a9-44df-b46e-638a9707a9f1-kube-api-access-qvblr\") pod \"metallb-operator-controller-manager-78d5b849b-c2rb5\" (UID: \"29cfdfc3-d4a9-44df-b46e-638a9707a9f1\") " pod="metallb-system/metallb-operator-controller-manager-78d5b849b-c2rb5" Feb 27 08:03:43 crc kubenswrapper[4612]: I0227 08:03:43.859715 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/29cfdfc3-d4a9-44df-b46e-638a9707a9f1-webhook-cert\") pod \"metallb-operator-controller-manager-78d5b849b-c2rb5\" (UID: \"29cfdfc3-d4a9-44df-b46e-638a9707a9f1\") " pod="metallb-system/metallb-operator-controller-manager-78d5b849b-c2rb5" Feb 27 08:03:43 crc kubenswrapper[4612]: I0227 08:03:43.859742 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/29cfdfc3-d4a9-44df-b46e-638a9707a9f1-apiservice-cert\") pod \"metallb-operator-controller-manager-78d5b849b-c2rb5\" (UID: \"29cfdfc3-d4a9-44df-b46e-638a9707a9f1\") " pod="metallb-system/metallb-operator-controller-manager-78d5b849b-c2rb5" Feb 27 08:03:43 crc kubenswrapper[4612]: I0227 08:03:43.960276 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qvblr\" (UniqueName: \"kubernetes.io/projected/29cfdfc3-d4a9-44df-b46e-638a9707a9f1-kube-api-access-qvblr\") pod \"metallb-operator-controller-manager-78d5b849b-c2rb5\" (UID: \"29cfdfc3-d4a9-44df-b46e-638a9707a9f1\") " pod="metallb-system/metallb-operator-controller-manager-78d5b849b-c2rb5" Feb 27 08:03:43 crc kubenswrapper[4612]: I0227 08:03:43.960346 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/29cfdfc3-d4a9-44df-b46e-638a9707a9f1-webhook-cert\") pod \"metallb-operator-controller-manager-78d5b849b-c2rb5\" (UID: \"29cfdfc3-d4a9-44df-b46e-638a9707a9f1\") " pod="metallb-system/metallb-operator-controller-manager-78d5b849b-c2rb5" Feb 27 08:03:43 crc kubenswrapper[4612]: I0227 08:03:43.960365 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/29cfdfc3-d4a9-44df-b46e-638a9707a9f1-apiservice-cert\") pod \"metallb-operator-controller-manager-78d5b849b-c2rb5\" (UID: \"29cfdfc3-d4a9-44df-b46e-638a9707a9f1\") " pod="metallb-system/metallb-operator-controller-manager-78d5b849b-c2rb5" Feb 27 08:03:43 crc kubenswrapper[4612]: I0227 08:03:43.969649 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/29cfdfc3-d4a9-44df-b46e-638a9707a9f1-apiservice-cert\") pod \"metallb-operator-controller-manager-78d5b849b-c2rb5\" (UID: \"29cfdfc3-d4a9-44df-b46e-638a9707a9f1\") " pod="metallb-system/metallb-operator-controller-manager-78d5b849b-c2rb5" Feb 27 08:03:43 crc kubenswrapper[4612]: I0227 08:03:43.969664 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/29cfdfc3-d4a9-44df-b46e-638a9707a9f1-webhook-cert\") pod \"metallb-operator-controller-manager-78d5b849b-c2rb5\" (UID: \"29cfdfc3-d4a9-44df-b46e-638a9707a9f1\") " pod="metallb-system/metallb-operator-controller-manager-78d5b849b-c2rb5" Feb 27 08:03:44 crc kubenswrapper[4612]: I0227 08:03:44.002461 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qvblr\" (UniqueName: \"kubernetes.io/projected/29cfdfc3-d4a9-44df-b46e-638a9707a9f1-kube-api-access-qvblr\") pod \"metallb-operator-controller-manager-78d5b849b-c2rb5\" (UID: \"29cfdfc3-d4a9-44df-b46e-638a9707a9f1\") " pod="metallb-system/metallb-operator-controller-manager-78d5b849b-c2rb5" Feb 27 08:03:44 crc kubenswrapper[4612]: I0227 08:03:44.080079 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-78d5b849b-c2rb5" Feb 27 08:03:44 crc kubenswrapper[4612]: I0227 08:03:44.225922 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-6f9ff8875f-jphph"] Feb 27 08:03:44 crc kubenswrapper[4612]: I0227 08:03:44.226815 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-6f9ff8875f-jphph" Feb 27 08:03:44 crc kubenswrapper[4612]: I0227 08:03:44.234412 4612 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Feb 27 08:03:44 crc kubenswrapper[4612]: I0227 08:03:44.235278 4612 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Feb 27 08:03:44 crc kubenswrapper[4612]: I0227 08:03:44.235922 4612 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-tkvb4" Feb 27 08:03:44 crc kubenswrapper[4612]: I0227 08:03:44.254305 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-6f9ff8875f-jphph"] Feb 27 08:03:44 crc kubenswrapper[4612]: I0227 08:03:44.365948 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/62df4a96-aac6-4169-bbc4-d41ea85b8722-webhook-cert\") pod \"metallb-operator-webhook-server-6f9ff8875f-jphph\" (UID: \"62df4a96-aac6-4169-bbc4-d41ea85b8722\") " pod="metallb-system/metallb-operator-webhook-server-6f9ff8875f-jphph" Feb 27 08:03:44 crc kubenswrapper[4612]: I0227 08:03:44.366254 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mk8dz\" (UniqueName: \"kubernetes.io/projected/62df4a96-aac6-4169-bbc4-d41ea85b8722-kube-api-access-mk8dz\") pod \"metallb-operator-webhook-server-6f9ff8875f-jphph\" (UID: \"62df4a96-aac6-4169-bbc4-d41ea85b8722\") " pod="metallb-system/metallb-operator-webhook-server-6f9ff8875f-jphph" Feb 27 08:03:44 crc kubenswrapper[4612]: I0227 08:03:44.366298 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/62df4a96-aac6-4169-bbc4-d41ea85b8722-apiservice-cert\") pod \"metallb-operator-webhook-server-6f9ff8875f-jphph\" (UID: \"62df4a96-aac6-4169-bbc4-d41ea85b8722\") " pod="metallb-system/metallb-operator-webhook-server-6f9ff8875f-jphph" Feb 27 08:03:44 crc kubenswrapper[4612]: I0227 08:03:44.466928 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mk8dz\" (UniqueName: \"kubernetes.io/projected/62df4a96-aac6-4169-bbc4-d41ea85b8722-kube-api-access-mk8dz\") pod \"metallb-operator-webhook-server-6f9ff8875f-jphph\" (UID: \"62df4a96-aac6-4169-bbc4-d41ea85b8722\") " pod="metallb-system/metallb-operator-webhook-server-6f9ff8875f-jphph" Feb 27 08:03:44 crc kubenswrapper[4612]: I0227 08:03:44.467001 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/62df4a96-aac6-4169-bbc4-d41ea85b8722-apiservice-cert\") pod \"metallb-operator-webhook-server-6f9ff8875f-jphph\" (UID: \"62df4a96-aac6-4169-bbc4-d41ea85b8722\") " pod="metallb-system/metallb-operator-webhook-server-6f9ff8875f-jphph" Feb 27 08:03:44 crc kubenswrapper[4612]: I0227 08:03:44.467039 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/62df4a96-aac6-4169-bbc4-d41ea85b8722-webhook-cert\") pod \"metallb-operator-webhook-server-6f9ff8875f-jphph\" (UID: \"62df4a96-aac6-4169-bbc4-d41ea85b8722\") " pod="metallb-system/metallb-operator-webhook-server-6f9ff8875f-jphph" Feb 27 08:03:44 crc kubenswrapper[4612]: I0227 08:03:44.471098 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/62df4a96-aac6-4169-bbc4-d41ea85b8722-apiservice-cert\") pod \"metallb-operator-webhook-server-6f9ff8875f-jphph\" (UID: \"62df4a96-aac6-4169-bbc4-d41ea85b8722\") " pod="metallb-system/metallb-operator-webhook-server-6f9ff8875f-jphph" Feb 27 08:03:44 crc kubenswrapper[4612]: I0227 08:03:44.472234 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/62df4a96-aac6-4169-bbc4-d41ea85b8722-webhook-cert\") pod \"metallb-operator-webhook-server-6f9ff8875f-jphph\" (UID: \"62df4a96-aac6-4169-bbc4-d41ea85b8722\") " pod="metallb-system/metallb-operator-webhook-server-6f9ff8875f-jphph" Feb 27 08:03:44 crc kubenswrapper[4612]: I0227 08:03:44.488831 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mk8dz\" (UniqueName: \"kubernetes.io/projected/62df4a96-aac6-4169-bbc4-d41ea85b8722-kube-api-access-mk8dz\") pod \"metallb-operator-webhook-server-6f9ff8875f-jphph\" (UID: \"62df4a96-aac6-4169-bbc4-d41ea85b8722\") " pod="metallb-system/metallb-operator-webhook-server-6f9ff8875f-jphph" Feb 27 08:03:44 crc kubenswrapper[4612]: I0227 08:03:44.501273 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-78d5b849b-c2rb5"] Feb 27 08:03:44 crc kubenswrapper[4612]: W0227 08:03:44.505923 4612 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod29cfdfc3_d4a9_44df_b46e_638a9707a9f1.slice/crio-6ac6a1f88a2b053260b05641f3939a1c0c48b629095dcfb492d04512f090bfff WatchSource:0}: Error finding container 6ac6a1f88a2b053260b05641f3939a1c0c48b629095dcfb492d04512f090bfff: Status 404 returned error can't find the container with id 6ac6a1f88a2b053260b05641f3939a1c0c48b629095dcfb492d04512f090bfff Feb 27 08:03:44 crc kubenswrapper[4612]: I0227 08:03:44.541346 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-6f9ff8875f-jphph" Feb 27 08:03:44 crc kubenswrapper[4612]: I0227 08:03:44.805220 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-6f9ff8875f-jphph"] Feb 27 08:03:45 crc kubenswrapper[4612]: I0227 08:03:45.380525 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-78d5b849b-c2rb5" event={"ID":"29cfdfc3-d4a9-44df-b46e-638a9707a9f1","Type":"ContainerStarted","Data":"6ac6a1f88a2b053260b05641f3939a1c0c48b629095dcfb492d04512f090bfff"} Feb 27 08:03:45 crc kubenswrapper[4612]: I0227 08:03:45.381807 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-6f9ff8875f-jphph" event={"ID":"62df4a96-aac6-4169-bbc4-d41ea85b8722","Type":"ContainerStarted","Data":"5079ef716ccdfed7f0eeffa97aa0b7b9b49212c3660ee81e019b9828e1db0f36"} Feb 27 08:03:45 crc kubenswrapper[4612]: I0227 08:03:45.713451 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-s5f86" Feb 27 08:03:45 crc kubenswrapper[4612]: I0227 08:03:45.714509 4612 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-s5f86" Feb 27 08:03:45 crc kubenswrapper[4612]: I0227 08:03:45.757227 4612 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-s5f86" Feb 27 08:03:46 crc kubenswrapper[4612]: I0227 08:03:46.436930 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-s5f86" Feb 27 08:03:47 crc kubenswrapper[4612]: I0227 08:03:47.586076 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-s5f86"] Feb 27 08:03:49 crc kubenswrapper[4612]: I0227 08:03:49.416673 4612 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-s5f86" podUID="5d964ece-5965-43d6-a48d-064f989f03d8" containerName="registry-server" containerID="cri-o://d5d77c07bca43580e509009658b455e44b1e69e64894e88afe40c9e1795648e0" gracePeriod=2 Feb 27 08:03:50 crc kubenswrapper[4612]: I0227 08:03:50.439937 4612 generic.go:334] "Generic (PLEG): container finished" podID="5d964ece-5965-43d6-a48d-064f989f03d8" containerID="d5d77c07bca43580e509009658b455e44b1e69e64894e88afe40c9e1795648e0" exitCode=0 Feb 27 08:03:50 crc kubenswrapper[4612]: I0227 08:03:50.439979 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-s5f86" event={"ID":"5d964ece-5965-43d6-a48d-064f989f03d8","Type":"ContainerDied","Data":"d5d77c07bca43580e509009658b455e44b1e69e64894e88afe40c9e1795648e0"} Feb 27 08:03:55 crc kubenswrapper[4612]: I0227 08:03:55.295526 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-s5f86" Feb 27 08:03:55 crc kubenswrapper[4612]: I0227 08:03:55.325762 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5d964ece-5965-43d6-a48d-064f989f03d8-utilities\") pod \"5d964ece-5965-43d6-a48d-064f989f03d8\" (UID: \"5d964ece-5965-43d6-a48d-064f989f03d8\") " Feb 27 08:03:55 crc kubenswrapper[4612]: I0227 08:03:55.325821 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5d964ece-5965-43d6-a48d-064f989f03d8-catalog-content\") pod \"5d964ece-5965-43d6-a48d-064f989f03d8\" (UID: \"5d964ece-5965-43d6-a48d-064f989f03d8\") " Feb 27 08:03:55 crc kubenswrapper[4612]: I0227 08:03:55.325880 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r56bl\" (UniqueName: \"kubernetes.io/projected/5d964ece-5965-43d6-a48d-064f989f03d8-kube-api-access-r56bl\") pod \"5d964ece-5965-43d6-a48d-064f989f03d8\" (UID: \"5d964ece-5965-43d6-a48d-064f989f03d8\") " Feb 27 08:03:55 crc kubenswrapper[4612]: I0227 08:03:55.326552 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5d964ece-5965-43d6-a48d-064f989f03d8-utilities" (OuterVolumeSpecName: "utilities") pod "5d964ece-5965-43d6-a48d-064f989f03d8" (UID: "5d964ece-5965-43d6-a48d-064f989f03d8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 08:03:55 crc kubenswrapper[4612]: I0227 08:03:55.336925 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5d964ece-5965-43d6-a48d-064f989f03d8-kube-api-access-r56bl" (OuterVolumeSpecName: "kube-api-access-r56bl") pod "5d964ece-5965-43d6-a48d-064f989f03d8" (UID: "5d964ece-5965-43d6-a48d-064f989f03d8"). InnerVolumeSpecName "kube-api-access-r56bl". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 08:03:55 crc kubenswrapper[4612]: I0227 08:03:55.390673 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5d964ece-5965-43d6-a48d-064f989f03d8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5d964ece-5965-43d6-a48d-064f989f03d8" (UID: "5d964ece-5965-43d6-a48d-064f989f03d8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 08:03:55 crc kubenswrapper[4612]: I0227 08:03:55.453089 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r56bl\" (UniqueName: \"kubernetes.io/projected/5d964ece-5965-43d6-a48d-064f989f03d8-kube-api-access-r56bl\") on node \"crc\" DevicePath \"\"" Feb 27 08:03:55 crc kubenswrapper[4612]: I0227 08:03:55.453120 4612 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5d964ece-5965-43d6-a48d-064f989f03d8-utilities\") on node \"crc\" DevicePath \"\"" Feb 27 08:03:55 crc kubenswrapper[4612]: I0227 08:03:55.453131 4612 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5d964ece-5965-43d6-a48d-064f989f03d8-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 27 08:03:55 crc kubenswrapper[4612]: I0227 08:03:55.478426 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-s5f86" event={"ID":"5d964ece-5965-43d6-a48d-064f989f03d8","Type":"ContainerDied","Data":"526bf0004bdcc11c7dc9042680a6f05051c6aed0ef6827ade12e594c36c9f45b"} Feb 27 08:03:55 crc kubenswrapper[4612]: I0227 08:03:55.478495 4612 scope.go:117] "RemoveContainer" containerID="d5d77c07bca43580e509009658b455e44b1e69e64894e88afe40c9e1795648e0" Feb 27 08:03:55 crc kubenswrapper[4612]: I0227 08:03:55.478629 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-s5f86" Feb 27 08:03:55 crc kubenswrapper[4612]: I0227 08:03:55.488252 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-78d5b849b-c2rb5" event={"ID":"29cfdfc3-d4a9-44df-b46e-638a9707a9f1","Type":"ContainerStarted","Data":"d5e69d0da6e7a523698b4c136d78a98a8cea08a3177f9a18572ba044e82f3ef0"} Feb 27 08:03:55 crc kubenswrapper[4612]: I0227 08:03:55.489336 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-78d5b849b-c2rb5" Feb 27 08:03:55 crc kubenswrapper[4612]: I0227 08:03:55.491153 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-6f9ff8875f-jphph" event={"ID":"62df4a96-aac6-4169-bbc4-d41ea85b8722","Type":"ContainerStarted","Data":"a5fe29f244efc17691f51302f20a291bf889ceeaf5d82e5b05c418b92e438ccd"} Feb 27 08:03:55 crc kubenswrapper[4612]: I0227 08:03:55.491540 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-6f9ff8875f-jphph" Feb 27 08:03:55 crc kubenswrapper[4612]: I0227 08:03:55.497260 4612 scope.go:117] "RemoveContainer" containerID="0224e64708afb96be780e9fe1212b17a47e500dbbb7c7995c029cf9af7ce7d4b" Feb 27 08:03:55 crc kubenswrapper[4612]: I0227 08:03:55.523666 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-78d5b849b-c2rb5" podStartSLOduration=1.9502280330000001 podStartE2EDuration="12.523652881s" podCreationTimestamp="2026-02-27 08:03:43 +0000 UTC" firstStartedPulling="2026-02-27 08:03:44.508896719 +0000 UTC m=+882.362826717" lastFinishedPulling="2026-02-27 08:03:55.082321567 +0000 UTC m=+892.936251565" observedRunningTime="2026-02-27 08:03:55.518058441 +0000 UTC m=+893.371988439" watchObservedRunningTime="2026-02-27 08:03:55.523652881 +0000 UTC m=+893.377582879" Feb 27 08:03:55 crc kubenswrapper[4612]: I0227 08:03:55.536483 4612 scope.go:117] "RemoveContainer" containerID="dbfbf57122d28a8f255e17140c3112baaadef3f59177b1c3d90991a1fae49d29" Feb 27 08:03:55 crc kubenswrapper[4612]: I0227 08:03:55.545369 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-s5f86"] Feb 27 08:03:55 crc kubenswrapper[4612]: I0227 08:03:55.558811 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-s5f86"] Feb 27 08:03:55 crc kubenswrapper[4612]: I0227 08:03:55.586272 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-6f9ff8875f-jphph" podStartSLOduration=1.3208606999999999 podStartE2EDuration="11.586252814s" podCreationTimestamp="2026-02-27 08:03:44 +0000 UTC" firstStartedPulling="2026-02-27 08:03:44.83146209 +0000 UTC m=+882.685392088" lastFinishedPulling="2026-02-27 08:03:55.096854204 +0000 UTC m=+892.950784202" observedRunningTime="2026-02-27 08:03:55.580601852 +0000 UTC m=+893.434531870" watchObservedRunningTime="2026-02-27 08:03:55.586252814 +0000 UTC m=+893.440182822" Feb 27 08:03:56 crc kubenswrapper[4612]: I0227 08:03:56.859924 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5d964ece-5965-43d6-a48d-064f989f03d8" path="/var/lib/kubelet/pods/5d964ece-5965-43d6-a48d-064f989f03d8/volumes" Feb 27 08:04:00 crc kubenswrapper[4612]: I0227 08:04:00.129658 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29536324-2tpql"] Feb 27 08:04:00 crc kubenswrapper[4612]: E0227 08:04:00.130156 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5d964ece-5965-43d6-a48d-064f989f03d8" containerName="extract-content" Feb 27 08:04:00 crc kubenswrapper[4612]: I0227 08:04:00.130168 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="5d964ece-5965-43d6-a48d-064f989f03d8" containerName="extract-content" Feb 27 08:04:00 crc kubenswrapper[4612]: E0227 08:04:00.130189 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5d964ece-5965-43d6-a48d-064f989f03d8" containerName="registry-server" Feb 27 08:04:00 crc kubenswrapper[4612]: I0227 08:04:00.130196 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="5d964ece-5965-43d6-a48d-064f989f03d8" containerName="registry-server" Feb 27 08:04:00 crc kubenswrapper[4612]: E0227 08:04:00.130209 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5d964ece-5965-43d6-a48d-064f989f03d8" containerName="extract-utilities" Feb 27 08:04:00 crc kubenswrapper[4612]: I0227 08:04:00.130215 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="5d964ece-5965-43d6-a48d-064f989f03d8" containerName="extract-utilities" Feb 27 08:04:00 crc kubenswrapper[4612]: I0227 08:04:00.130307 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="5d964ece-5965-43d6-a48d-064f989f03d8" containerName="registry-server" Feb 27 08:04:00 crc kubenswrapper[4612]: I0227 08:04:00.130641 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536324-2tpql" Feb 27 08:04:00 crc kubenswrapper[4612]: I0227 08:04:00.133012 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 27 08:04:00 crc kubenswrapper[4612]: I0227 08:04:00.133024 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-5zwfx" Feb 27 08:04:00 crc kubenswrapper[4612]: I0227 08:04:00.134024 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 27 08:04:00 crc kubenswrapper[4612]: I0227 08:04:00.153271 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536324-2tpql"] Feb 27 08:04:00 crc kubenswrapper[4612]: I0227 08:04:00.215489 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jqzdn\" (UniqueName: \"kubernetes.io/projected/79cb5a3d-2162-4f09-8508-c6102ab5e4cf-kube-api-access-jqzdn\") pod \"auto-csr-approver-29536324-2tpql\" (UID: \"79cb5a3d-2162-4f09-8508-c6102ab5e4cf\") " pod="openshift-infra/auto-csr-approver-29536324-2tpql" Feb 27 08:04:00 crc kubenswrapper[4612]: I0227 08:04:00.317009 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jqzdn\" (UniqueName: \"kubernetes.io/projected/79cb5a3d-2162-4f09-8508-c6102ab5e4cf-kube-api-access-jqzdn\") pod \"auto-csr-approver-29536324-2tpql\" (UID: \"79cb5a3d-2162-4f09-8508-c6102ab5e4cf\") " pod="openshift-infra/auto-csr-approver-29536324-2tpql" Feb 27 08:04:00 crc kubenswrapper[4612]: I0227 08:04:00.339436 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jqzdn\" (UniqueName: \"kubernetes.io/projected/79cb5a3d-2162-4f09-8508-c6102ab5e4cf-kube-api-access-jqzdn\") pod \"auto-csr-approver-29536324-2tpql\" (UID: \"79cb5a3d-2162-4f09-8508-c6102ab5e4cf\") " pod="openshift-infra/auto-csr-approver-29536324-2tpql" Feb 27 08:04:00 crc kubenswrapper[4612]: I0227 08:04:00.458912 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536324-2tpql" Feb 27 08:04:00 crc kubenswrapper[4612]: I0227 08:04:00.663150 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536324-2tpql"] Feb 27 08:04:00 crc kubenswrapper[4612]: W0227 08:04:00.673868 4612 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod79cb5a3d_2162_4f09_8508_c6102ab5e4cf.slice/crio-60e9480ac597482a404062e7bc35e0b820378a40725595065672038a5d5506d2 WatchSource:0}: Error finding container 60e9480ac597482a404062e7bc35e0b820378a40725595065672038a5d5506d2: Status 404 returned error can't find the container with id 60e9480ac597482a404062e7bc35e0b820378a40725595065672038a5d5506d2 Feb 27 08:04:01 crc kubenswrapper[4612]: I0227 08:04:01.531917 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536324-2tpql" event={"ID":"79cb5a3d-2162-4f09-8508-c6102ab5e4cf","Type":"ContainerStarted","Data":"60e9480ac597482a404062e7bc35e0b820378a40725595065672038a5d5506d2"} Feb 27 08:04:02 crc kubenswrapper[4612]: I0227 08:04:02.539892 4612 generic.go:334] "Generic (PLEG): container finished" podID="79cb5a3d-2162-4f09-8508-c6102ab5e4cf" containerID="98c86f4e5a52fa64298df451c4bba1d93fe8e52894afdbdbddb9b042118fb3f1" exitCode=0 Feb 27 08:04:02 crc kubenswrapper[4612]: I0227 08:04:02.539987 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536324-2tpql" event={"ID":"79cb5a3d-2162-4f09-8508-c6102ab5e4cf","Type":"ContainerDied","Data":"98c86f4e5a52fa64298df451c4bba1d93fe8e52894afdbdbddb9b042118fb3f1"} Feb 27 08:04:03 crc kubenswrapper[4612]: I0227 08:04:03.817329 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536324-2tpql" Feb 27 08:04:03 crc kubenswrapper[4612]: I0227 08:04:03.863620 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jqzdn\" (UniqueName: \"kubernetes.io/projected/79cb5a3d-2162-4f09-8508-c6102ab5e4cf-kube-api-access-jqzdn\") pod \"79cb5a3d-2162-4f09-8508-c6102ab5e4cf\" (UID: \"79cb5a3d-2162-4f09-8508-c6102ab5e4cf\") " Feb 27 08:04:03 crc kubenswrapper[4612]: I0227 08:04:03.869391 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/79cb5a3d-2162-4f09-8508-c6102ab5e4cf-kube-api-access-jqzdn" (OuterVolumeSpecName: "kube-api-access-jqzdn") pod "79cb5a3d-2162-4f09-8508-c6102ab5e4cf" (UID: "79cb5a3d-2162-4f09-8508-c6102ab5e4cf"). InnerVolumeSpecName "kube-api-access-jqzdn". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 08:04:03 crc kubenswrapper[4612]: I0227 08:04:03.965627 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jqzdn\" (UniqueName: \"kubernetes.io/projected/79cb5a3d-2162-4f09-8508-c6102ab5e4cf-kube-api-access-jqzdn\") on node \"crc\" DevicePath \"\"" Feb 27 08:04:04 crc kubenswrapper[4612]: I0227 08:04:04.552578 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536324-2tpql" event={"ID":"79cb5a3d-2162-4f09-8508-c6102ab5e4cf","Type":"ContainerDied","Data":"60e9480ac597482a404062e7bc35e0b820378a40725595065672038a5d5506d2"} Feb 27 08:04:04 crc kubenswrapper[4612]: I0227 08:04:04.552615 4612 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="60e9480ac597482a404062e7bc35e0b820378a40725595065672038a5d5506d2" Feb 27 08:04:04 crc kubenswrapper[4612]: I0227 08:04:04.552652 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536324-2tpql" Feb 27 08:04:04 crc kubenswrapper[4612]: I0227 08:04:04.889664 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29536318-gjl85"] Feb 27 08:04:04 crc kubenswrapper[4612]: I0227 08:04:04.895385 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29536318-gjl85"] Feb 27 08:04:06 crc kubenswrapper[4612]: I0227 08:04:06.859785 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d76961be-0bf0-45b8-afed-089eaa480485" path="/var/lib/kubelet/pods/d76961be-0bf0-45b8-afed-089eaa480485/volumes" Feb 27 08:04:14 crc kubenswrapper[4612]: I0227 08:04:14.547397 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-6f9ff8875f-jphph" Feb 27 08:04:34 crc kubenswrapper[4612]: I0227 08:04:34.083222 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-78d5b849b-c2rb5" Feb 27 08:04:34 crc kubenswrapper[4612]: I0227 08:04:34.916646 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-8drht"] Feb 27 08:04:34 crc kubenswrapper[4612]: E0227 08:04:34.917522 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="79cb5a3d-2162-4f09-8508-c6102ab5e4cf" containerName="oc" Feb 27 08:04:34 crc kubenswrapper[4612]: I0227 08:04:34.917546 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="79cb5a3d-2162-4f09-8508-c6102ab5e4cf" containerName="oc" Feb 27 08:04:34 crc kubenswrapper[4612]: I0227 08:04:34.917885 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="79cb5a3d-2162-4f09-8508-c6102ab5e4cf" containerName="oc" Feb 27 08:04:34 crc kubenswrapper[4612]: I0227 08:04:34.922072 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-8drht" Feb 27 08:04:34 crc kubenswrapper[4612]: I0227 08:04:34.929021 4612 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Feb 27 08:04:34 crc kubenswrapper[4612]: I0227 08:04:34.929452 4612 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-f8q9l" Feb 27 08:04:34 crc kubenswrapper[4612]: I0227 08:04:34.930802 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Feb 27 08:04:34 crc kubenswrapper[4612]: I0227 08:04:34.935255 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-7f989f654f-x2k2n"] Feb 27 08:04:34 crc kubenswrapper[4612]: I0227 08:04:34.937876 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7f989f654f-x2k2n" Feb 27 08:04:34 crc kubenswrapper[4612]: I0227 08:04:34.952136 4612 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Feb 27 08:04:34 crc kubenswrapper[4612]: I0227 08:04:34.955533 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7f989f654f-x2k2n"] Feb 27 08:04:35 crc kubenswrapper[4612]: I0227 08:04:35.026359 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/8fe63fdd-8b6f-4ecf-81d2-959ae060d567-frr-conf\") pod \"frr-k8s-8drht\" (UID: \"8fe63fdd-8b6f-4ecf-81d2-959ae060d567\") " pod="metallb-system/frr-k8s-8drht" Feb 27 08:04:35 crc kubenswrapper[4612]: I0227 08:04:35.026411 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/161b64e1-f414-4519-9fc1-c564d1cc0316-cert\") pod \"frr-k8s-webhook-server-7f989f654f-x2k2n\" (UID: \"161b64e1-f414-4519-9fc1-c564d1cc0316\") " pod="metallb-system/frr-k8s-webhook-server-7f989f654f-x2k2n" Feb 27 08:04:35 crc kubenswrapper[4612]: I0227 08:04:35.026460 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8fe63fdd-8b6f-4ecf-81d2-959ae060d567-metrics-certs\") pod \"frr-k8s-8drht\" (UID: \"8fe63fdd-8b6f-4ecf-81d2-959ae060d567\") " pod="metallb-system/frr-k8s-8drht" Feb 27 08:04:35 crc kubenswrapper[4612]: I0227 08:04:35.026482 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/8fe63fdd-8b6f-4ecf-81d2-959ae060d567-metrics\") pod \"frr-k8s-8drht\" (UID: \"8fe63fdd-8b6f-4ecf-81d2-959ae060d567\") " pod="metallb-system/frr-k8s-8drht" Feb 27 08:04:35 crc kubenswrapper[4612]: I0227 08:04:35.026509 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6jdsc\" (UniqueName: \"kubernetes.io/projected/8fe63fdd-8b6f-4ecf-81d2-959ae060d567-kube-api-access-6jdsc\") pod \"frr-k8s-8drht\" (UID: \"8fe63fdd-8b6f-4ecf-81d2-959ae060d567\") " pod="metallb-system/frr-k8s-8drht" Feb 27 08:04:35 crc kubenswrapper[4612]: I0227 08:04:35.026533 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/8fe63fdd-8b6f-4ecf-81d2-959ae060d567-frr-startup\") pod \"frr-k8s-8drht\" (UID: \"8fe63fdd-8b6f-4ecf-81d2-959ae060d567\") " pod="metallb-system/frr-k8s-8drht" Feb 27 08:04:35 crc kubenswrapper[4612]: I0227 08:04:35.026558 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-577dh\" (UniqueName: \"kubernetes.io/projected/161b64e1-f414-4519-9fc1-c564d1cc0316-kube-api-access-577dh\") pod \"frr-k8s-webhook-server-7f989f654f-x2k2n\" (UID: \"161b64e1-f414-4519-9fc1-c564d1cc0316\") " pod="metallb-system/frr-k8s-webhook-server-7f989f654f-x2k2n" Feb 27 08:04:35 crc kubenswrapper[4612]: I0227 08:04:35.026631 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/8fe63fdd-8b6f-4ecf-81d2-959ae060d567-reloader\") pod \"frr-k8s-8drht\" (UID: \"8fe63fdd-8b6f-4ecf-81d2-959ae060d567\") " pod="metallb-system/frr-k8s-8drht" Feb 27 08:04:35 crc kubenswrapper[4612]: I0227 08:04:35.026731 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/8fe63fdd-8b6f-4ecf-81d2-959ae060d567-frr-sockets\") pod \"frr-k8s-8drht\" (UID: \"8fe63fdd-8b6f-4ecf-81d2-959ae060d567\") " pod="metallb-system/frr-k8s-8drht" Feb 27 08:04:35 crc kubenswrapper[4612]: I0227 08:04:35.047264 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-zlmtg"] Feb 27 08:04:35 crc kubenswrapper[4612]: I0227 08:04:35.048435 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-zlmtg" Feb 27 08:04:35 crc kubenswrapper[4612]: I0227 08:04:35.055272 4612 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Feb 27 08:04:35 crc kubenswrapper[4612]: I0227 08:04:35.055298 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Feb 27 08:04:35 crc kubenswrapper[4612]: I0227 08:04:35.055914 4612 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Feb 27 08:04:35 crc kubenswrapper[4612]: I0227 08:04:35.055929 4612 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-9bpmd" Feb 27 08:04:35 crc kubenswrapper[4612]: I0227 08:04:35.058673 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-86ddb6bd46-gk8pf"] Feb 27 08:04:35 crc kubenswrapper[4612]: I0227 08:04:35.059485 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-86ddb6bd46-gk8pf" Feb 27 08:04:35 crc kubenswrapper[4612]: I0227 08:04:35.061834 4612 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Feb 27 08:04:35 crc kubenswrapper[4612]: I0227 08:04:35.070036 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-86ddb6bd46-gk8pf"] Feb 27 08:04:35 crc kubenswrapper[4612]: I0227 08:04:35.127213 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/8fe63fdd-8b6f-4ecf-81d2-959ae060d567-frr-sockets\") pod \"frr-k8s-8drht\" (UID: \"8fe63fdd-8b6f-4ecf-81d2-959ae060d567\") " pod="metallb-system/frr-k8s-8drht" Feb 27 08:04:35 crc kubenswrapper[4612]: I0227 08:04:35.127265 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-878lm\" (UniqueName: \"kubernetes.io/projected/2b49ec1b-e464-4e71-badd-04acfa13ef75-kube-api-access-878lm\") pod \"controller-86ddb6bd46-gk8pf\" (UID: \"2b49ec1b-e464-4e71-badd-04acfa13ef75\") " pod="metallb-system/controller-86ddb6bd46-gk8pf" Feb 27 08:04:35 crc kubenswrapper[4612]: I0227 08:04:35.127287 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9b1f5f78-3c58-42fb-9a8f-0166eab4f500-metrics-certs\") pod \"speaker-zlmtg\" (UID: \"9b1f5f78-3c58-42fb-9a8f-0166eab4f500\") " pod="metallb-system/speaker-zlmtg" Feb 27 08:04:35 crc kubenswrapper[4612]: I0227 08:04:35.127312 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/8fe63fdd-8b6f-4ecf-81d2-959ae060d567-frr-conf\") pod \"frr-k8s-8drht\" (UID: \"8fe63fdd-8b6f-4ecf-81d2-959ae060d567\") " pod="metallb-system/frr-k8s-8drht" Feb 27 08:04:35 crc kubenswrapper[4612]: I0227 08:04:35.127333 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/161b64e1-f414-4519-9fc1-c564d1cc0316-cert\") pod \"frr-k8s-webhook-server-7f989f654f-x2k2n\" (UID: \"161b64e1-f414-4519-9fc1-c564d1cc0316\") " pod="metallb-system/frr-k8s-webhook-server-7f989f654f-x2k2n" Feb 27 08:04:35 crc kubenswrapper[4612]: I0227 08:04:35.127349 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/9b1f5f78-3c58-42fb-9a8f-0166eab4f500-metallb-excludel2\") pod \"speaker-zlmtg\" (UID: \"9b1f5f78-3c58-42fb-9a8f-0166eab4f500\") " pod="metallb-system/speaker-zlmtg" Feb 27 08:04:35 crc kubenswrapper[4612]: I0227 08:04:35.127373 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/9b1f5f78-3c58-42fb-9a8f-0166eab4f500-memberlist\") pod \"speaker-zlmtg\" (UID: \"9b1f5f78-3c58-42fb-9a8f-0166eab4f500\") " pod="metallb-system/speaker-zlmtg" Feb 27 08:04:35 crc kubenswrapper[4612]: I0227 08:04:35.127395 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6sgsb\" (UniqueName: \"kubernetes.io/projected/9b1f5f78-3c58-42fb-9a8f-0166eab4f500-kube-api-access-6sgsb\") pod \"speaker-zlmtg\" (UID: \"9b1f5f78-3c58-42fb-9a8f-0166eab4f500\") " pod="metallb-system/speaker-zlmtg" Feb 27 08:04:35 crc kubenswrapper[4612]: I0227 08:04:35.127417 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2b49ec1b-e464-4e71-badd-04acfa13ef75-metrics-certs\") pod \"controller-86ddb6bd46-gk8pf\" (UID: \"2b49ec1b-e464-4e71-badd-04acfa13ef75\") " pod="metallb-system/controller-86ddb6bd46-gk8pf" Feb 27 08:04:35 crc kubenswrapper[4612]: I0227 08:04:35.127440 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8fe63fdd-8b6f-4ecf-81d2-959ae060d567-metrics-certs\") pod \"frr-k8s-8drht\" (UID: \"8fe63fdd-8b6f-4ecf-81d2-959ae060d567\") " pod="metallb-system/frr-k8s-8drht" Feb 27 08:04:35 crc kubenswrapper[4612]: I0227 08:04:35.127458 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/8fe63fdd-8b6f-4ecf-81d2-959ae060d567-metrics\") pod \"frr-k8s-8drht\" (UID: \"8fe63fdd-8b6f-4ecf-81d2-959ae060d567\") " pod="metallb-system/frr-k8s-8drht" Feb 27 08:04:35 crc kubenswrapper[4612]: I0227 08:04:35.127480 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6jdsc\" (UniqueName: \"kubernetes.io/projected/8fe63fdd-8b6f-4ecf-81d2-959ae060d567-kube-api-access-6jdsc\") pod \"frr-k8s-8drht\" (UID: \"8fe63fdd-8b6f-4ecf-81d2-959ae060d567\") " pod="metallb-system/frr-k8s-8drht" Feb 27 08:04:35 crc kubenswrapper[4612]: I0227 08:04:35.127501 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/8fe63fdd-8b6f-4ecf-81d2-959ae060d567-frr-startup\") pod \"frr-k8s-8drht\" (UID: \"8fe63fdd-8b6f-4ecf-81d2-959ae060d567\") " pod="metallb-system/frr-k8s-8drht" Feb 27 08:04:35 crc kubenswrapper[4612]: I0227 08:04:35.127520 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-577dh\" (UniqueName: \"kubernetes.io/projected/161b64e1-f414-4519-9fc1-c564d1cc0316-kube-api-access-577dh\") pod \"frr-k8s-webhook-server-7f989f654f-x2k2n\" (UID: \"161b64e1-f414-4519-9fc1-c564d1cc0316\") " pod="metallb-system/frr-k8s-webhook-server-7f989f654f-x2k2n" Feb 27 08:04:35 crc kubenswrapper[4612]: I0227 08:04:35.127540 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/8fe63fdd-8b6f-4ecf-81d2-959ae060d567-reloader\") pod \"frr-k8s-8drht\" (UID: \"8fe63fdd-8b6f-4ecf-81d2-959ae060d567\") " pod="metallb-system/frr-k8s-8drht" Feb 27 08:04:35 crc kubenswrapper[4612]: I0227 08:04:35.127556 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/2b49ec1b-e464-4e71-badd-04acfa13ef75-cert\") pod \"controller-86ddb6bd46-gk8pf\" (UID: \"2b49ec1b-e464-4e71-badd-04acfa13ef75\") " pod="metallb-system/controller-86ddb6bd46-gk8pf" Feb 27 08:04:35 crc kubenswrapper[4612]: I0227 08:04:35.127998 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/8fe63fdd-8b6f-4ecf-81d2-959ae060d567-frr-sockets\") pod \"frr-k8s-8drht\" (UID: \"8fe63fdd-8b6f-4ecf-81d2-959ae060d567\") " pod="metallb-system/frr-k8s-8drht" Feb 27 08:04:35 crc kubenswrapper[4612]: I0227 08:04:35.128204 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/8fe63fdd-8b6f-4ecf-81d2-959ae060d567-frr-conf\") pod \"frr-k8s-8drht\" (UID: \"8fe63fdd-8b6f-4ecf-81d2-959ae060d567\") " pod="metallb-system/frr-k8s-8drht" Feb 27 08:04:35 crc kubenswrapper[4612]: E0227 08:04:35.128276 4612 secret.go:188] Couldn't get secret metallb-system/frr-k8s-webhook-server-cert: secret "frr-k8s-webhook-server-cert" not found Feb 27 08:04:35 crc kubenswrapper[4612]: E0227 08:04:35.128323 4612 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/161b64e1-f414-4519-9fc1-c564d1cc0316-cert podName:161b64e1-f414-4519-9fc1-c564d1cc0316 nodeName:}" failed. No retries permitted until 2026-02-27 08:04:35.628304374 +0000 UTC m=+933.482234372 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/161b64e1-f414-4519-9fc1-c564d1cc0316-cert") pod "frr-k8s-webhook-server-7f989f654f-x2k2n" (UID: "161b64e1-f414-4519-9fc1-c564d1cc0316") : secret "frr-k8s-webhook-server-cert" not found Feb 27 08:04:35 crc kubenswrapper[4612]: E0227 08:04:35.128605 4612 secret.go:188] Couldn't get secret metallb-system/frr-k8s-certs-secret: secret "frr-k8s-certs-secret" not found Feb 27 08:04:35 crc kubenswrapper[4612]: E0227 08:04:35.128631 4612 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8fe63fdd-8b6f-4ecf-81d2-959ae060d567-metrics-certs podName:8fe63fdd-8b6f-4ecf-81d2-959ae060d567 nodeName:}" failed. No retries permitted until 2026-02-27 08:04:35.628623063 +0000 UTC m=+933.482553061 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/8fe63fdd-8b6f-4ecf-81d2-959ae060d567-metrics-certs") pod "frr-k8s-8drht" (UID: "8fe63fdd-8b6f-4ecf-81d2-959ae060d567") : secret "frr-k8s-certs-secret" not found Feb 27 08:04:35 crc kubenswrapper[4612]: I0227 08:04:35.128852 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/8fe63fdd-8b6f-4ecf-81d2-959ae060d567-metrics\") pod \"frr-k8s-8drht\" (UID: \"8fe63fdd-8b6f-4ecf-81d2-959ae060d567\") " pod="metallb-system/frr-k8s-8drht" Feb 27 08:04:35 crc kubenswrapper[4612]: I0227 08:04:35.129724 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/8fe63fdd-8b6f-4ecf-81d2-959ae060d567-frr-startup\") pod \"frr-k8s-8drht\" (UID: \"8fe63fdd-8b6f-4ecf-81d2-959ae060d567\") " pod="metallb-system/frr-k8s-8drht" Feb 27 08:04:35 crc kubenswrapper[4612]: I0227 08:04:35.130047 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/8fe63fdd-8b6f-4ecf-81d2-959ae060d567-reloader\") pod \"frr-k8s-8drht\" (UID: \"8fe63fdd-8b6f-4ecf-81d2-959ae060d567\") " pod="metallb-system/frr-k8s-8drht" Feb 27 08:04:35 crc kubenswrapper[4612]: I0227 08:04:35.154840 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6jdsc\" (UniqueName: \"kubernetes.io/projected/8fe63fdd-8b6f-4ecf-81d2-959ae060d567-kube-api-access-6jdsc\") pod \"frr-k8s-8drht\" (UID: \"8fe63fdd-8b6f-4ecf-81d2-959ae060d567\") " pod="metallb-system/frr-k8s-8drht" Feb 27 08:04:35 crc kubenswrapper[4612]: I0227 08:04:35.156317 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-577dh\" (UniqueName: \"kubernetes.io/projected/161b64e1-f414-4519-9fc1-c564d1cc0316-kube-api-access-577dh\") pod \"frr-k8s-webhook-server-7f989f654f-x2k2n\" (UID: \"161b64e1-f414-4519-9fc1-c564d1cc0316\") " pod="metallb-system/frr-k8s-webhook-server-7f989f654f-x2k2n" Feb 27 08:04:35 crc kubenswrapper[4612]: I0227 08:04:35.227899 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2b49ec1b-e464-4e71-badd-04acfa13ef75-metrics-certs\") pod \"controller-86ddb6bd46-gk8pf\" (UID: \"2b49ec1b-e464-4e71-badd-04acfa13ef75\") " pod="metallb-system/controller-86ddb6bd46-gk8pf" Feb 27 08:04:35 crc kubenswrapper[4612]: I0227 08:04:35.227975 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/2b49ec1b-e464-4e71-badd-04acfa13ef75-cert\") pod \"controller-86ddb6bd46-gk8pf\" (UID: \"2b49ec1b-e464-4e71-badd-04acfa13ef75\") " pod="metallb-system/controller-86ddb6bd46-gk8pf" Feb 27 08:04:35 crc kubenswrapper[4612]: I0227 08:04:35.228029 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9b1f5f78-3c58-42fb-9a8f-0166eab4f500-metrics-certs\") pod \"speaker-zlmtg\" (UID: \"9b1f5f78-3c58-42fb-9a8f-0166eab4f500\") " pod="metallb-system/speaker-zlmtg" Feb 27 08:04:35 crc kubenswrapper[4612]: I0227 08:04:35.228051 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-878lm\" (UniqueName: \"kubernetes.io/projected/2b49ec1b-e464-4e71-badd-04acfa13ef75-kube-api-access-878lm\") pod \"controller-86ddb6bd46-gk8pf\" (UID: \"2b49ec1b-e464-4e71-badd-04acfa13ef75\") " pod="metallb-system/controller-86ddb6bd46-gk8pf" Feb 27 08:04:35 crc kubenswrapper[4612]: I0227 08:04:35.228080 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/9b1f5f78-3c58-42fb-9a8f-0166eab4f500-metallb-excludel2\") pod \"speaker-zlmtg\" (UID: \"9b1f5f78-3c58-42fb-9a8f-0166eab4f500\") " pod="metallb-system/speaker-zlmtg" Feb 27 08:04:35 crc kubenswrapper[4612]: I0227 08:04:35.228381 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/9b1f5f78-3c58-42fb-9a8f-0166eab4f500-memberlist\") pod \"speaker-zlmtg\" (UID: \"9b1f5f78-3c58-42fb-9a8f-0166eab4f500\") " pod="metallb-system/speaker-zlmtg" Feb 27 08:04:35 crc kubenswrapper[4612]: I0227 08:04:35.228408 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6sgsb\" (UniqueName: \"kubernetes.io/projected/9b1f5f78-3c58-42fb-9a8f-0166eab4f500-kube-api-access-6sgsb\") pod \"speaker-zlmtg\" (UID: \"9b1f5f78-3c58-42fb-9a8f-0166eab4f500\") " pod="metallb-system/speaker-zlmtg" Feb 27 08:04:35 crc kubenswrapper[4612]: E0227 08:04:35.228503 4612 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Feb 27 08:04:35 crc kubenswrapper[4612]: E0227 08:04:35.228581 4612 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9b1f5f78-3c58-42fb-9a8f-0166eab4f500-memberlist podName:9b1f5f78-3c58-42fb-9a8f-0166eab4f500 nodeName:}" failed. No retries permitted until 2026-02-27 08:04:35.728558524 +0000 UTC m=+933.582488522 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/9b1f5f78-3c58-42fb-9a8f-0166eab4f500-memberlist") pod "speaker-zlmtg" (UID: "9b1f5f78-3c58-42fb-9a8f-0166eab4f500") : secret "metallb-memberlist" not found Feb 27 08:04:35 crc kubenswrapper[4612]: I0227 08:04:35.228860 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/9b1f5f78-3c58-42fb-9a8f-0166eab4f500-metallb-excludel2\") pod \"speaker-zlmtg\" (UID: \"9b1f5f78-3c58-42fb-9a8f-0166eab4f500\") " pod="metallb-system/speaker-zlmtg" Feb 27 08:04:35 crc kubenswrapper[4612]: I0227 08:04:35.229810 4612 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Feb 27 08:04:35 crc kubenswrapper[4612]: I0227 08:04:35.232582 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2b49ec1b-e464-4e71-badd-04acfa13ef75-metrics-certs\") pod \"controller-86ddb6bd46-gk8pf\" (UID: \"2b49ec1b-e464-4e71-badd-04acfa13ef75\") " pod="metallb-system/controller-86ddb6bd46-gk8pf" Feb 27 08:04:35 crc kubenswrapper[4612]: I0227 08:04:35.232893 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/9b1f5f78-3c58-42fb-9a8f-0166eab4f500-metrics-certs\") pod \"speaker-zlmtg\" (UID: \"9b1f5f78-3c58-42fb-9a8f-0166eab4f500\") " pod="metallb-system/speaker-zlmtg" Feb 27 08:04:35 crc kubenswrapper[4612]: I0227 08:04:35.243319 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/2b49ec1b-e464-4e71-badd-04acfa13ef75-cert\") pod \"controller-86ddb6bd46-gk8pf\" (UID: \"2b49ec1b-e464-4e71-badd-04acfa13ef75\") " pod="metallb-system/controller-86ddb6bd46-gk8pf" Feb 27 08:04:35 crc kubenswrapper[4612]: I0227 08:04:35.263302 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6sgsb\" (UniqueName: \"kubernetes.io/projected/9b1f5f78-3c58-42fb-9a8f-0166eab4f500-kube-api-access-6sgsb\") pod \"speaker-zlmtg\" (UID: \"9b1f5f78-3c58-42fb-9a8f-0166eab4f500\") " pod="metallb-system/speaker-zlmtg" Feb 27 08:04:35 crc kubenswrapper[4612]: I0227 08:04:35.278069 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-878lm\" (UniqueName: \"kubernetes.io/projected/2b49ec1b-e464-4e71-badd-04acfa13ef75-kube-api-access-878lm\") pod \"controller-86ddb6bd46-gk8pf\" (UID: \"2b49ec1b-e464-4e71-badd-04acfa13ef75\") " pod="metallb-system/controller-86ddb6bd46-gk8pf" Feb 27 08:04:35 crc kubenswrapper[4612]: I0227 08:04:35.384951 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-86ddb6bd46-gk8pf" Feb 27 08:04:35 crc kubenswrapper[4612]: I0227 08:04:35.569107 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-86ddb6bd46-gk8pf"] Feb 27 08:04:35 crc kubenswrapper[4612]: I0227 08:04:35.633161 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8fe63fdd-8b6f-4ecf-81d2-959ae060d567-metrics-certs\") pod \"frr-k8s-8drht\" (UID: \"8fe63fdd-8b6f-4ecf-81d2-959ae060d567\") " pod="metallb-system/frr-k8s-8drht" Feb 27 08:04:35 crc kubenswrapper[4612]: I0227 08:04:35.633268 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/161b64e1-f414-4519-9fc1-c564d1cc0316-cert\") pod \"frr-k8s-webhook-server-7f989f654f-x2k2n\" (UID: \"161b64e1-f414-4519-9fc1-c564d1cc0316\") " pod="metallb-system/frr-k8s-webhook-server-7f989f654f-x2k2n" Feb 27 08:04:35 crc kubenswrapper[4612]: I0227 08:04:35.638828 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8fe63fdd-8b6f-4ecf-81d2-959ae060d567-metrics-certs\") pod \"frr-k8s-8drht\" (UID: \"8fe63fdd-8b6f-4ecf-81d2-959ae060d567\") " pod="metallb-system/frr-k8s-8drht" Feb 27 08:04:35 crc kubenswrapper[4612]: I0227 08:04:35.639446 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/161b64e1-f414-4519-9fc1-c564d1cc0316-cert\") pod \"frr-k8s-webhook-server-7f989f654f-x2k2n\" (UID: \"161b64e1-f414-4519-9fc1-c564d1cc0316\") " pod="metallb-system/frr-k8s-webhook-server-7f989f654f-x2k2n" Feb 27 08:04:35 crc kubenswrapper[4612]: I0227 08:04:35.734665 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/9b1f5f78-3c58-42fb-9a8f-0166eab4f500-memberlist\") pod \"speaker-zlmtg\" (UID: \"9b1f5f78-3c58-42fb-9a8f-0166eab4f500\") " pod="metallb-system/speaker-zlmtg" Feb 27 08:04:35 crc kubenswrapper[4612]: E0227 08:04:35.735038 4612 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Feb 27 08:04:35 crc kubenswrapper[4612]: E0227 08:04:35.735087 4612 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9b1f5f78-3c58-42fb-9a8f-0166eab4f500-memberlist podName:9b1f5f78-3c58-42fb-9a8f-0166eab4f500 nodeName:}" failed. No retries permitted until 2026-02-27 08:04:36.735071974 +0000 UTC m=+934.589001972 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/9b1f5f78-3c58-42fb-9a8f-0166eab4f500-memberlist") pod "speaker-zlmtg" (UID: "9b1f5f78-3c58-42fb-9a8f-0166eab4f500") : secret "metallb-memberlist" not found Feb 27 08:04:35 crc kubenswrapper[4612]: I0227 08:04:35.744945 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-86ddb6bd46-gk8pf" event={"ID":"2b49ec1b-e464-4e71-badd-04acfa13ef75","Type":"ContainerStarted","Data":"170d832482ff0d5438f932fb9a85de46c5008ec4a1e31534aac43a31d64ab144"} Feb 27 08:04:35 crc kubenswrapper[4612]: I0227 08:04:35.745008 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-86ddb6bd46-gk8pf" event={"ID":"2b49ec1b-e464-4e71-badd-04acfa13ef75","Type":"ContainerStarted","Data":"7e55ef523f725d106629ec719809ac2059f9aef61513253cd5b3cdbcec17ac2f"} Feb 27 08:04:35 crc kubenswrapper[4612]: I0227 08:04:35.862025 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-8drht" Feb 27 08:04:35 crc kubenswrapper[4612]: I0227 08:04:35.897626 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7f989f654f-x2k2n" Feb 27 08:04:36 crc kubenswrapper[4612]: I0227 08:04:36.097428 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7f989f654f-x2k2n"] Feb 27 08:04:36 crc kubenswrapper[4612]: W0227 08:04:36.104445 4612 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod161b64e1_f414_4519_9fc1_c564d1cc0316.slice/crio-99da99f407f6e2aceb502a7cf191ddbe719167a8f1ee775692dad9ba03966928 WatchSource:0}: Error finding container 99da99f407f6e2aceb502a7cf191ddbe719167a8f1ee775692dad9ba03966928: Status 404 returned error can't find the container with id 99da99f407f6e2aceb502a7cf191ddbe719167a8f1ee775692dad9ba03966928 Feb 27 08:04:36 crc kubenswrapper[4612]: I0227 08:04:36.751563 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-8drht" event={"ID":"8fe63fdd-8b6f-4ecf-81d2-959ae060d567","Type":"ContainerStarted","Data":"5bcdad6655c46171ec7021387a6b2129d9461efb02e29ecb34c1ecdf17a62e19"} Feb 27 08:04:36 crc kubenswrapper[4612]: I0227 08:04:36.754685 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-86ddb6bd46-gk8pf" event={"ID":"2b49ec1b-e464-4e71-badd-04acfa13ef75","Type":"ContainerStarted","Data":"2e056d6f823f7dc46f7354f3776cb3eba76d8eb32b94764b9606c9dca72ec382"} Feb 27 08:04:36 crc kubenswrapper[4612]: I0227 08:04:36.754871 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-86ddb6bd46-gk8pf" Feb 27 08:04:36 crc kubenswrapper[4612]: I0227 08:04:36.755442 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7f989f654f-x2k2n" event={"ID":"161b64e1-f414-4519-9fc1-c564d1cc0316","Type":"ContainerStarted","Data":"99da99f407f6e2aceb502a7cf191ddbe719167a8f1ee775692dad9ba03966928"} Feb 27 08:04:36 crc kubenswrapper[4612]: I0227 08:04:36.768656 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/9b1f5f78-3c58-42fb-9a8f-0166eab4f500-memberlist\") pod \"speaker-zlmtg\" (UID: \"9b1f5f78-3c58-42fb-9a8f-0166eab4f500\") " pod="metallb-system/speaker-zlmtg" Feb 27 08:04:36 crc kubenswrapper[4612]: I0227 08:04:36.772176 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-86ddb6bd46-gk8pf" podStartSLOduration=1.772156002 podStartE2EDuration="1.772156002s" podCreationTimestamp="2026-02-27 08:04:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 08:04:36.767787887 +0000 UTC m=+934.621717885" watchObservedRunningTime="2026-02-27 08:04:36.772156002 +0000 UTC m=+934.626086000" Feb 27 08:04:36 crc kubenswrapper[4612]: I0227 08:04:36.775376 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/9b1f5f78-3c58-42fb-9a8f-0166eab4f500-memberlist\") pod \"speaker-zlmtg\" (UID: \"9b1f5f78-3c58-42fb-9a8f-0166eab4f500\") " pod="metallb-system/speaker-zlmtg" Feb 27 08:04:36 crc kubenswrapper[4612]: I0227 08:04:36.868026 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-zlmtg" Feb 27 08:04:37 crc kubenswrapper[4612]: I0227 08:04:37.767285 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-zlmtg" event={"ID":"9b1f5f78-3c58-42fb-9a8f-0166eab4f500","Type":"ContainerStarted","Data":"1402ad3f0b823b372681b08acbc3cb625e829268cc3b076cad5ff1e77c9ecac0"} Feb 27 08:04:37 crc kubenswrapper[4612]: I0227 08:04:37.767635 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-zlmtg" event={"ID":"9b1f5f78-3c58-42fb-9a8f-0166eab4f500","Type":"ContainerStarted","Data":"f76d0d82585d170a27c396c38fc2f93186506c649453b666b992714d578e7a5e"} Feb 27 08:04:37 crc kubenswrapper[4612]: I0227 08:04:37.767650 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-zlmtg" event={"ID":"9b1f5f78-3c58-42fb-9a8f-0166eab4f500","Type":"ContainerStarted","Data":"2799110a036ca36d2f548ad1a34b09e14444bc1557116b5a5b3e6d1e559e7b97"} Feb 27 08:04:37 crc kubenswrapper[4612]: I0227 08:04:37.768018 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-zlmtg" Feb 27 08:04:37 crc kubenswrapper[4612]: I0227 08:04:37.787764 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-zlmtg" podStartSLOduration=2.787742574 podStartE2EDuration="2.787742574s" podCreationTimestamp="2026-02-27 08:04:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 08:04:37.784363178 +0000 UTC m=+935.638293186" watchObservedRunningTime="2026-02-27 08:04:37.787742574 +0000 UTC m=+935.641672572" Feb 27 08:04:43 crc kubenswrapper[4612]: I0227 08:04:43.821465 4612 generic.go:334] "Generic (PLEG): container finished" podID="8fe63fdd-8b6f-4ecf-81d2-959ae060d567" containerID="b89ca2bfe2bb81e622f3e5868d5d1c05f0ead9ba59343048137b9d0458b81d98" exitCode=0 Feb 27 08:04:43 crc kubenswrapper[4612]: I0227 08:04:43.821711 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-8drht" event={"ID":"8fe63fdd-8b6f-4ecf-81d2-959ae060d567","Type":"ContainerDied","Data":"b89ca2bfe2bb81e622f3e5868d5d1c05f0ead9ba59343048137b9d0458b81d98"} Feb 27 08:04:43 crc kubenswrapper[4612]: I0227 08:04:43.824730 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7f989f654f-x2k2n" event={"ID":"161b64e1-f414-4519-9fc1-c564d1cc0316","Type":"ContainerStarted","Data":"6d7381653c22ac84e3d48efa9410dcec0238d0a93b0721ece71f76bf4466ff00"} Feb 27 08:04:43 crc kubenswrapper[4612]: I0227 08:04:43.825210 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-7f989f654f-x2k2n" Feb 27 08:04:43 crc kubenswrapper[4612]: I0227 08:04:43.867465 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-7f989f654f-x2k2n" podStartSLOduration=2.423554633 podStartE2EDuration="9.867448602s" podCreationTimestamp="2026-02-27 08:04:34 +0000 UTC" firstStartedPulling="2026-02-27 08:04:36.10662222 +0000 UTC m=+933.960552218" lastFinishedPulling="2026-02-27 08:04:43.550516189 +0000 UTC m=+941.404446187" observedRunningTime="2026-02-27 08:04:43.867401431 +0000 UTC m=+941.721331439" watchObservedRunningTime="2026-02-27 08:04:43.867448602 +0000 UTC m=+941.721378600" Feb 27 08:04:44 crc kubenswrapper[4612]: I0227 08:04:44.835892 4612 generic.go:334] "Generic (PLEG): container finished" podID="8fe63fdd-8b6f-4ecf-81d2-959ae060d567" containerID="0957d02d54b8c3b8678dc36cc3265faa57b00643bcc762d963ad0974d6313d76" exitCode=0 Feb 27 08:04:44 crc kubenswrapper[4612]: I0227 08:04:44.835991 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-8drht" event={"ID":"8fe63fdd-8b6f-4ecf-81d2-959ae060d567","Type":"ContainerDied","Data":"0957d02d54b8c3b8678dc36cc3265faa57b00643bcc762d963ad0974d6313d76"} Feb 27 08:04:45 crc kubenswrapper[4612]: I0227 08:04:45.391344 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-86ddb6bd46-gk8pf" Feb 27 08:04:45 crc kubenswrapper[4612]: I0227 08:04:45.849014 4612 generic.go:334] "Generic (PLEG): container finished" podID="8fe63fdd-8b6f-4ecf-81d2-959ae060d567" containerID="661378c3516b262b12124ce441f4f6342f73ce8d3d83cb70cae7a5e2184b78bd" exitCode=0 Feb 27 08:04:45 crc kubenswrapper[4612]: I0227 08:04:45.849072 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-8drht" event={"ID":"8fe63fdd-8b6f-4ecf-81d2-959ae060d567","Type":"ContainerDied","Data":"661378c3516b262b12124ce441f4f6342f73ce8d3d83cb70cae7a5e2184b78bd"} Feb 27 08:04:46 crc kubenswrapper[4612]: I0227 08:04:46.026937 4612 patch_prober.go:28] interesting pod/machine-config-daemon-924vb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 27 08:04:46 crc kubenswrapper[4612]: I0227 08:04:46.027231 4612 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 27 08:04:46 crc kubenswrapper[4612]: I0227 08:04:46.883132 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-8drht" event={"ID":"8fe63fdd-8b6f-4ecf-81d2-959ae060d567","Type":"ContainerStarted","Data":"fbf6b0ca3564d6ee52781b33007e7ca284e08f3dc0b1f4d764e4ba476ec2b35b"} Feb 27 08:04:46 crc kubenswrapper[4612]: I0227 08:04:46.883484 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-8drht" event={"ID":"8fe63fdd-8b6f-4ecf-81d2-959ae060d567","Type":"ContainerStarted","Data":"ccdd85c19b5159b85dbe77749b341f14258590f80ff819e88210966a34bdd9c7"} Feb 27 08:04:46 crc kubenswrapper[4612]: I0227 08:04:46.883500 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-8drht" event={"ID":"8fe63fdd-8b6f-4ecf-81d2-959ae060d567","Type":"ContainerStarted","Data":"805a7b4d171f666fffef8263791d6cb473511d7f213213fbe9bb6d7d695a9d61"} Feb 27 08:04:46 crc kubenswrapper[4612]: I0227 08:04:46.883512 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-8drht" event={"ID":"8fe63fdd-8b6f-4ecf-81d2-959ae060d567","Type":"ContainerStarted","Data":"89af7c91aa6a2b887d99ca51ba27c40c32eaf4e70e9b97eff31bafed579f3834"} Feb 27 08:04:47 crc kubenswrapper[4612]: I0227 08:04:47.892643 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-8drht" event={"ID":"8fe63fdd-8b6f-4ecf-81d2-959ae060d567","Type":"ContainerStarted","Data":"62e463ad78c730405c08fa518475ce0153917ef72ae201a70f9c950e3556fed3"} Feb 27 08:04:47 crc kubenswrapper[4612]: I0227 08:04:47.892720 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-8drht" event={"ID":"8fe63fdd-8b6f-4ecf-81d2-959ae060d567","Type":"ContainerStarted","Data":"82fe14e5c7fa8c8e289d1ab6f3bca5016ffecf539091775f6817f000a76411bb"} Feb 27 08:04:47 crc kubenswrapper[4612]: I0227 08:04:47.892866 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-8drht" Feb 27 08:04:50 crc kubenswrapper[4612]: I0227 08:04:50.863215 4612 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-8drht" Feb 27 08:04:50 crc kubenswrapper[4612]: I0227 08:04:50.924991 4612 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-8drht" Feb 27 08:04:50 crc kubenswrapper[4612]: I0227 08:04:50.943708 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-8drht" podStartSLOduration=9.314432822 podStartE2EDuration="16.943680607s" podCreationTimestamp="2026-02-27 08:04:34 +0000 UTC" firstStartedPulling="2026-02-27 08:04:35.956655267 +0000 UTC m=+933.810585265" lastFinishedPulling="2026-02-27 08:04:43.585903052 +0000 UTC m=+941.439833050" observedRunningTime="2026-02-27 08:04:47.917326924 +0000 UTC m=+945.771256932" watchObservedRunningTime="2026-02-27 08:04:50.943680607 +0000 UTC m=+948.797610605" Feb 27 08:04:55 crc kubenswrapper[4612]: I0227 08:04:55.865248 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-8drht" Feb 27 08:04:55 crc kubenswrapper[4612]: I0227 08:04:55.903878 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-7f989f654f-x2k2n" Feb 27 08:04:56 crc kubenswrapper[4612]: I0227 08:04:56.880541 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-zlmtg" Feb 27 08:04:59 crc kubenswrapper[4612]: I0227 08:04:59.918573 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-v64tw"] Feb 27 08:04:59 crc kubenswrapper[4612]: I0227 08:04:59.920653 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-v64tw" Feb 27 08:04:59 crc kubenswrapper[4612]: I0227 08:04:59.926647 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-v64tw"] Feb 27 08:04:59 crc kubenswrapper[4612]: I0227 08:04:59.927185 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Feb 27 08:04:59 crc kubenswrapper[4612]: I0227 08:04:59.927488 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-gk66n" Feb 27 08:04:59 crc kubenswrapper[4612]: I0227 08:04:59.927655 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Feb 27 08:05:00 crc kubenswrapper[4612]: I0227 08:05:00.076801 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c8n76\" (UniqueName: \"kubernetes.io/projected/e3c78923-c929-4743-9b3e-2d9f93a37bf3-kube-api-access-c8n76\") pod \"openstack-operator-index-v64tw\" (UID: \"e3c78923-c929-4743-9b3e-2d9f93a37bf3\") " pod="openstack-operators/openstack-operator-index-v64tw" Feb 27 08:05:00 crc kubenswrapper[4612]: I0227 08:05:00.178271 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c8n76\" (UniqueName: \"kubernetes.io/projected/e3c78923-c929-4743-9b3e-2d9f93a37bf3-kube-api-access-c8n76\") pod \"openstack-operator-index-v64tw\" (UID: \"e3c78923-c929-4743-9b3e-2d9f93a37bf3\") " pod="openstack-operators/openstack-operator-index-v64tw" Feb 27 08:05:00 crc kubenswrapper[4612]: I0227 08:05:00.201965 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c8n76\" (UniqueName: \"kubernetes.io/projected/e3c78923-c929-4743-9b3e-2d9f93a37bf3-kube-api-access-c8n76\") pod \"openstack-operator-index-v64tw\" (UID: \"e3c78923-c929-4743-9b3e-2d9f93a37bf3\") " pod="openstack-operators/openstack-operator-index-v64tw" Feb 27 08:05:00 crc kubenswrapper[4612]: I0227 08:05:00.288287 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-v64tw" Feb 27 08:05:00 crc kubenswrapper[4612]: I0227 08:05:00.694356 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-v64tw"] Feb 27 08:05:01 crc kubenswrapper[4612]: I0227 08:05:01.253399 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-v64tw" event={"ID":"e3c78923-c929-4743-9b3e-2d9f93a37bf3","Type":"ContainerStarted","Data":"a64d6b6618af53a8c6f0d6be48a79819243ce0298088a4c874d6839423fdd709"} Feb 27 08:05:03 crc kubenswrapper[4612]: I0227 08:05:03.085659 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-v64tw"] Feb 27 08:05:03 crc kubenswrapper[4612]: I0227 08:05:03.696331 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-f9cqc"] Feb 27 08:05:03 crc kubenswrapper[4612]: I0227 08:05:03.697293 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-f9cqc" Feb 27 08:05:03 crc kubenswrapper[4612]: I0227 08:05:03.702276 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-f9cqc"] Feb 27 08:05:03 crc kubenswrapper[4612]: I0227 08:05:03.788072 4612 scope.go:117] "RemoveContainer" containerID="5cd4755114c2182b6ac4cf17cfea7c8aa357aac1c10adadb9c49b8bf54cea10e" Feb 27 08:05:03 crc kubenswrapper[4612]: I0227 08:05:03.830591 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2pcvh\" (UniqueName: \"kubernetes.io/projected/5ed1acaa-e783-4d8d-bac0-adf0ab4a3b0f-kube-api-access-2pcvh\") pod \"openstack-operator-index-f9cqc\" (UID: \"5ed1acaa-e783-4d8d-bac0-adf0ab4a3b0f\") " pod="openstack-operators/openstack-operator-index-f9cqc" Feb 27 08:05:03 crc kubenswrapper[4612]: I0227 08:05:03.932470 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2pcvh\" (UniqueName: \"kubernetes.io/projected/5ed1acaa-e783-4d8d-bac0-adf0ab4a3b0f-kube-api-access-2pcvh\") pod \"openstack-operator-index-f9cqc\" (UID: \"5ed1acaa-e783-4d8d-bac0-adf0ab4a3b0f\") " pod="openstack-operators/openstack-operator-index-f9cqc" Feb 27 08:05:03 crc kubenswrapper[4612]: I0227 08:05:03.953755 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2pcvh\" (UniqueName: \"kubernetes.io/projected/5ed1acaa-e783-4d8d-bac0-adf0ab4a3b0f-kube-api-access-2pcvh\") pod \"openstack-operator-index-f9cqc\" (UID: \"5ed1acaa-e783-4d8d-bac0-adf0ab4a3b0f\") " pod="openstack-operators/openstack-operator-index-f9cqc" Feb 27 08:05:04 crc kubenswrapper[4612]: I0227 08:05:04.015852 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-f9cqc" Feb 27 08:05:04 crc kubenswrapper[4612]: I0227 08:05:04.276445 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-v64tw" event={"ID":"e3c78923-c929-4743-9b3e-2d9f93a37bf3","Type":"ContainerStarted","Data":"fb2ab2ee8c9da7c8144963ab3d26499e1e42efdb43f398ff14597056fd315424"} Feb 27 08:05:04 crc kubenswrapper[4612]: I0227 08:05:04.276567 4612 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/openstack-operator-index-v64tw" podUID="e3c78923-c929-4743-9b3e-2d9f93a37bf3" containerName="registry-server" containerID="cri-o://fb2ab2ee8c9da7c8144963ab3d26499e1e42efdb43f398ff14597056fd315424" gracePeriod=2 Feb 27 08:05:04 crc kubenswrapper[4612]: I0227 08:05:04.302120 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-v64tw" podStartSLOduration=2.4879872069999998 podStartE2EDuration="5.302100255s" podCreationTimestamp="2026-02-27 08:04:59 +0000 UTC" firstStartedPulling="2026-02-27 08:05:00.70171308 +0000 UTC m=+958.555643078" lastFinishedPulling="2026-02-27 08:05:03.515826118 +0000 UTC m=+961.369756126" observedRunningTime="2026-02-27 08:05:04.294436496 +0000 UTC m=+962.148366494" watchObservedRunningTime="2026-02-27 08:05:04.302100255 +0000 UTC m=+962.156030273" Feb 27 08:05:04 crc kubenswrapper[4612]: I0227 08:05:04.424172 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-f9cqc"] Feb 27 08:05:04 crc kubenswrapper[4612]: W0227 08:05:04.432113 4612 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5ed1acaa_e783_4d8d_bac0_adf0ab4a3b0f.slice/crio-e0f227cbd4fd78df98d7786e81c3f76b6dd087df254c89ba00063e97305d1e9c WatchSource:0}: Error finding container e0f227cbd4fd78df98d7786e81c3f76b6dd087df254c89ba00063e97305d1e9c: Status 404 returned error can't find the container with id e0f227cbd4fd78df98d7786e81c3f76b6dd087df254c89ba00063e97305d1e9c Feb 27 08:05:04 crc kubenswrapper[4612]: I0227 08:05:04.666172 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-v64tw" Feb 27 08:05:04 crc kubenswrapper[4612]: I0227 08:05:04.742805 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c8n76\" (UniqueName: \"kubernetes.io/projected/e3c78923-c929-4743-9b3e-2d9f93a37bf3-kube-api-access-c8n76\") pod \"e3c78923-c929-4743-9b3e-2d9f93a37bf3\" (UID: \"e3c78923-c929-4743-9b3e-2d9f93a37bf3\") " Feb 27 08:05:04 crc kubenswrapper[4612]: I0227 08:05:04.747731 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e3c78923-c929-4743-9b3e-2d9f93a37bf3-kube-api-access-c8n76" (OuterVolumeSpecName: "kube-api-access-c8n76") pod "e3c78923-c929-4743-9b3e-2d9f93a37bf3" (UID: "e3c78923-c929-4743-9b3e-2d9f93a37bf3"). InnerVolumeSpecName "kube-api-access-c8n76". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 08:05:04 crc kubenswrapper[4612]: I0227 08:05:04.844554 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c8n76\" (UniqueName: \"kubernetes.io/projected/e3c78923-c929-4743-9b3e-2d9f93a37bf3-kube-api-access-c8n76\") on node \"crc\" DevicePath \"\"" Feb 27 08:05:05 crc kubenswrapper[4612]: I0227 08:05:05.293685 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-f9cqc" event={"ID":"5ed1acaa-e783-4d8d-bac0-adf0ab4a3b0f","Type":"ContainerStarted","Data":"a35667078c752816bbc3902579732d0898eb91ed95556559ce157b425fdd927e"} Feb 27 08:05:05 crc kubenswrapper[4612]: I0227 08:05:05.293830 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-f9cqc" event={"ID":"5ed1acaa-e783-4d8d-bac0-adf0ab4a3b0f","Type":"ContainerStarted","Data":"e0f227cbd4fd78df98d7786e81c3f76b6dd087df254c89ba00063e97305d1e9c"} Feb 27 08:05:05 crc kubenswrapper[4612]: I0227 08:05:05.298796 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-v64tw" event={"ID":"e3c78923-c929-4743-9b3e-2d9f93a37bf3","Type":"ContainerDied","Data":"fb2ab2ee8c9da7c8144963ab3d26499e1e42efdb43f398ff14597056fd315424"} Feb 27 08:05:05 crc kubenswrapper[4612]: I0227 08:05:05.299034 4612 scope.go:117] "RemoveContainer" containerID="fb2ab2ee8c9da7c8144963ab3d26499e1e42efdb43f398ff14597056fd315424" Feb 27 08:05:05 crc kubenswrapper[4612]: I0227 08:05:05.299152 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-v64tw" Feb 27 08:05:05 crc kubenswrapper[4612]: I0227 08:05:05.298915 4612 generic.go:334] "Generic (PLEG): container finished" podID="e3c78923-c929-4743-9b3e-2d9f93a37bf3" containerID="fb2ab2ee8c9da7c8144963ab3d26499e1e42efdb43f398ff14597056fd315424" exitCode=0 Feb 27 08:05:05 crc kubenswrapper[4612]: I0227 08:05:05.299201 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-v64tw" event={"ID":"e3c78923-c929-4743-9b3e-2d9f93a37bf3","Type":"ContainerDied","Data":"a64d6b6618af53a8c6f0d6be48a79819243ce0298088a4c874d6839423fdd709"} Feb 27 08:05:05 crc kubenswrapper[4612]: I0227 08:05:05.311763 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-f9cqc" podStartSLOduration=2.245187562 podStartE2EDuration="2.311748818s" podCreationTimestamp="2026-02-27 08:05:03 +0000 UTC" firstStartedPulling="2026-02-27 08:05:04.437089029 +0000 UTC m=+962.291019027" lastFinishedPulling="2026-02-27 08:05:04.503650275 +0000 UTC m=+962.357580283" observedRunningTime="2026-02-27 08:05:05.311358126 +0000 UTC m=+963.165288144" watchObservedRunningTime="2026-02-27 08:05:05.311748818 +0000 UTC m=+963.165678816" Feb 27 08:05:05 crc kubenswrapper[4612]: I0227 08:05:05.329215 4612 scope.go:117] "RemoveContainer" containerID="fb2ab2ee8c9da7c8144963ab3d26499e1e42efdb43f398ff14597056fd315424" Feb 27 08:05:05 crc kubenswrapper[4612]: E0227 08:05:05.330285 4612 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fb2ab2ee8c9da7c8144963ab3d26499e1e42efdb43f398ff14597056fd315424\": container with ID starting with fb2ab2ee8c9da7c8144963ab3d26499e1e42efdb43f398ff14597056fd315424 not found: ID does not exist" containerID="fb2ab2ee8c9da7c8144963ab3d26499e1e42efdb43f398ff14597056fd315424" Feb 27 08:05:05 crc kubenswrapper[4612]: I0227 08:05:05.330355 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fb2ab2ee8c9da7c8144963ab3d26499e1e42efdb43f398ff14597056fd315424"} err="failed to get container status \"fb2ab2ee8c9da7c8144963ab3d26499e1e42efdb43f398ff14597056fd315424\": rpc error: code = NotFound desc = could not find container \"fb2ab2ee8c9da7c8144963ab3d26499e1e42efdb43f398ff14597056fd315424\": container with ID starting with fb2ab2ee8c9da7c8144963ab3d26499e1e42efdb43f398ff14597056fd315424 not found: ID does not exist" Feb 27 08:05:05 crc kubenswrapper[4612]: I0227 08:05:05.331061 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-v64tw"] Feb 27 08:05:05 crc kubenswrapper[4612]: I0227 08:05:05.334436 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/openstack-operator-index-v64tw"] Feb 27 08:05:06 crc kubenswrapper[4612]: I0227 08:05:06.863719 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e3c78923-c929-4743-9b3e-2d9f93a37bf3" path="/var/lib/kubelet/pods/e3c78923-c929-4743-9b3e-2d9f93a37bf3/volumes" Feb 27 08:05:14 crc kubenswrapper[4612]: I0227 08:05:14.016250 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-f9cqc" Feb 27 08:05:14 crc kubenswrapper[4612]: I0227 08:05:14.016666 4612 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-f9cqc" Feb 27 08:05:14 crc kubenswrapper[4612]: I0227 08:05:14.060604 4612 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-f9cqc" Feb 27 08:05:14 crc kubenswrapper[4612]: I0227 08:05:14.398080 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-f9cqc" Feb 27 08:05:15 crc kubenswrapper[4612]: I0227 08:05:15.353791 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/1e2105768d85fd59d354c5be43e261a5f6290b57040396b6c3a32dc181djh99"] Feb 27 08:05:15 crc kubenswrapper[4612]: E0227 08:05:15.354151 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e3c78923-c929-4743-9b3e-2d9f93a37bf3" containerName="registry-server" Feb 27 08:05:15 crc kubenswrapper[4612]: I0227 08:05:15.354174 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="e3c78923-c929-4743-9b3e-2d9f93a37bf3" containerName="registry-server" Feb 27 08:05:15 crc kubenswrapper[4612]: I0227 08:05:15.354417 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="e3c78923-c929-4743-9b3e-2d9f93a37bf3" containerName="registry-server" Feb 27 08:05:15 crc kubenswrapper[4612]: I0227 08:05:15.355920 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/1e2105768d85fd59d354c5be43e261a5f6290b57040396b6c3a32dc181djh99" Feb 27 08:05:15 crc kubenswrapper[4612]: I0227 08:05:15.359321 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-6z8js" Feb 27 08:05:15 crc kubenswrapper[4612]: I0227 08:05:15.382020 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/1e2105768d85fd59d354c5be43e261a5f6290b57040396b6c3a32dc181djh99"] Feb 27 08:05:15 crc kubenswrapper[4612]: I0227 08:05:15.511519 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hjrn6\" (UniqueName: \"kubernetes.io/projected/dfad17e6-8a21-44be-8328-2a156da54578-kube-api-access-hjrn6\") pod \"1e2105768d85fd59d354c5be43e261a5f6290b57040396b6c3a32dc181djh99\" (UID: \"dfad17e6-8a21-44be-8328-2a156da54578\") " pod="openstack-operators/1e2105768d85fd59d354c5be43e261a5f6290b57040396b6c3a32dc181djh99" Feb 27 08:05:15 crc kubenswrapper[4612]: I0227 08:05:15.511585 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/dfad17e6-8a21-44be-8328-2a156da54578-util\") pod \"1e2105768d85fd59d354c5be43e261a5f6290b57040396b6c3a32dc181djh99\" (UID: \"dfad17e6-8a21-44be-8328-2a156da54578\") " pod="openstack-operators/1e2105768d85fd59d354c5be43e261a5f6290b57040396b6c3a32dc181djh99" Feb 27 08:05:15 crc kubenswrapper[4612]: I0227 08:05:15.511632 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/dfad17e6-8a21-44be-8328-2a156da54578-bundle\") pod \"1e2105768d85fd59d354c5be43e261a5f6290b57040396b6c3a32dc181djh99\" (UID: \"dfad17e6-8a21-44be-8328-2a156da54578\") " pod="openstack-operators/1e2105768d85fd59d354c5be43e261a5f6290b57040396b6c3a32dc181djh99" Feb 27 08:05:15 crc kubenswrapper[4612]: I0227 08:05:15.613665 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hjrn6\" (UniqueName: \"kubernetes.io/projected/dfad17e6-8a21-44be-8328-2a156da54578-kube-api-access-hjrn6\") pod \"1e2105768d85fd59d354c5be43e261a5f6290b57040396b6c3a32dc181djh99\" (UID: \"dfad17e6-8a21-44be-8328-2a156da54578\") " pod="openstack-operators/1e2105768d85fd59d354c5be43e261a5f6290b57040396b6c3a32dc181djh99" Feb 27 08:05:15 crc kubenswrapper[4612]: I0227 08:05:15.613735 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/dfad17e6-8a21-44be-8328-2a156da54578-util\") pod \"1e2105768d85fd59d354c5be43e261a5f6290b57040396b6c3a32dc181djh99\" (UID: \"dfad17e6-8a21-44be-8328-2a156da54578\") " pod="openstack-operators/1e2105768d85fd59d354c5be43e261a5f6290b57040396b6c3a32dc181djh99" Feb 27 08:05:15 crc kubenswrapper[4612]: I0227 08:05:15.613781 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/dfad17e6-8a21-44be-8328-2a156da54578-bundle\") pod \"1e2105768d85fd59d354c5be43e261a5f6290b57040396b6c3a32dc181djh99\" (UID: \"dfad17e6-8a21-44be-8328-2a156da54578\") " pod="openstack-operators/1e2105768d85fd59d354c5be43e261a5f6290b57040396b6c3a32dc181djh99" Feb 27 08:05:15 crc kubenswrapper[4612]: I0227 08:05:15.614322 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/dfad17e6-8a21-44be-8328-2a156da54578-bundle\") pod \"1e2105768d85fd59d354c5be43e261a5f6290b57040396b6c3a32dc181djh99\" (UID: \"dfad17e6-8a21-44be-8328-2a156da54578\") " pod="openstack-operators/1e2105768d85fd59d354c5be43e261a5f6290b57040396b6c3a32dc181djh99" Feb 27 08:05:15 crc kubenswrapper[4612]: I0227 08:05:15.614792 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/dfad17e6-8a21-44be-8328-2a156da54578-util\") pod \"1e2105768d85fd59d354c5be43e261a5f6290b57040396b6c3a32dc181djh99\" (UID: \"dfad17e6-8a21-44be-8328-2a156da54578\") " pod="openstack-operators/1e2105768d85fd59d354c5be43e261a5f6290b57040396b6c3a32dc181djh99" Feb 27 08:05:15 crc kubenswrapper[4612]: I0227 08:05:15.657772 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hjrn6\" (UniqueName: \"kubernetes.io/projected/dfad17e6-8a21-44be-8328-2a156da54578-kube-api-access-hjrn6\") pod \"1e2105768d85fd59d354c5be43e261a5f6290b57040396b6c3a32dc181djh99\" (UID: \"dfad17e6-8a21-44be-8328-2a156da54578\") " pod="openstack-operators/1e2105768d85fd59d354c5be43e261a5f6290b57040396b6c3a32dc181djh99" Feb 27 08:05:15 crc kubenswrapper[4612]: I0227 08:05:15.688300 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/1e2105768d85fd59d354c5be43e261a5f6290b57040396b6c3a32dc181djh99" Feb 27 08:05:16 crc kubenswrapper[4612]: I0227 08:05:16.027449 4612 patch_prober.go:28] interesting pod/machine-config-daemon-924vb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 27 08:05:16 crc kubenswrapper[4612]: I0227 08:05:16.028195 4612 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 27 08:05:16 crc kubenswrapper[4612]: I0227 08:05:16.149880 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/1e2105768d85fd59d354c5be43e261a5f6290b57040396b6c3a32dc181djh99"] Feb 27 08:05:16 crc kubenswrapper[4612]: I0227 08:05:16.387935 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/1e2105768d85fd59d354c5be43e261a5f6290b57040396b6c3a32dc181djh99" event={"ID":"dfad17e6-8a21-44be-8328-2a156da54578","Type":"ContainerStarted","Data":"c0bb7612778109c26a4b67ff704732e71d7884fbb61f6393babb4a08e07c1c58"} Feb 27 08:05:16 crc kubenswrapper[4612]: I0227 08:05:16.388000 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/1e2105768d85fd59d354c5be43e261a5f6290b57040396b6c3a32dc181djh99" event={"ID":"dfad17e6-8a21-44be-8328-2a156da54578","Type":"ContainerStarted","Data":"705bb3e79bbc4d5f85eb336f536998eeb2829ade7378953134e7de7870170d64"} Feb 27 08:05:17 crc kubenswrapper[4612]: I0227 08:05:17.400813 4612 generic.go:334] "Generic (PLEG): container finished" podID="dfad17e6-8a21-44be-8328-2a156da54578" containerID="c0bb7612778109c26a4b67ff704732e71d7884fbb61f6393babb4a08e07c1c58" exitCode=0 Feb 27 08:05:17 crc kubenswrapper[4612]: I0227 08:05:17.400864 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/1e2105768d85fd59d354c5be43e261a5f6290b57040396b6c3a32dc181djh99" event={"ID":"dfad17e6-8a21-44be-8328-2a156da54578","Type":"ContainerDied","Data":"c0bb7612778109c26a4b67ff704732e71d7884fbb61f6393babb4a08e07c1c58"} Feb 27 08:05:18 crc kubenswrapper[4612]: I0227 08:05:18.413333 4612 generic.go:334] "Generic (PLEG): container finished" podID="dfad17e6-8a21-44be-8328-2a156da54578" containerID="5ef3dcda2ee2daf0f7ad570ae5af04d0ed0e2358c2656adfd6964087d5f65d1d" exitCode=0 Feb 27 08:05:18 crc kubenswrapper[4612]: I0227 08:05:18.413436 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/1e2105768d85fd59d354c5be43e261a5f6290b57040396b6c3a32dc181djh99" event={"ID":"dfad17e6-8a21-44be-8328-2a156da54578","Type":"ContainerDied","Data":"5ef3dcda2ee2daf0f7ad570ae5af04d0ed0e2358c2656adfd6964087d5f65d1d"} Feb 27 08:05:19 crc kubenswrapper[4612]: I0227 08:05:19.423348 4612 generic.go:334] "Generic (PLEG): container finished" podID="dfad17e6-8a21-44be-8328-2a156da54578" containerID="9d1e4c885d6cec87182ca607b75d865b62579d592b94c63845a6528f1cb05e78" exitCode=0 Feb 27 08:05:19 crc kubenswrapper[4612]: I0227 08:05:19.423529 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/1e2105768d85fd59d354c5be43e261a5f6290b57040396b6c3a32dc181djh99" event={"ID":"dfad17e6-8a21-44be-8328-2a156da54578","Type":"ContainerDied","Data":"9d1e4c885d6cec87182ca607b75d865b62579d592b94c63845a6528f1cb05e78"} Feb 27 08:05:20 crc kubenswrapper[4612]: I0227 08:05:20.707881 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/1e2105768d85fd59d354c5be43e261a5f6290b57040396b6c3a32dc181djh99" Feb 27 08:05:20 crc kubenswrapper[4612]: I0227 08:05:20.889941 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hjrn6\" (UniqueName: \"kubernetes.io/projected/dfad17e6-8a21-44be-8328-2a156da54578-kube-api-access-hjrn6\") pod \"dfad17e6-8a21-44be-8328-2a156da54578\" (UID: \"dfad17e6-8a21-44be-8328-2a156da54578\") " Feb 27 08:05:20 crc kubenswrapper[4612]: I0227 08:05:20.890414 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/dfad17e6-8a21-44be-8328-2a156da54578-bundle\") pod \"dfad17e6-8a21-44be-8328-2a156da54578\" (UID: \"dfad17e6-8a21-44be-8328-2a156da54578\") " Feb 27 08:05:20 crc kubenswrapper[4612]: I0227 08:05:20.890441 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/dfad17e6-8a21-44be-8328-2a156da54578-util\") pod \"dfad17e6-8a21-44be-8328-2a156da54578\" (UID: \"dfad17e6-8a21-44be-8328-2a156da54578\") " Feb 27 08:05:20 crc kubenswrapper[4612]: I0227 08:05:20.892106 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dfad17e6-8a21-44be-8328-2a156da54578-bundle" (OuterVolumeSpecName: "bundle") pod "dfad17e6-8a21-44be-8328-2a156da54578" (UID: "dfad17e6-8a21-44be-8328-2a156da54578"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 08:05:20 crc kubenswrapper[4612]: I0227 08:05:20.901052 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dfad17e6-8a21-44be-8328-2a156da54578-kube-api-access-hjrn6" (OuterVolumeSpecName: "kube-api-access-hjrn6") pod "dfad17e6-8a21-44be-8328-2a156da54578" (UID: "dfad17e6-8a21-44be-8328-2a156da54578"). InnerVolumeSpecName "kube-api-access-hjrn6". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 08:05:20 crc kubenswrapper[4612]: I0227 08:05:20.911894 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dfad17e6-8a21-44be-8328-2a156da54578-util" (OuterVolumeSpecName: "util") pod "dfad17e6-8a21-44be-8328-2a156da54578" (UID: "dfad17e6-8a21-44be-8328-2a156da54578"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 08:05:20 crc kubenswrapper[4612]: I0227 08:05:20.991730 4612 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/dfad17e6-8a21-44be-8328-2a156da54578-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 08:05:20 crc kubenswrapper[4612]: I0227 08:05:20.991753 4612 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/dfad17e6-8a21-44be-8328-2a156da54578-util\") on node \"crc\" DevicePath \"\"" Feb 27 08:05:20 crc kubenswrapper[4612]: I0227 08:05:20.991763 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hjrn6\" (UniqueName: \"kubernetes.io/projected/dfad17e6-8a21-44be-8328-2a156da54578-kube-api-access-hjrn6\") on node \"crc\" DevicePath \"\"" Feb 27 08:05:21 crc kubenswrapper[4612]: I0227 08:05:21.441674 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/1e2105768d85fd59d354c5be43e261a5f6290b57040396b6c3a32dc181djh99" event={"ID":"dfad17e6-8a21-44be-8328-2a156da54578","Type":"ContainerDied","Data":"705bb3e79bbc4d5f85eb336f536998eeb2829ade7378953134e7de7870170d64"} Feb 27 08:05:21 crc kubenswrapper[4612]: I0227 08:05:21.441776 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/1e2105768d85fd59d354c5be43e261a5f6290b57040396b6c3a32dc181djh99" Feb 27 08:05:21 crc kubenswrapper[4612]: I0227 08:05:21.441779 4612 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="705bb3e79bbc4d5f85eb336f536998eeb2829ade7378953134e7de7870170d64" Feb 27 08:05:28 crc kubenswrapper[4612]: I0227 08:05:28.166593 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-init-8667ddcbf9-7stg7"] Feb 27 08:05:28 crc kubenswrapper[4612]: E0227 08:05:28.167346 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dfad17e6-8a21-44be-8328-2a156da54578" containerName="extract" Feb 27 08:05:28 crc kubenswrapper[4612]: I0227 08:05:28.167365 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="dfad17e6-8a21-44be-8328-2a156da54578" containerName="extract" Feb 27 08:05:28 crc kubenswrapper[4612]: E0227 08:05:28.167389 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dfad17e6-8a21-44be-8328-2a156da54578" containerName="util" Feb 27 08:05:28 crc kubenswrapper[4612]: I0227 08:05:28.167397 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="dfad17e6-8a21-44be-8328-2a156da54578" containerName="util" Feb 27 08:05:28 crc kubenswrapper[4612]: E0227 08:05:28.167410 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dfad17e6-8a21-44be-8328-2a156da54578" containerName="pull" Feb 27 08:05:28 crc kubenswrapper[4612]: I0227 08:05:28.167416 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="dfad17e6-8a21-44be-8328-2a156da54578" containerName="pull" Feb 27 08:05:28 crc kubenswrapper[4612]: I0227 08:05:28.167527 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="dfad17e6-8a21-44be-8328-2a156da54578" containerName="extract" Feb 27 08:05:28 crc kubenswrapper[4612]: I0227 08:05:28.167950 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-init-8667ddcbf9-7stg7" Feb 27 08:05:28 crc kubenswrapper[4612]: I0227 08:05:28.171532 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-init-dockercfg-w75cz" Feb 27 08:05:28 crc kubenswrapper[4612]: I0227 08:05:28.191058 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-init-8667ddcbf9-7stg7"] Feb 27 08:05:28 crc kubenswrapper[4612]: I0227 08:05:28.292423 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nbrlk\" (UniqueName: \"kubernetes.io/projected/a8e2b07a-a4cf-4a98-afeb-2a7554b687ee-kube-api-access-nbrlk\") pod \"openstack-operator-controller-init-8667ddcbf9-7stg7\" (UID: \"a8e2b07a-a4cf-4a98-afeb-2a7554b687ee\") " pod="openstack-operators/openstack-operator-controller-init-8667ddcbf9-7stg7" Feb 27 08:05:28 crc kubenswrapper[4612]: I0227 08:05:28.394090 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nbrlk\" (UniqueName: \"kubernetes.io/projected/a8e2b07a-a4cf-4a98-afeb-2a7554b687ee-kube-api-access-nbrlk\") pod \"openstack-operator-controller-init-8667ddcbf9-7stg7\" (UID: \"a8e2b07a-a4cf-4a98-afeb-2a7554b687ee\") " pod="openstack-operators/openstack-operator-controller-init-8667ddcbf9-7stg7" Feb 27 08:05:28 crc kubenswrapper[4612]: I0227 08:05:28.427622 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nbrlk\" (UniqueName: \"kubernetes.io/projected/a8e2b07a-a4cf-4a98-afeb-2a7554b687ee-kube-api-access-nbrlk\") pod \"openstack-operator-controller-init-8667ddcbf9-7stg7\" (UID: \"a8e2b07a-a4cf-4a98-afeb-2a7554b687ee\") " pod="openstack-operators/openstack-operator-controller-init-8667ddcbf9-7stg7" Feb 27 08:05:28 crc kubenswrapper[4612]: I0227 08:05:28.484230 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-init-8667ddcbf9-7stg7" Feb 27 08:05:28 crc kubenswrapper[4612]: I0227 08:05:28.964669 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-init-8667ddcbf9-7stg7"] Feb 27 08:05:29 crc kubenswrapper[4612]: I0227 08:05:29.502556 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-init-8667ddcbf9-7stg7" event={"ID":"a8e2b07a-a4cf-4a98-afeb-2a7554b687ee","Type":"ContainerStarted","Data":"e9d31ce2eb89f620170b783eb0e95bafc21d48176d2cc6534208a8d23c08a53c"} Feb 27 08:05:33 crc kubenswrapper[4612]: I0227 08:05:33.527437 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-init-8667ddcbf9-7stg7" event={"ID":"a8e2b07a-a4cf-4a98-afeb-2a7554b687ee","Type":"ContainerStarted","Data":"ee8c7f19c5dd9336dc62a657d6c0578c09a09ff5de7944d1908ff5602b652a1e"} Feb 27 08:05:33 crc kubenswrapper[4612]: I0227 08:05:33.528090 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-init-8667ddcbf9-7stg7" Feb 27 08:05:33 crc kubenswrapper[4612]: I0227 08:05:33.563293 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-init-8667ddcbf9-7stg7" podStartSLOduration=1.601976743 podStartE2EDuration="5.563274659s" podCreationTimestamp="2026-02-27 08:05:28 +0000 UTC" firstStartedPulling="2026-02-27 08:05:28.986275317 +0000 UTC m=+986.840205315" lastFinishedPulling="2026-02-27 08:05:32.947573233 +0000 UTC m=+990.801503231" observedRunningTime="2026-02-27 08:05:33.558517262 +0000 UTC m=+991.412447260" watchObservedRunningTime="2026-02-27 08:05:33.563274659 +0000 UTC m=+991.417204657" Feb 27 08:05:38 crc kubenswrapper[4612]: I0227 08:05:38.487479 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-init-8667ddcbf9-7stg7" Feb 27 08:05:46 crc kubenswrapper[4612]: I0227 08:05:46.027300 4612 patch_prober.go:28] interesting pod/machine-config-daemon-924vb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 27 08:05:46 crc kubenswrapper[4612]: I0227 08:05:46.027625 4612 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 27 08:05:46 crc kubenswrapper[4612]: I0227 08:05:46.027674 4612 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-924vb" Feb 27 08:05:46 crc kubenswrapper[4612]: I0227 08:05:46.028251 4612 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"3ee7350e6fe9f892458f31227d045e092d0ee2f4312ea6cc953bb5da78fe0fa7"} pod="openshift-machine-config-operator/machine-config-daemon-924vb" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 27 08:05:46 crc kubenswrapper[4612]: I0227 08:05:46.028296 4612 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" containerName="machine-config-daemon" containerID="cri-o://3ee7350e6fe9f892458f31227d045e092d0ee2f4312ea6cc953bb5da78fe0fa7" gracePeriod=600 Feb 27 08:05:46 crc kubenswrapper[4612]: I0227 08:05:46.603500 4612 generic.go:334] "Generic (PLEG): container finished" podID="24599373-7adc-4a1b-8bb4-797bf726f43d" containerID="3ee7350e6fe9f892458f31227d045e092d0ee2f4312ea6cc953bb5da78fe0fa7" exitCode=0 Feb 27 08:05:46 crc kubenswrapper[4612]: I0227 08:05:46.603588 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-924vb" event={"ID":"24599373-7adc-4a1b-8bb4-797bf726f43d","Type":"ContainerDied","Data":"3ee7350e6fe9f892458f31227d045e092d0ee2f4312ea6cc953bb5da78fe0fa7"} Feb 27 08:05:46 crc kubenswrapper[4612]: I0227 08:05:46.603838 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-924vb" event={"ID":"24599373-7adc-4a1b-8bb4-797bf726f43d","Type":"ContainerStarted","Data":"8e7aef6fa8ec4b3d8f700070d994aef18691e45f133524f54d3ffb63aa703a66"} Feb 27 08:05:46 crc kubenswrapper[4612]: I0227 08:05:46.603860 4612 scope.go:117] "RemoveContainer" containerID="b0982c79ab30408b51bb931dbdd8a46865a7a758c57125e452728efbaccdd1ec" Feb 27 08:05:57 crc kubenswrapper[4612]: I0227 08:05:57.520445 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-55d77d7b5c-462m7"] Feb 27 08:05:57 crc kubenswrapper[4612]: I0227 08:05:57.521757 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-55d77d7b5c-462m7" Feb 27 08:05:57 crc kubenswrapper[4612]: I0227 08:05:57.529750 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-jhbbj" Feb 27 08:05:57 crc kubenswrapper[4612]: I0227 08:05:57.533322 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-868647ff47-mlgz7"] Feb 27 08:05:57 crc kubenswrapper[4612]: I0227 08:05:57.534240 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-868647ff47-mlgz7" Feb 27 08:05:57 crc kubenswrapper[4612]: I0227 08:05:57.536319 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-9v7ws" Feb 27 08:05:57 crc kubenswrapper[4612]: I0227 08:05:57.548782 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-868647ff47-mlgz7"] Feb 27 08:05:57 crc kubenswrapper[4612]: I0227 08:05:57.572190 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-784b5bb6c5-g4p5v"] Feb 27 08:05:57 crc kubenswrapper[4612]: I0227 08:05:57.573071 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-784b5bb6c5-g4p5v" Feb 27 08:05:57 crc kubenswrapper[4612]: I0227 08:05:57.575793 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-2kvtq" Feb 27 08:05:57 crc kubenswrapper[4612]: I0227 08:05:57.579327 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-6d8bf5c495-t95hp"] Feb 27 08:05:57 crc kubenswrapper[4612]: I0227 08:05:57.580062 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-6d8bf5c495-t95hp" Feb 27 08:05:57 crc kubenswrapper[4612]: I0227 08:05:57.591826 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-784b5bb6c5-g4p5v"] Feb 27 08:05:57 crc kubenswrapper[4612]: I0227 08:05:57.595042 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-rgvkh" Feb 27 08:05:57 crc kubenswrapper[4612]: I0227 08:05:57.620884 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-55d77d7b5c-462m7"] Feb 27 08:05:57 crc kubenswrapper[4612]: I0227 08:05:57.626774 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-6d8bf5c495-t95hp"] Feb 27 08:05:57 crc kubenswrapper[4612]: I0227 08:05:57.629527 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-69f49c598c-7f9bj"] Feb 27 08:05:57 crc kubenswrapper[4612]: I0227 08:05:57.630209 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-69f49c598c-7f9bj" Feb 27 08:05:57 crc kubenswrapper[4612]: I0227 08:05:57.632120 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-mpl9h" Feb 27 08:05:57 crc kubenswrapper[4612]: I0227 08:05:57.649329 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-5b9b8895d5-t4rct"] Feb 27 08:05:57 crc kubenswrapper[4612]: I0227 08:05:57.650025 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-5b9b8895d5-t4rct" Feb 27 08:05:57 crc kubenswrapper[4612]: I0227 08:05:57.653169 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-s4gj6" Feb 27 08:05:57 crc kubenswrapper[4612]: I0227 08:05:57.658581 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-69f49c598c-7f9bj"] Feb 27 08:05:57 crc kubenswrapper[4612]: I0227 08:05:57.670752 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-79d975b745-plm9k"] Feb 27 08:05:57 crc kubenswrapper[4612]: I0227 08:05:57.671609 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-79d975b745-plm9k" Feb 27 08:05:57 crc kubenswrapper[4612]: I0227 08:05:57.673472 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Feb 27 08:05:57 crc kubenswrapper[4612]: I0227 08:05:57.675024 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-g2x8p" Feb 27 08:05:57 crc kubenswrapper[4612]: I0227 08:05:57.699857 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-79d975b745-plm9k"] Feb 27 08:05:57 crc kubenswrapper[4612]: I0227 08:05:57.704529 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f5pcx\" (UniqueName: \"kubernetes.io/projected/63d519c2-d2f9-4349-ad82-0399e472d78a-kube-api-access-f5pcx\") pod \"designate-operator-controller-manager-6d8bf5c495-t95hp\" (UID: \"63d519c2-d2f9-4349-ad82-0399e472d78a\") " pod="openstack-operators/designate-operator-controller-manager-6d8bf5c495-t95hp" Feb 27 08:05:57 crc kubenswrapper[4612]: I0227 08:05:57.704601 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-br2gx\" (UniqueName: \"kubernetes.io/projected/2ddc15ea-b7c1-44e7-b795-392890e9c14e-kube-api-access-br2gx\") pod \"barbican-operator-controller-manager-868647ff47-mlgz7\" (UID: \"2ddc15ea-b7c1-44e7-b795-392890e9c14e\") " pod="openstack-operators/barbican-operator-controller-manager-868647ff47-mlgz7" Feb 27 08:05:57 crc kubenswrapper[4612]: I0227 08:05:57.704797 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8m9j9\" (UniqueName: \"kubernetes.io/projected/71c482bc-7818-4193-a80b-323682cdcfb3-kube-api-access-8m9j9\") pod \"glance-operator-controller-manager-784b5bb6c5-g4p5v\" (UID: \"71c482bc-7818-4193-a80b-323682cdcfb3\") " pod="openstack-operators/glance-operator-controller-manager-784b5bb6c5-g4p5v" Feb 27 08:05:57 crc kubenswrapper[4612]: I0227 08:05:57.704836 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b5897\" (UniqueName: \"kubernetes.io/projected/36fc734e-4b3e-46bb-b914-896b1ec5c2ee-kube-api-access-b5897\") pod \"cinder-operator-controller-manager-55d77d7b5c-462m7\" (UID: \"36fc734e-4b3e-46bb-b914-896b1ec5c2ee\") " pod="openstack-operators/cinder-operator-controller-manager-55d77d7b5c-462m7" Feb 27 08:05:57 crc kubenswrapper[4612]: I0227 08:05:57.756756 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-5b9b8895d5-t4rct"] Feb 27 08:05:57 crc kubenswrapper[4612]: I0227 08:05:57.788883 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-554564d7fc-8kpwf"] Feb 27 08:05:57 crc kubenswrapper[4612]: I0227 08:05:57.789851 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-b4d948c87-2hm5l"] Feb 27 08:05:57 crc kubenswrapper[4612]: I0227 08:05:57.790233 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-554564d7fc-8kpwf" Feb 27 08:05:57 crc kubenswrapper[4612]: I0227 08:05:57.790300 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-b4d948c87-2hm5l" Feb 27 08:05:57 crc kubenswrapper[4612]: I0227 08:05:57.799536 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-hqdgn" Feb 27 08:05:57 crc kubenswrapper[4612]: I0227 08:05:57.799750 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-fr5lp" Feb 27 08:05:57 crc kubenswrapper[4612]: I0227 08:05:57.806112 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f5pcx\" (UniqueName: \"kubernetes.io/projected/63d519c2-d2f9-4349-ad82-0399e472d78a-kube-api-access-f5pcx\") pod \"designate-operator-controller-manager-6d8bf5c495-t95hp\" (UID: \"63d519c2-d2f9-4349-ad82-0399e472d78a\") " pod="openstack-operators/designate-operator-controller-manager-6d8bf5c495-t95hp" Feb 27 08:05:57 crc kubenswrapper[4612]: I0227 08:05:57.806164 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-br2gx\" (UniqueName: \"kubernetes.io/projected/2ddc15ea-b7c1-44e7-b795-392890e9c14e-kube-api-access-br2gx\") pod \"barbican-operator-controller-manager-868647ff47-mlgz7\" (UID: \"2ddc15ea-b7c1-44e7-b795-392890e9c14e\") " pod="openstack-operators/barbican-operator-controller-manager-868647ff47-mlgz7" Feb 27 08:05:57 crc kubenswrapper[4612]: I0227 08:05:57.806190 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zcj9h\" (UniqueName: \"kubernetes.io/projected/98296d45-b487-47c3-a44a-fa63ba52426b-kube-api-access-zcj9h\") pod \"infra-operator-controller-manager-79d975b745-plm9k\" (UID: \"98296d45-b487-47c3-a44a-fa63ba52426b\") " pod="openstack-operators/infra-operator-controller-manager-79d975b745-plm9k" Feb 27 08:05:57 crc kubenswrapper[4612]: I0227 08:05:57.806223 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tmdcx\" (UniqueName: \"kubernetes.io/projected/1903598c-2428-4e90-ab81-9a47fa5af9ef-kube-api-access-tmdcx\") pod \"horizon-operator-controller-manager-5b9b8895d5-t4rct\" (UID: \"1903598c-2428-4e90-ab81-9a47fa5af9ef\") " pod="openstack-operators/horizon-operator-controller-manager-5b9b8895d5-t4rct" Feb 27 08:05:57 crc kubenswrapper[4612]: I0227 08:05:57.806244 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8m9j9\" (UniqueName: \"kubernetes.io/projected/71c482bc-7818-4193-a80b-323682cdcfb3-kube-api-access-8m9j9\") pod \"glance-operator-controller-manager-784b5bb6c5-g4p5v\" (UID: \"71c482bc-7818-4193-a80b-323682cdcfb3\") " pod="openstack-operators/glance-operator-controller-manager-784b5bb6c5-g4p5v" Feb 27 08:05:57 crc kubenswrapper[4612]: I0227 08:05:57.806267 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gmvbz\" (UniqueName: \"kubernetes.io/projected/6065186a-5435-4db3-8adc-698bec4358ef-kube-api-access-gmvbz\") pod \"heat-operator-controller-manager-69f49c598c-7f9bj\" (UID: \"6065186a-5435-4db3-8adc-698bec4358ef\") " pod="openstack-operators/heat-operator-controller-manager-69f49c598c-7f9bj" Feb 27 08:05:57 crc kubenswrapper[4612]: I0227 08:05:57.806286 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b5897\" (UniqueName: \"kubernetes.io/projected/36fc734e-4b3e-46bb-b914-896b1ec5c2ee-kube-api-access-b5897\") pod \"cinder-operator-controller-manager-55d77d7b5c-462m7\" (UID: \"36fc734e-4b3e-46bb-b914-896b1ec5c2ee\") " pod="openstack-operators/cinder-operator-controller-manager-55d77d7b5c-462m7" Feb 27 08:05:57 crc kubenswrapper[4612]: I0227 08:05:57.806302 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/98296d45-b487-47c3-a44a-fa63ba52426b-cert\") pod \"infra-operator-controller-manager-79d975b745-plm9k\" (UID: \"98296d45-b487-47c3-a44a-fa63ba52426b\") " pod="openstack-operators/infra-operator-controller-manager-79d975b745-plm9k" Feb 27 08:05:57 crc kubenswrapper[4612]: I0227 08:05:57.807771 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-67d996989d-4dhhw"] Feb 27 08:05:57 crc kubenswrapper[4612]: I0227 08:05:57.808576 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-67d996989d-4dhhw" Feb 27 08:05:57 crc kubenswrapper[4612]: I0227 08:05:57.812855 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-nwsnp" Feb 27 08:05:57 crc kubenswrapper[4612]: I0227 08:05:57.829392 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-554564d7fc-8kpwf"] Feb 27 08:05:57 crc kubenswrapper[4612]: I0227 08:05:57.847786 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f5pcx\" (UniqueName: \"kubernetes.io/projected/63d519c2-d2f9-4349-ad82-0399e472d78a-kube-api-access-f5pcx\") pod \"designate-operator-controller-manager-6d8bf5c495-t95hp\" (UID: \"63d519c2-d2f9-4349-ad82-0399e472d78a\") " pod="openstack-operators/designate-operator-controller-manager-6d8bf5c495-t95hp" Feb 27 08:05:57 crc kubenswrapper[4612]: I0227 08:05:57.849445 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8m9j9\" (UniqueName: \"kubernetes.io/projected/71c482bc-7818-4193-a80b-323682cdcfb3-kube-api-access-8m9j9\") pod \"glance-operator-controller-manager-784b5bb6c5-g4p5v\" (UID: \"71c482bc-7818-4193-a80b-323682cdcfb3\") " pod="openstack-operators/glance-operator-controller-manager-784b5bb6c5-g4p5v" Feb 27 08:05:57 crc kubenswrapper[4612]: I0227 08:05:57.850625 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-br2gx\" (UniqueName: \"kubernetes.io/projected/2ddc15ea-b7c1-44e7-b795-392890e9c14e-kube-api-access-br2gx\") pod \"barbican-operator-controller-manager-868647ff47-mlgz7\" (UID: \"2ddc15ea-b7c1-44e7-b795-392890e9c14e\") " pod="openstack-operators/barbican-operator-controller-manager-868647ff47-mlgz7" Feb 27 08:05:57 crc kubenswrapper[4612]: I0227 08:05:57.850924 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-868647ff47-mlgz7" Feb 27 08:05:57 crc kubenswrapper[4612]: I0227 08:05:57.851302 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b5897\" (UniqueName: \"kubernetes.io/projected/36fc734e-4b3e-46bb-b914-896b1ec5c2ee-kube-api-access-b5897\") pod \"cinder-operator-controller-manager-55d77d7b5c-462m7\" (UID: \"36fc734e-4b3e-46bb-b914-896b1ec5c2ee\") " pod="openstack-operators/cinder-operator-controller-manager-55d77d7b5c-462m7" Feb 27 08:05:57 crc kubenswrapper[4612]: I0227 08:05:57.867998 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-b4d948c87-2hm5l"] Feb 27 08:05:57 crc kubenswrapper[4612]: I0227 08:05:57.874797 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-67d996989d-4dhhw"] Feb 27 08:05:57 crc kubenswrapper[4612]: I0227 08:05:57.895469 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-784b5bb6c5-g4p5v" Feb 27 08:05:57 crc kubenswrapper[4612]: I0227 08:05:57.909967 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-6994f66f48-qs4l7"] Feb 27 08:05:57 crc kubenswrapper[4612]: I0227 08:05:57.910990 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-6994f66f48-qs4l7" Feb 27 08:05:57 crc kubenswrapper[4612]: I0227 08:05:57.913720 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-75k5x" Feb 27 08:05:57 crc kubenswrapper[4612]: I0227 08:05:57.914143 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-6d8bf5c495-t95hp" Feb 27 08:05:57 crc kubenswrapper[4612]: I0227 08:05:57.915439 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fqgp9\" (UniqueName: \"kubernetes.io/projected/b60d4209-cd53-4cfe-99f7-7e7351041857-kube-api-access-fqgp9\") pod \"manila-operator-controller-manager-67d996989d-4dhhw\" (UID: \"b60d4209-cd53-4cfe-99f7-7e7351041857\") " pod="openstack-operators/manila-operator-controller-manager-67d996989d-4dhhw" Feb 27 08:05:57 crc kubenswrapper[4612]: I0227 08:05:57.926302 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-6bd4687957-2qmzb"] Feb 27 08:05:57 crc kubenswrapper[4612]: I0227 08:05:57.941943 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-6bd4687957-2qmzb" Feb 27 08:05:57 crc kubenswrapper[4612]: I0227 08:05:57.944953 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-cfpgc" Feb 27 08:05:57 crc kubenswrapper[4612]: I0227 08:05:57.945132 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-6994f66f48-qs4l7"] Feb 27 08:05:57 crc kubenswrapper[4612]: I0227 08:05:57.953905 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zcj9h\" (UniqueName: \"kubernetes.io/projected/98296d45-b487-47c3-a44a-fa63ba52426b-kube-api-access-zcj9h\") pod \"infra-operator-controller-manager-79d975b745-plm9k\" (UID: \"98296d45-b487-47c3-a44a-fa63ba52426b\") " pod="openstack-operators/infra-operator-controller-manager-79d975b745-plm9k" Feb 27 08:05:57 crc kubenswrapper[4612]: I0227 08:05:57.953983 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lflq8\" (UniqueName: \"kubernetes.io/projected/259df4bd-af1b-4c61-a5bd-0ca28ac531d4-kube-api-access-lflq8\") pod \"keystone-operator-controller-manager-b4d948c87-2hm5l\" (UID: \"259df4bd-af1b-4c61-a5bd-0ca28ac531d4\") " pod="openstack-operators/keystone-operator-controller-manager-b4d948c87-2hm5l" Feb 27 08:05:57 crc kubenswrapper[4612]: I0227 08:05:57.954056 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hw2ws\" (UniqueName: \"kubernetes.io/projected/f62fea1f-71de-4872-9baa-42ad2448d5f5-kube-api-access-hw2ws\") pod \"mariadb-operator-controller-manager-6994f66f48-qs4l7\" (UID: \"f62fea1f-71de-4872-9baa-42ad2448d5f5\") " pod="openstack-operators/mariadb-operator-controller-manager-6994f66f48-qs4l7" Feb 27 08:05:57 crc kubenswrapper[4612]: I0227 08:05:57.954096 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tmdcx\" (UniqueName: \"kubernetes.io/projected/1903598c-2428-4e90-ab81-9a47fa5af9ef-kube-api-access-tmdcx\") pod \"horizon-operator-controller-manager-5b9b8895d5-t4rct\" (UID: \"1903598c-2428-4e90-ab81-9a47fa5af9ef\") " pod="openstack-operators/horizon-operator-controller-manager-5b9b8895d5-t4rct" Feb 27 08:05:57 crc kubenswrapper[4612]: I0227 08:05:57.954129 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j284h\" (UniqueName: \"kubernetes.io/projected/bd687bfa-c843-4d11-80f9-cedcdce41e3d-kube-api-access-j284h\") pod \"ironic-operator-controller-manager-554564d7fc-8kpwf\" (UID: \"bd687bfa-c843-4d11-80f9-cedcdce41e3d\") " pod="openstack-operators/ironic-operator-controller-manager-554564d7fc-8kpwf" Feb 27 08:05:57 crc kubenswrapper[4612]: I0227 08:05:57.954196 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gmvbz\" (UniqueName: \"kubernetes.io/projected/6065186a-5435-4db3-8adc-698bec4358ef-kube-api-access-gmvbz\") pod \"heat-operator-controller-manager-69f49c598c-7f9bj\" (UID: \"6065186a-5435-4db3-8adc-698bec4358ef\") " pod="openstack-operators/heat-operator-controller-manager-69f49c598c-7f9bj" Feb 27 08:05:57 crc kubenswrapper[4612]: I0227 08:05:57.954235 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/98296d45-b487-47c3-a44a-fa63ba52426b-cert\") pod \"infra-operator-controller-manager-79d975b745-plm9k\" (UID: \"98296d45-b487-47c3-a44a-fa63ba52426b\") " pod="openstack-operators/infra-operator-controller-manager-79d975b745-plm9k" Feb 27 08:05:57 crc kubenswrapper[4612]: E0227 08:05:57.954417 4612 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Feb 27 08:05:57 crc kubenswrapper[4612]: E0227 08:05:57.954480 4612 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/98296d45-b487-47c3-a44a-fa63ba52426b-cert podName:98296d45-b487-47c3-a44a-fa63ba52426b nodeName:}" failed. No retries permitted until 2026-02-27 08:05:58.454457611 +0000 UTC m=+1016.308387609 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/98296d45-b487-47c3-a44a-fa63ba52426b-cert") pod "infra-operator-controller-manager-79d975b745-plm9k" (UID: "98296d45-b487-47c3-a44a-fa63ba52426b") : secret "infra-operator-webhook-server-cert" not found Feb 27 08:05:57 crc kubenswrapper[4612]: I0227 08:05:57.959897 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-6bd4687957-2qmzb"] Feb 27 08:05:57 crc kubenswrapper[4612]: I0227 08:05:57.974933 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-567668f5cf-tjqwp"] Feb 27 08:05:57 crc kubenswrapper[4612]: I0227 08:05:57.975843 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-567668f5cf-tjqwp" Feb 27 08:05:57 crc kubenswrapper[4612]: I0227 08:05:57.994064 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-l7lcl" Feb 27 08:05:58 crc kubenswrapper[4612]: I0227 08:05:58.000308 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-567668f5cf-tjqwp"] Feb 27 08:05:58 crc kubenswrapper[4612]: I0227 08:05:58.014723 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-659dc6bbfc-vlczg"] Feb 27 08:05:58 crc kubenswrapper[4612]: I0227 08:05:58.017378 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-659dc6bbfc-vlczg" Feb 27 08:05:58 crc kubenswrapper[4612]: I0227 08:05:58.021376 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gmvbz\" (UniqueName: \"kubernetes.io/projected/6065186a-5435-4db3-8adc-698bec4358ef-kube-api-access-gmvbz\") pod \"heat-operator-controller-manager-69f49c598c-7f9bj\" (UID: \"6065186a-5435-4db3-8adc-698bec4358ef\") " pod="openstack-operators/heat-operator-controller-manager-69f49c598c-7f9bj" Feb 27 08:05:58 crc kubenswrapper[4612]: I0227 08:05:58.021827 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zcj9h\" (UniqueName: \"kubernetes.io/projected/98296d45-b487-47c3-a44a-fa63ba52426b-kube-api-access-zcj9h\") pod \"infra-operator-controller-manager-79d975b745-plm9k\" (UID: \"98296d45-b487-47c3-a44a-fa63ba52426b\") " pod="openstack-operators/infra-operator-controller-manager-79d975b745-plm9k" Feb 27 08:05:58 crc kubenswrapper[4612]: I0227 08:05:58.022083 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-hrqj2" Feb 27 08:05:58 crc kubenswrapper[4612]: I0227 08:05:58.036390 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tmdcx\" (UniqueName: \"kubernetes.io/projected/1903598c-2428-4e90-ab81-9a47fa5af9ef-kube-api-access-tmdcx\") pod \"horizon-operator-controller-manager-5b9b8895d5-t4rct\" (UID: \"1903598c-2428-4e90-ab81-9a47fa5af9ef\") " pod="openstack-operators/horizon-operator-controller-manager-5b9b8895d5-t4rct" Feb 27 08:05:58 crc kubenswrapper[4612]: I0227 08:05:58.042119 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-659dc6bbfc-vlczg"] Feb 27 08:05:58 crc kubenswrapper[4612]: I0227 08:05:58.059213 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lflq8\" (UniqueName: \"kubernetes.io/projected/259df4bd-af1b-4c61-a5bd-0ca28ac531d4-kube-api-access-lflq8\") pod \"keystone-operator-controller-manager-b4d948c87-2hm5l\" (UID: \"259df4bd-af1b-4c61-a5bd-0ca28ac531d4\") " pod="openstack-operators/keystone-operator-controller-manager-b4d948c87-2hm5l" Feb 27 08:05:58 crc kubenswrapper[4612]: I0227 08:05:58.059276 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8vwbr\" (UniqueName: \"kubernetes.io/projected/d84adbda-b53c-46c7-9196-682910a0009c-kube-api-access-8vwbr\") pod \"nova-operator-controller-manager-567668f5cf-tjqwp\" (UID: \"d84adbda-b53c-46c7-9196-682910a0009c\") " pod="openstack-operators/nova-operator-controller-manager-567668f5cf-tjqwp" Feb 27 08:05:58 crc kubenswrapper[4612]: I0227 08:05:58.059309 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hw2ws\" (UniqueName: \"kubernetes.io/projected/f62fea1f-71de-4872-9baa-42ad2448d5f5-kube-api-access-hw2ws\") pod \"mariadb-operator-controller-manager-6994f66f48-qs4l7\" (UID: \"f62fea1f-71de-4872-9baa-42ad2448d5f5\") " pod="openstack-operators/mariadb-operator-controller-manager-6994f66f48-qs4l7" Feb 27 08:05:58 crc kubenswrapper[4612]: I0227 08:05:58.059339 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j284h\" (UniqueName: \"kubernetes.io/projected/bd687bfa-c843-4d11-80f9-cedcdce41e3d-kube-api-access-j284h\") pod \"ironic-operator-controller-manager-554564d7fc-8kpwf\" (UID: \"bd687bfa-c843-4d11-80f9-cedcdce41e3d\") " pod="openstack-operators/ironic-operator-controller-manager-554564d7fc-8kpwf" Feb 27 08:05:58 crc kubenswrapper[4612]: I0227 08:05:58.059407 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2ltcz\" (UniqueName: \"kubernetes.io/projected/bd923861-d2ad-4a06-b0c7-613605afefbb-kube-api-access-2ltcz\") pod \"octavia-operator-controller-manager-659dc6bbfc-vlczg\" (UID: \"bd923861-d2ad-4a06-b0c7-613605afefbb\") " pod="openstack-operators/octavia-operator-controller-manager-659dc6bbfc-vlczg" Feb 27 08:05:58 crc kubenswrapper[4612]: I0227 08:05:58.059439 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fqgp9\" (UniqueName: \"kubernetes.io/projected/b60d4209-cd53-4cfe-99f7-7e7351041857-kube-api-access-fqgp9\") pod \"manila-operator-controller-manager-67d996989d-4dhhw\" (UID: \"b60d4209-cd53-4cfe-99f7-7e7351041857\") " pod="openstack-operators/manila-operator-controller-manager-67d996989d-4dhhw" Feb 27 08:05:58 crc kubenswrapper[4612]: I0227 08:05:58.059479 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ws5gl\" (UniqueName: \"kubernetes.io/projected/f292b2e2-2b32-4e4e-bfca-d5db60f0b358-kube-api-access-ws5gl\") pod \"neutron-operator-controller-manager-6bd4687957-2qmzb\" (UID: \"f292b2e2-2b32-4e4e-bfca-d5db60f0b358\") " pod="openstack-operators/neutron-operator-controller-manager-6bd4687957-2qmzb" Feb 27 08:05:58 crc kubenswrapper[4612]: I0227 08:05:58.064807 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-5955d8c787-q74ht"] Feb 27 08:05:58 crc kubenswrapper[4612]: I0227 08:05:58.065720 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-5955d8c787-q74ht" Feb 27 08:05:58 crc kubenswrapper[4612]: I0227 08:05:58.074254 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-v8wmf" Feb 27 08:05:58 crc kubenswrapper[4612]: I0227 08:05:58.077297 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-8497b45c89-dvq8j"] Feb 27 08:05:58 crc kubenswrapper[4612]: I0227 08:05:58.078151 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-8497b45c89-dvq8j" Feb 27 08:05:58 crc kubenswrapper[4612]: I0227 08:05:58.081582 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-pm4sp" Feb 27 08:05:58 crc kubenswrapper[4612]: I0227 08:05:58.087801 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9clp9dx"] Feb 27 08:05:58 crc kubenswrapper[4612]: I0227 08:05:58.088773 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9clp9dx" Feb 27 08:05:58 crc kubenswrapper[4612]: I0227 08:05:58.091812 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Feb 27 08:05:58 crc kubenswrapper[4612]: I0227 08:05:58.094630 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-nhbw6" Feb 27 08:05:58 crc kubenswrapper[4612]: I0227 08:05:58.095528 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j284h\" (UniqueName: \"kubernetes.io/projected/bd687bfa-c843-4d11-80f9-cedcdce41e3d-kube-api-access-j284h\") pod \"ironic-operator-controller-manager-554564d7fc-8kpwf\" (UID: \"bd687bfa-c843-4d11-80f9-cedcdce41e3d\") " pod="openstack-operators/ironic-operator-controller-manager-554564d7fc-8kpwf" Feb 27 08:05:58 crc kubenswrapper[4612]: I0227 08:05:58.107604 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fqgp9\" (UniqueName: \"kubernetes.io/projected/b60d4209-cd53-4cfe-99f7-7e7351041857-kube-api-access-fqgp9\") pod \"manila-operator-controller-manager-67d996989d-4dhhw\" (UID: \"b60d4209-cd53-4cfe-99f7-7e7351041857\") " pod="openstack-operators/manila-operator-controller-manager-67d996989d-4dhhw" Feb 27 08:05:58 crc kubenswrapper[4612]: I0227 08:05:58.113152 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-68f46476f-r8lqs"] Feb 27 08:05:58 crc kubenswrapper[4612]: I0227 08:05:58.118414 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hw2ws\" (UniqueName: \"kubernetes.io/projected/f62fea1f-71de-4872-9baa-42ad2448d5f5-kube-api-access-hw2ws\") pod \"mariadb-operator-controller-manager-6994f66f48-qs4l7\" (UID: \"f62fea1f-71de-4872-9baa-42ad2448d5f5\") " pod="openstack-operators/mariadb-operator-controller-manager-6994f66f48-qs4l7" Feb 27 08:05:58 crc kubenswrapper[4612]: I0227 08:05:58.127074 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-68f46476f-r8lqs" Feb 27 08:05:58 crc kubenswrapper[4612]: I0227 08:05:58.128093 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lflq8\" (UniqueName: \"kubernetes.io/projected/259df4bd-af1b-4c61-a5bd-0ca28ac531d4-kube-api-access-lflq8\") pod \"keystone-operator-controller-manager-b4d948c87-2hm5l\" (UID: \"259df4bd-af1b-4c61-a5bd-0ca28ac531d4\") " pod="openstack-operators/keystone-operator-controller-manager-b4d948c87-2hm5l" Feb 27 08:05:58 crc kubenswrapper[4612]: I0227 08:05:58.131261 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-5955d8c787-q74ht"] Feb 27 08:05:58 crc kubenswrapper[4612]: I0227 08:05:58.133083 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-t6ns9" Feb 27 08:05:58 crc kubenswrapper[4612]: I0227 08:05:58.137205 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-55d77d7b5c-462m7" Feb 27 08:05:58 crc kubenswrapper[4612]: I0227 08:05:58.138527 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-8497b45c89-dvq8j"] Feb 27 08:05:58 crc kubenswrapper[4612]: I0227 08:05:58.143712 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-b4d948c87-2hm5l" Feb 27 08:05:58 crc kubenswrapper[4612]: I0227 08:05:58.145647 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9clp9dx"] Feb 27 08:05:58 crc kubenswrapper[4612]: I0227 08:05:58.154065 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-589c568786-qclqp"] Feb 27 08:05:58 crc kubenswrapper[4612]: I0227 08:05:58.155860 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-589c568786-qclqp" Feb 27 08:05:58 crc kubenswrapper[4612]: I0227 08:05:58.168635 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-4jmgt" Feb 27 08:05:58 crc kubenswrapper[4612]: I0227 08:05:58.169386 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ws5gl\" (UniqueName: \"kubernetes.io/projected/f292b2e2-2b32-4e4e-bfca-d5db60f0b358-kube-api-access-ws5gl\") pod \"neutron-operator-controller-manager-6bd4687957-2qmzb\" (UID: \"f292b2e2-2b32-4e4e-bfca-d5db60f0b358\") " pod="openstack-operators/neutron-operator-controller-manager-6bd4687957-2qmzb" Feb 27 08:05:58 crc kubenswrapper[4612]: I0227 08:05:58.170920 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sfh76\" (UniqueName: \"kubernetes.io/projected/8414b96c-7e5b-4824-998e-549a0c43d9d5-kube-api-access-sfh76\") pod \"placement-operator-controller-manager-8497b45c89-dvq8j\" (UID: \"8414b96c-7e5b-4824-998e-549a0c43d9d5\") " pod="openstack-operators/placement-operator-controller-manager-8497b45c89-dvq8j" Feb 27 08:05:58 crc kubenswrapper[4612]: I0227 08:05:58.170955 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8vwbr\" (UniqueName: \"kubernetes.io/projected/d84adbda-b53c-46c7-9196-682910a0009c-kube-api-access-8vwbr\") pod \"nova-operator-controller-manager-567668f5cf-tjqwp\" (UID: \"d84adbda-b53c-46c7-9196-682910a0009c\") " pod="openstack-operators/nova-operator-controller-manager-567668f5cf-tjqwp" Feb 27 08:05:58 crc kubenswrapper[4612]: I0227 08:05:58.170988 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4wp6l\" (UniqueName: \"kubernetes.io/projected/e0304d11-114b-4aa3-86ca-1747c60b5bfc-kube-api-access-4wp6l\") pod \"ovn-operator-controller-manager-5955d8c787-q74ht\" (UID: \"e0304d11-114b-4aa3-86ca-1747c60b5bfc\") " pod="openstack-operators/ovn-operator-controller-manager-5955d8c787-q74ht" Feb 27 08:05:58 crc kubenswrapper[4612]: I0227 08:05:58.171008 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/cf5ef40d-5e5c-4b2c-9570-63e552800b4e-cert\") pod \"openstack-baremetal-operator-controller-manager-7c6767dc9clp9dx\" (UID: \"cf5ef40d-5e5c-4b2c-9570-63e552800b4e\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9clp9dx" Feb 27 08:05:58 crc kubenswrapper[4612]: I0227 08:05:58.171066 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2ltcz\" (UniqueName: \"kubernetes.io/projected/bd923861-d2ad-4a06-b0c7-613605afefbb-kube-api-access-2ltcz\") pod \"octavia-operator-controller-manager-659dc6bbfc-vlczg\" (UID: \"bd923861-d2ad-4a06-b0c7-613605afefbb\") " pod="openstack-operators/octavia-operator-controller-manager-659dc6bbfc-vlczg" Feb 27 08:05:58 crc kubenswrapper[4612]: I0227 08:05:58.171100 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-26lww\" (UniqueName: \"kubernetes.io/projected/cf5ef40d-5e5c-4b2c-9570-63e552800b4e-kube-api-access-26lww\") pod \"openstack-baremetal-operator-controller-manager-7c6767dc9clp9dx\" (UID: \"cf5ef40d-5e5c-4b2c-9570-63e552800b4e\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9clp9dx" Feb 27 08:05:58 crc kubenswrapper[4612]: I0227 08:05:58.199854 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-589c568786-qclqp"] Feb 27 08:05:58 crc kubenswrapper[4612]: I0227 08:05:58.218534 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ws5gl\" (UniqueName: \"kubernetes.io/projected/f292b2e2-2b32-4e4e-bfca-d5db60f0b358-kube-api-access-ws5gl\") pod \"neutron-operator-controller-manager-6bd4687957-2qmzb\" (UID: \"f292b2e2-2b32-4e4e-bfca-d5db60f0b358\") " pod="openstack-operators/neutron-operator-controller-manager-6bd4687957-2qmzb" Feb 27 08:05:58 crc kubenswrapper[4612]: I0227 08:05:58.232294 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2ltcz\" (UniqueName: \"kubernetes.io/projected/bd923861-d2ad-4a06-b0c7-613605afefbb-kube-api-access-2ltcz\") pod \"octavia-operator-controller-manager-659dc6bbfc-vlczg\" (UID: \"bd923861-d2ad-4a06-b0c7-613605afefbb\") " pod="openstack-operators/octavia-operator-controller-manager-659dc6bbfc-vlczg" Feb 27 08:05:58 crc kubenswrapper[4612]: I0227 08:05:58.236636 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-68f46476f-r8lqs"] Feb 27 08:05:58 crc kubenswrapper[4612]: I0227 08:05:58.238260 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8vwbr\" (UniqueName: \"kubernetes.io/projected/d84adbda-b53c-46c7-9196-682910a0009c-kube-api-access-8vwbr\") pod \"nova-operator-controller-manager-567668f5cf-tjqwp\" (UID: \"d84adbda-b53c-46c7-9196-682910a0009c\") " pod="openstack-operators/nova-operator-controller-manager-567668f5cf-tjqwp" Feb 27 08:05:58 crc kubenswrapper[4612]: I0227 08:05:58.243780 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-69f49c598c-7f9bj" Feb 27 08:05:58 crc kubenswrapper[4612]: I0227 08:05:58.245642 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-554564d7fc-8kpwf" Feb 27 08:05:58 crc kubenswrapper[4612]: I0227 08:05:58.252865 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-5dc6794d5b-s2jnh"] Feb 27 08:05:58 crc kubenswrapper[4612]: I0227 08:05:58.253786 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5dc6794d5b-s2jnh" Feb 27 08:05:58 crc kubenswrapper[4612]: I0227 08:05:58.263434 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-q9lt9" Feb 27 08:05:58 crc kubenswrapper[4612]: I0227 08:05:58.265470 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-5b9b8895d5-t4rct" Feb 27 08:05:58 crc kubenswrapper[4612]: I0227 08:05:58.284979 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5dc6794d5b-s2jnh"] Feb 27 08:05:58 crc kubenswrapper[4612]: I0227 08:05:58.305224 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-26lww\" (UniqueName: \"kubernetes.io/projected/cf5ef40d-5e5c-4b2c-9570-63e552800b4e-kube-api-access-26lww\") pod \"openstack-baremetal-operator-controller-manager-7c6767dc9clp9dx\" (UID: \"cf5ef40d-5e5c-4b2c-9570-63e552800b4e\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9clp9dx" Feb 27 08:05:58 crc kubenswrapper[4612]: I0227 08:05:58.305556 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k5p4x\" (UniqueName: \"kubernetes.io/projected/713a994e-490d-47b4-86ab-1203a6b806fc-kube-api-access-k5p4x\") pod \"swift-operator-controller-manager-68f46476f-r8lqs\" (UID: \"713a994e-490d-47b4-86ab-1203a6b806fc\") " pod="openstack-operators/swift-operator-controller-manager-68f46476f-r8lqs" Feb 27 08:05:58 crc kubenswrapper[4612]: I0227 08:05:58.305667 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sfh76\" (UniqueName: \"kubernetes.io/projected/8414b96c-7e5b-4824-998e-549a0c43d9d5-kube-api-access-sfh76\") pod \"placement-operator-controller-manager-8497b45c89-dvq8j\" (UID: \"8414b96c-7e5b-4824-998e-549a0c43d9d5\") " pod="openstack-operators/placement-operator-controller-manager-8497b45c89-dvq8j" Feb 27 08:05:58 crc kubenswrapper[4612]: I0227 08:05:58.305766 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4wp6l\" (UniqueName: \"kubernetes.io/projected/e0304d11-114b-4aa3-86ca-1747c60b5bfc-kube-api-access-4wp6l\") pod \"ovn-operator-controller-manager-5955d8c787-q74ht\" (UID: \"e0304d11-114b-4aa3-86ca-1747c60b5bfc\") " pod="openstack-operators/ovn-operator-controller-manager-5955d8c787-q74ht" Feb 27 08:05:58 crc kubenswrapper[4612]: I0227 08:05:58.305842 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2rk2j\" (UniqueName: \"kubernetes.io/projected/ee3f21f6-2f91-45c6-8c80-5fb8f89a41ce-kube-api-access-2rk2j\") pod \"telemetry-operator-controller-manager-589c568786-qclqp\" (UID: \"ee3f21f6-2f91-45c6-8c80-5fb8f89a41ce\") " pod="openstack-operators/telemetry-operator-controller-manager-589c568786-qclqp" Feb 27 08:05:58 crc kubenswrapper[4612]: I0227 08:05:58.305925 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/cf5ef40d-5e5c-4b2c-9570-63e552800b4e-cert\") pod \"openstack-baremetal-operator-controller-manager-7c6767dc9clp9dx\" (UID: \"cf5ef40d-5e5c-4b2c-9570-63e552800b4e\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9clp9dx" Feb 27 08:05:58 crc kubenswrapper[4612]: E0227 08:05:58.306187 4612 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 27 08:05:58 crc kubenswrapper[4612]: E0227 08:05:58.308886 4612 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/cf5ef40d-5e5c-4b2c-9570-63e552800b4e-cert podName:cf5ef40d-5e5c-4b2c-9570-63e552800b4e nodeName:}" failed. No retries permitted until 2026-02-27 08:05:58.808865046 +0000 UTC m=+1016.662795044 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/cf5ef40d-5e5c-4b2c-9570-63e552800b4e-cert") pod "openstack-baremetal-operator-controller-manager-7c6767dc9clp9dx" (UID: "cf5ef40d-5e5c-4b2c-9570-63e552800b4e") : secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 27 08:05:58 crc kubenswrapper[4612]: I0227 08:05:58.331468 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-67d996989d-4dhhw" Feb 27 08:05:58 crc kubenswrapper[4612]: I0227 08:05:58.342634 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-bccc79885-zdtzv"] Feb 27 08:05:58 crc kubenswrapper[4612]: I0227 08:05:58.344862 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-bccc79885-zdtzv" Feb 27 08:05:58 crc kubenswrapper[4612]: I0227 08:05:58.345382 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-6994f66f48-qs4l7" Feb 27 08:05:58 crc kubenswrapper[4612]: I0227 08:05:58.351353 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-mfrtk" Feb 27 08:05:58 crc kubenswrapper[4612]: I0227 08:05:58.366122 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-bccc79885-zdtzv"] Feb 27 08:05:58 crc kubenswrapper[4612]: I0227 08:05:58.366929 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sfh76\" (UniqueName: \"kubernetes.io/projected/8414b96c-7e5b-4824-998e-549a0c43d9d5-kube-api-access-sfh76\") pod \"placement-operator-controller-manager-8497b45c89-dvq8j\" (UID: \"8414b96c-7e5b-4824-998e-549a0c43d9d5\") " pod="openstack-operators/placement-operator-controller-manager-8497b45c89-dvq8j" Feb 27 08:05:58 crc kubenswrapper[4612]: I0227 08:05:58.373715 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-6bd4687957-2qmzb" Feb 27 08:05:58 crc kubenswrapper[4612]: I0227 08:05:58.397685 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-567668f5cf-tjqwp" Feb 27 08:05:58 crc kubenswrapper[4612]: I0227 08:05:58.403212 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-659dc6bbfc-vlczg" Feb 27 08:05:58 crc kubenswrapper[4612]: I0227 08:05:58.407736 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hnztn\" (UniqueName: \"kubernetes.io/projected/ad4fca59-6a87-4ebc-8681-d5a3a40f0b82-kube-api-access-hnztn\") pod \"test-operator-controller-manager-5dc6794d5b-s2jnh\" (UID: \"ad4fca59-6a87-4ebc-8681-d5a3a40f0b82\") " pod="openstack-operators/test-operator-controller-manager-5dc6794d5b-s2jnh" Feb 27 08:05:58 crc kubenswrapper[4612]: I0227 08:05:58.407961 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2rk2j\" (UniqueName: \"kubernetes.io/projected/ee3f21f6-2f91-45c6-8c80-5fb8f89a41ce-kube-api-access-2rk2j\") pod \"telemetry-operator-controller-manager-589c568786-qclqp\" (UID: \"ee3f21f6-2f91-45c6-8c80-5fb8f89a41ce\") " pod="openstack-operators/telemetry-operator-controller-manager-589c568786-qclqp" Feb 27 08:05:58 crc kubenswrapper[4612]: I0227 08:05:58.408114 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k5p4x\" (UniqueName: \"kubernetes.io/projected/713a994e-490d-47b4-86ab-1203a6b806fc-kube-api-access-k5p4x\") pod \"swift-operator-controller-manager-68f46476f-r8lqs\" (UID: \"713a994e-490d-47b4-86ab-1203a6b806fc\") " pod="openstack-operators/swift-operator-controller-manager-68f46476f-r8lqs" Feb 27 08:05:58 crc kubenswrapper[4612]: I0227 08:05:58.418798 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-26lww\" (UniqueName: \"kubernetes.io/projected/cf5ef40d-5e5c-4b2c-9570-63e552800b4e-kube-api-access-26lww\") pod \"openstack-baremetal-operator-controller-manager-7c6767dc9clp9dx\" (UID: \"cf5ef40d-5e5c-4b2c-9570-63e552800b4e\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9clp9dx" Feb 27 08:05:58 crc kubenswrapper[4612]: I0227 08:05:58.422382 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4wp6l\" (UniqueName: \"kubernetes.io/projected/e0304d11-114b-4aa3-86ca-1747c60b5bfc-kube-api-access-4wp6l\") pod \"ovn-operator-controller-manager-5955d8c787-q74ht\" (UID: \"e0304d11-114b-4aa3-86ca-1747c60b5bfc\") " pod="openstack-operators/ovn-operator-controller-manager-5955d8c787-q74ht" Feb 27 08:05:58 crc kubenswrapper[4612]: I0227 08:05:58.433204 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-5955d8c787-q74ht" Feb 27 08:05:58 crc kubenswrapper[4612]: I0227 08:05:58.444510 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2rk2j\" (UniqueName: \"kubernetes.io/projected/ee3f21f6-2f91-45c6-8c80-5fb8f89a41ce-kube-api-access-2rk2j\") pod \"telemetry-operator-controller-manager-589c568786-qclqp\" (UID: \"ee3f21f6-2f91-45c6-8c80-5fb8f89a41ce\") " pod="openstack-operators/telemetry-operator-controller-manager-589c568786-qclqp" Feb 27 08:05:58 crc kubenswrapper[4612]: I0227 08:05:58.446156 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-69f9884c8f-v4kkz"] Feb 27 08:05:58 crc kubenswrapper[4612]: I0227 08:05:58.447312 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-589c568786-qclqp" Feb 27 08:05:58 crc kubenswrapper[4612]: I0227 08:05:58.460406 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k5p4x\" (UniqueName: \"kubernetes.io/projected/713a994e-490d-47b4-86ab-1203a6b806fc-kube-api-access-k5p4x\") pod \"swift-operator-controller-manager-68f46476f-r8lqs\" (UID: \"713a994e-490d-47b4-86ab-1203a6b806fc\") " pod="openstack-operators/swift-operator-controller-manager-68f46476f-r8lqs" Feb 27 08:05:58 crc kubenswrapper[4612]: I0227 08:05:58.462494 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-69f9884c8f-v4kkz"] Feb 27 08:05:58 crc kubenswrapper[4612]: I0227 08:05:58.462601 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-69f9884c8f-v4kkz" Feb 27 08:05:58 crc kubenswrapper[4612]: I0227 08:05:58.468487 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Feb 27 08:05:58 crc kubenswrapper[4612]: I0227 08:05:58.468869 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-kmvsq" Feb 27 08:05:58 crc kubenswrapper[4612]: I0227 08:05:58.469071 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"metrics-server-cert" Feb 27 08:05:58 crc kubenswrapper[4612]: I0227 08:05:58.503342 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-8497b45c89-dvq8j" Feb 27 08:05:58 crc kubenswrapper[4612]: I0227 08:05:58.507892 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-8s6n2"] Feb 27 08:05:58 crc kubenswrapper[4612]: I0227 08:05:58.508682 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-8s6n2" Feb 27 08:05:58 crc kubenswrapper[4612]: I0227 08:05:58.510950 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hnztn\" (UniqueName: \"kubernetes.io/projected/ad4fca59-6a87-4ebc-8681-d5a3a40f0b82-kube-api-access-hnztn\") pod \"test-operator-controller-manager-5dc6794d5b-s2jnh\" (UID: \"ad4fca59-6a87-4ebc-8681-d5a3a40f0b82\") " pod="openstack-operators/test-operator-controller-manager-5dc6794d5b-s2jnh" Feb 27 08:05:58 crc kubenswrapper[4612]: I0227 08:05:58.511019 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wn2qx\" (UniqueName: \"kubernetes.io/projected/a4619f7b-7242-459b-8558-f84c2893ee55-kube-api-access-wn2qx\") pod \"watcher-operator-controller-manager-bccc79885-zdtzv\" (UID: \"a4619f7b-7242-459b-8558-f84c2893ee55\") " pod="openstack-operators/watcher-operator-controller-manager-bccc79885-zdtzv" Feb 27 08:05:58 crc kubenswrapper[4612]: I0227 08:05:58.511083 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/98296d45-b487-47c3-a44a-fa63ba52426b-cert\") pod \"infra-operator-controller-manager-79d975b745-plm9k\" (UID: \"98296d45-b487-47c3-a44a-fa63ba52426b\") " pod="openstack-operators/infra-operator-controller-manager-79d975b745-plm9k" Feb 27 08:05:58 crc kubenswrapper[4612]: E0227 08:05:58.512586 4612 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Feb 27 08:05:58 crc kubenswrapper[4612]: E0227 08:05:58.512721 4612 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/98296d45-b487-47c3-a44a-fa63ba52426b-cert podName:98296d45-b487-47c3-a44a-fa63ba52426b nodeName:}" failed. No retries permitted until 2026-02-27 08:05:59.512685991 +0000 UTC m=+1017.366615989 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/98296d45-b487-47c3-a44a-fa63ba52426b-cert") pod "infra-operator-controller-manager-79d975b745-plm9k" (UID: "98296d45-b487-47c3-a44a-fa63ba52426b") : secret "infra-operator-webhook-server-cert" not found Feb 27 08:05:58 crc kubenswrapper[4612]: I0227 08:05:58.526081 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-bt5ln" Feb 27 08:05:58 crc kubenswrapper[4612]: I0227 08:05:58.573764 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-8s6n2"] Feb 27 08:05:58 crc kubenswrapper[4612]: I0227 08:05:58.585199 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hnztn\" (UniqueName: \"kubernetes.io/projected/ad4fca59-6a87-4ebc-8681-d5a3a40f0b82-kube-api-access-hnztn\") pod \"test-operator-controller-manager-5dc6794d5b-s2jnh\" (UID: \"ad4fca59-6a87-4ebc-8681-d5a3a40f0b82\") " pod="openstack-operators/test-operator-controller-manager-5dc6794d5b-s2jnh" Feb 27 08:05:58 crc kubenswrapper[4612]: I0227 08:05:58.620254 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/41aed211-3a1f-4f7c-86ca-53e445dcd585-webhook-certs\") pod \"openstack-operator-controller-manager-69f9884c8f-v4kkz\" (UID: \"41aed211-3a1f-4f7c-86ca-53e445dcd585\") " pod="openstack-operators/openstack-operator-controller-manager-69f9884c8f-v4kkz" Feb 27 08:05:58 crc kubenswrapper[4612]: I0227 08:05:58.620326 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wn2qx\" (UniqueName: \"kubernetes.io/projected/a4619f7b-7242-459b-8558-f84c2893ee55-kube-api-access-wn2qx\") pod \"watcher-operator-controller-manager-bccc79885-zdtzv\" (UID: \"a4619f7b-7242-459b-8558-f84c2893ee55\") " pod="openstack-operators/watcher-operator-controller-manager-bccc79885-zdtzv" Feb 27 08:05:58 crc kubenswrapper[4612]: I0227 08:05:58.620391 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9qr52\" (UniqueName: \"kubernetes.io/projected/41aed211-3a1f-4f7c-86ca-53e445dcd585-kube-api-access-9qr52\") pod \"openstack-operator-controller-manager-69f9884c8f-v4kkz\" (UID: \"41aed211-3a1f-4f7c-86ca-53e445dcd585\") " pod="openstack-operators/openstack-operator-controller-manager-69f9884c8f-v4kkz" Feb 27 08:05:58 crc kubenswrapper[4612]: I0227 08:05:58.620408 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/41aed211-3a1f-4f7c-86ca-53e445dcd585-metrics-certs\") pod \"openstack-operator-controller-manager-69f9884c8f-v4kkz\" (UID: \"41aed211-3a1f-4f7c-86ca-53e445dcd585\") " pod="openstack-operators/openstack-operator-controller-manager-69f9884c8f-v4kkz" Feb 27 08:05:58 crc kubenswrapper[4612]: I0227 08:05:58.620452 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5684q\" (UniqueName: \"kubernetes.io/projected/57ba67ae-3e41-4e2b-9ee6-30b2643805dd-kube-api-access-5684q\") pod \"rabbitmq-cluster-operator-manager-668c99d594-8s6n2\" (UID: \"57ba67ae-3e41-4e2b-9ee6-30b2643805dd\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-8s6n2" Feb 27 08:05:58 crc kubenswrapper[4612]: I0227 08:05:58.667183 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wn2qx\" (UniqueName: \"kubernetes.io/projected/a4619f7b-7242-459b-8558-f84c2893ee55-kube-api-access-wn2qx\") pod \"watcher-operator-controller-manager-bccc79885-zdtzv\" (UID: \"a4619f7b-7242-459b-8558-f84c2893ee55\") " pod="openstack-operators/watcher-operator-controller-manager-bccc79885-zdtzv" Feb 27 08:05:58 crc kubenswrapper[4612]: I0227 08:05:58.713561 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-68f46476f-r8lqs" Feb 27 08:05:58 crc kubenswrapper[4612]: I0227 08:05:58.721556 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5684q\" (UniqueName: \"kubernetes.io/projected/57ba67ae-3e41-4e2b-9ee6-30b2643805dd-kube-api-access-5684q\") pod \"rabbitmq-cluster-operator-manager-668c99d594-8s6n2\" (UID: \"57ba67ae-3e41-4e2b-9ee6-30b2643805dd\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-8s6n2" Feb 27 08:05:58 crc kubenswrapper[4612]: I0227 08:05:58.721626 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/41aed211-3a1f-4f7c-86ca-53e445dcd585-webhook-certs\") pod \"openstack-operator-controller-manager-69f9884c8f-v4kkz\" (UID: \"41aed211-3a1f-4f7c-86ca-53e445dcd585\") " pod="openstack-operators/openstack-operator-controller-manager-69f9884c8f-v4kkz" Feb 27 08:05:58 crc kubenswrapper[4612]: I0227 08:05:58.721683 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9qr52\" (UniqueName: \"kubernetes.io/projected/41aed211-3a1f-4f7c-86ca-53e445dcd585-kube-api-access-9qr52\") pod \"openstack-operator-controller-manager-69f9884c8f-v4kkz\" (UID: \"41aed211-3a1f-4f7c-86ca-53e445dcd585\") " pod="openstack-operators/openstack-operator-controller-manager-69f9884c8f-v4kkz" Feb 27 08:05:58 crc kubenswrapper[4612]: I0227 08:05:58.721736 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/41aed211-3a1f-4f7c-86ca-53e445dcd585-metrics-certs\") pod \"openstack-operator-controller-manager-69f9884c8f-v4kkz\" (UID: \"41aed211-3a1f-4f7c-86ca-53e445dcd585\") " pod="openstack-operators/openstack-operator-controller-manager-69f9884c8f-v4kkz" Feb 27 08:05:58 crc kubenswrapper[4612]: E0227 08:05:58.721839 4612 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Feb 27 08:05:58 crc kubenswrapper[4612]: E0227 08:05:58.722440 4612 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/41aed211-3a1f-4f7c-86ca-53e445dcd585-metrics-certs podName:41aed211-3a1f-4f7c-86ca-53e445dcd585 nodeName:}" failed. No retries permitted until 2026-02-27 08:05:59.222425115 +0000 UTC m=+1017.076355103 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/41aed211-3a1f-4f7c-86ca-53e445dcd585-metrics-certs") pod "openstack-operator-controller-manager-69f9884c8f-v4kkz" (UID: "41aed211-3a1f-4f7c-86ca-53e445dcd585") : secret "metrics-server-cert" not found Feb 27 08:05:58 crc kubenswrapper[4612]: E0227 08:05:58.722834 4612 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Feb 27 08:05:58 crc kubenswrapper[4612]: E0227 08:05:58.722858 4612 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/41aed211-3a1f-4f7c-86ca-53e445dcd585-webhook-certs podName:41aed211-3a1f-4f7c-86ca-53e445dcd585 nodeName:}" failed. No retries permitted until 2026-02-27 08:05:59.222851357 +0000 UTC m=+1017.076781345 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/41aed211-3a1f-4f7c-86ca-53e445dcd585-webhook-certs") pod "openstack-operator-controller-manager-69f9884c8f-v4kkz" (UID: "41aed211-3a1f-4f7c-86ca-53e445dcd585") : secret "webhook-server-cert" not found Feb 27 08:05:58 crc kubenswrapper[4612]: I0227 08:05:58.750629 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9qr52\" (UniqueName: \"kubernetes.io/projected/41aed211-3a1f-4f7c-86ca-53e445dcd585-kube-api-access-9qr52\") pod \"openstack-operator-controller-manager-69f9884c8f-v4kkz\" (UID: \"41aed211-3a1f-4f7c-86ca-53e445dcd585\") " pod="openstack-operators/openstack-operator-controller-manager-69f9884c8f-v4kkz" Feb 27 08:05:58 crc kubenswrapper[4612]: I0227 08:05:58.752512 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5684q\" (UniqueName: \"kubernetes.io/projected/57ba67ae-3e41-4e2b-9ee6-30b2643805dd-kube-api-access-5684q\") pod \"rabbitmq-cluster-operator-manager-668c99d594-8s6n2\" (UID: \"57ba67ae-3e41-4e2b-9ee6-30b2643805dd\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-8s6n2" Feb 27 08:05:58 crc kubenswrapper[4612]: I0227 08:05:58.778521 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5dc6794d5b-s2jnh" Feb 27 08:05:58 crc kubenswrapper[4612]: I0227 08:05:58.821061 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-bccc79885-zdtzv" Feb 27 08:05:58 crc kubenswrapper[4612]: I0227 08:05:58.823501 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/cf5ef40d-5e5c-4b2c-9570-63e552800b4e-cert\") pod \"openstack-baremetal-operator-controller-manager-7c6767dc9clp9dx\" (UID: \"cf5ef40d-5e5c-4b2c-9570-63e552800b4e\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9clp9dx" Feb 27 08:05:58 crc kubenswrapper[4612]: E0227 08:05:58.824625 4612 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 27 08:05:58 crc kubenswrapper[4612]: E0227 08:05:58.824709 4612 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/cf5ef40d-5e5c-4b2c-9570-63e552800b4e-cert podName:cf5ef40d-5e5c-4b2c-9570-63e552800b4e nodeName:}" failed. No retries permitted until 2026-02-27 08:05:59.824675992 +0000 UTC m=+1017.678605990 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/cf5ef40d-5e5c-4b2c-9570-63e552800b4e-cert") pod "openstack-baremetal-operator-controller-manager-7c6767dc9clp9dx" (UID: "cf5ef40d-5e5c-4b2c-9570-63e552800b4e") : secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 27 08:05:58 crc kubenswrapper[4612]: I0227 08:05:58.924632 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-8s6n2" Feb 27 08:05:58 crc kubenswrapper[4612]: I0227 08:05:58.929832 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-868647ff47-mlgz7"] Feb 27 08:05:59 crc kubenswrapper[4612]: I0227 08:05:59.037809 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-784b5bb6c5-g4p5v"] Feb 27 08:05:59 crc kubenswrapper[4612]: I0227 08:05:59.238738 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/41aed211-3a1f-4f7c-86ca-53e445dcd585-metrics-certs\") pod \"openstack-operator-controller-manager-69f9884c8f-v4kkz\" (UID: \"41aed211-3a1f-4f7c-86ca-53e445dcd585\") " pod="openstack-operators/openstack-operator-controller-manager-69f9884c8f-v4kkz" Feb 27 08:05:59 crc kubenswrapper[4612]: I0227 08:05:59.238882 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/41aed211-3a1f-4f7c-86ca-53e445dcd585-webhook-certs\") pod \"openstack-operator-controller-manager-69f9884c8f-v4kkz\" (UID: \"41aed211-3a1f-4f7c-86ca-53e445dcd585\") " pod="openstack-operators/openstack-operator-controller-manager-69f9884c8f-v4kkz" Feb 27 08:05:59 crc kubenswrapper[4612]: E0227 08:05:59.239094 4612 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Feb 27 08:05:59 crc kubenswrapper[4612]: E0227 08:05:59.239154 4612 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/41aed211-3a1f-4f7c-86ca-53e445dcd585-webhook-certs podName:41aed211-3a1f-4f7c-86ca-53e445dcd585 nodeName:}" failed. No retries permitted until 2026-02-27 08:06:00.239135387 +0000 UTC m=+1018.093065385 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/41aed211-3a1f-4f7c-86ca-53e445dcd585-webhook-certs") pod "openstack-operator-controller-manager-69f9884c8f-v4kkz" (UID: "41aed211-3a1f-4f7c-86ca-53e445dcd585") : secret "webhook-server-cert" not found Feb 27 08:05:59 crc kubenswrapper[4612]: E0227 08:05:59.239548 4612 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Feb 27 08:05:59 crc kubenswrapper[4612]: E0227 08:05:59.239583 4612 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/41aed211-3a1f-4f7c-86ca-53e445dcd585-metrics-certs podName:41aed211-3a1f-4f7c-86ca-53e445dcd585 nodeName:}" failed. No retries permitted until 2026-02-27 08:06:00.239571859 +0000 UTC m=+1018.093501857 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/41aed211-3a1f-4f7c-86ca-53e445dcd585-metrics-certs") pod "openstack-operator-controller-manager-69f9884c8f-v4kkz" (UID: "41aed211-3a1f-4f7c-86ca-53e445dcd585") : secret "metrics-server-cert" not found Feb 27 08:05:59 crc kubenswrapper[4612]: I0227 08:05:59.299305 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-6d8bf5c495-t95hp"] Feb 27 08:05:59 crc kubenswrapper[4612]: I0227 08:05:59.452199 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-554564d7fc-8kpwf"] Feb 27 08:05:59 crc kubenswrapper[4612]: W0227 08:05:59.457821 4612 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbd687bfa_c843_4d11_80f9_cedcdce41e3d.slice/crio-979dbd2d00521d21fbf2d12c01038e7a7e9e4b5efa76476053bb0ee09ef51983 WatchSource:0}: Error finding container 979dbd2d00521d21fbf2d12c01038e7a7e9e4b5efa76476053bb0ee09ef51983: Status 404 returned error can't find the container with id 979dbd2d00521d21fbf2d12c01038e7a7e9e4b5efa76476053bb0ee09ef51983 Feb 27 08:05:59 crc kubenswrapper[4612]: I0227 08:05:59.476407 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-67d996989d-4dhhw"] Feb 27 08:05:59 crc kubenswrapper[4612]: I0227 08:05:59.493517 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-69f49c598c-7f9bj"] Feb 27 08:05:59 crc kubenswrapper[4612]: W0227 08:05:59.508342 4612 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6065186a_5435_4db3_8adc_698bec4358ef.slice/crio-9e84f3f2588a6b77bcfcd803fed5fbeb363e672cdad67948887dad90000b39b9 WatchSource:0}: Error finding container 9e84f3f2588a6b77bcfcd803fed5fbeb363e672cdad67948887dad90000b39b9: Status 404 returned error can't find the container with id 9e84f3f2588a6b77bcfcd803fed5fbeb363e672cdad67948887dad90000b39b9 Feb 27 08:05:59 crc kubenswrapper[4612]: I0227 08:05:59.513167 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-b4d948c87-2hm5l"] Feb 27 08:05:59 crc kubenswrapper[4612]: W0227 08:05:59.513274 4612 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod259df4bd_af1b_4c61_a5bd_0ca28ac531d4.slice/crio-b7723488fb18b53b3538afa34e41f065c87a0a0f4a45fe0c605d3cb1aecdb27a WatchSource:0}: Error finding container b7723488fb18b53b3538afa34e41f065c87a0a0f4a45fe0c605d3cb1aecdb27a: Status 404 returned error can't find the container with id b7723488fb18b53b3538afa34e41f065c87a0a0f4a45fe0c605d3cb1aecdb27a Feb 27 08:05:59 crc kubenswrapper[4612]: I0227 08:05:59.543155 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/98296d45-b487-47c3-a44a-fa63ba52426b-cert\") pod \"infra-operator-controller-manager-79d975b745-plm9k\" (UID: \"98296d45-b487-47c3-a44a-fa63ba52426b\") " pod="openstack-operators/infra-operator-controller-manager-79d975b745-plm9k" Feb 27 08:05:59 crc kubenswrapper[4612]: E0227 08:05:59.543367 4612 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Feb 27 08:05:59 crc kubenswrapper[4612]: E0227 08:05:59.543456 4612 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/98296d45-b487-47c3-a44a-fa63ba52426b-cert podName:98296d45-b487-47c3-a44a-fa63ba52426b nodeName:}" failed. No retries permitted until 2026-02-27 08:06:01.543423427 +0000 UTC m=+1019.397353425 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/98296d45-b487-47c3-a44a-fa63ba52426b-cert") pod "infra-operator-controller-manager-79d975b745-plm9k" (UID: "98296d45-b487-47c3-a44a-fa63ba52426b") : secret "infra-operator-webhook-server-cert" not found Feb 27 08:05:59 crc kubenswrapper[4612]: I0227 08:05:59.666462 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-8497b45c89-dvq8j"] Feb 27 08:05:59 crc kubenswrapper[4612]: I0227 08:05:59.670766 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-5b9b8895d5-t4rct"] Feb 27 08:05:59 crc kubenswrapper[4612]: W0227 08:05:59.673454 4612 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1903598c_2428_4e90_ab81_9a47fa5af9ef.slice/crio-b8d649536abf883514a146fdf3f37b6a7e5f54047a0821bad83866c456814be7 WatchSource:0}: Error finding container b8d649536abf883514a146fdf3f37b6a7e5f54047a0821bad83866c456814be7: Status 404 returned error can't find the container with id b8d649536abf883514a146fdf3f37b6a7e5f54047a0821bad83866c456814be7 Feb 27 08:05:59 crc kubenswrapper[4612]: I0227 08:05:59.725579 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-8497b45c89-dvq8j" event={"ID":"8414b96c-7e5b-4824-998e-549a0c43d9d5","Type":"ContainerStarted","Data":"b0741f6433c5441570e5ece68ac74c078812934364371796daaf5976e72c09be"} Feb 27 08:05:59 crc kubenswrapper[4612]: I0227 08:05:59.728547 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-b4d948c87-2hm5l" event={"ID":"259df4bd-af1b-4c61-a5bd-0ca28ac531d4","Type":"ContainerStarted","Data":"b7723488fb18b53b3538afa34e41f065c87a0a0f4a45fe0c605d3cb1aecdb27a"} Feb 27 08:05:59 crc kubenswrapper[4612]: I0227 08:05:59.731065 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-5b9b8895d5-t4rct" event={"ID":"1903598c-2428-4e90-ab81-9a47fa5af9ef","Type":"ContainerStarted","Data":"b8d649536abf883514a146fdf3f37b6a7e5f54047a0821bad83866c456814be7"} Feb 27 08:05:59 crc kubenswrapper[4612]: I0227 08:05:59.736195 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-67d996989d-4dhhw" event={"ID":"b60d4209-cd53-4cfe-99f7-7e7351041857","Type":"ContainerStarted","Data":"ca5e874efdde593847c03de21ef3062c7d9e749d67fa1680a760a57a51cb62c7"} Feb 27 08:05:59 crc kubenswrapper[4612]: I0227 08:05:59.738173 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-69f49c598c-7f9bj" event={"ID":"6065186a-5435-4db3-8adc-698bec4358ef","Type":"ContainerStarted","Data":"9e84f3f2588a6b77bcfcd803fed5fbeb363e672cdad67948887dad90000b39b9"} Feb 27 08:05:59 crc kubenswrapper[4612]: I0227 08:05:59.740793 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-554564d7fc-8kpwf" event={"ID":"bd687bfa-c843-4d11-80f9-cedcdce41e3d","Type":"ContainerStarted","Data":"979dbd2d00521d21fbf2d12c01038e7a7e9e4b5efa76476053bb0ee09ef51983"} Feb 27 08:05:59 crc kubenswrapper[4612]: I0227 08:05:59.742910 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-784b5bb6c5-g4p5v" event={"ID":"71c482bc-7818-4193-a80b-323682cdcfb3","Type":"ContainerStarted","Data":"56c8f765352d48e611907981d60805a3c8c89af0658d22da39eda3470eb22dd2"} Feb 27 08:05:59 crc kubenswrapper[4612]: I0227 08:05:59.745854 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-6d8bf5c495-t95hp" event={"ID":"63d519c2-d2f9-4349-ad82-0399e472d78a","Type":"ContainerStarted","Data":"32a42d410b81a6658985119ed73dd74c7433f3330000390b32fcbae7799b762a"} Feb 27 08:05:59 crc kubenswrapper[4612]: I0227 08:05:59.748505 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-868647ff47-mlgz7" event={"ID":"2ddc15ea-b7c1-44e7-b795-392890e9c14e","Type":"ContainerStarted","Data":"acae751d7a8a577509a92c7d06e01c45ed19d3ffc2d4e71d72f8ab2df8c48967"} Feb 27 08:05:59 crc kubenswrapper[4612]: I0227 08:05:59.846086 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/cf5ef40d-5e5c-4b2c-9570-63e552800b4e-cert\") pod \"openstack-baremetal-operator-controller-manager-7c6767dc9clp9dx\" (UID: \"cf5ef40d-5e5c-4b2c-9570-63e552800b4e\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9clp9dx" Feb 27 08:05:59 crc kubenswrapper[4612]: E0227 08:05:59.846272 4612 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 27 08:05:59 crc kubenswrapper[4612]: E0227 08:05:59.846320 4612 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/cf5ef40d-5e5c-4b2c-9570-63e552800b4e-cert podName:cf5ef40d-5e5c-4b2c-9570-63e552800b4e nodeName:}" failed. No retries permitted until 2026-02-27 08:06:01.846305278 +0000 UTC m=+1019.700235276 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/cf5ef40d-5e5c-4b2c-9570-63e552800b4e-cert") pod "openstack-baremetal-operator-controller-manager-7c6767dc9clp9dx" (UID: "cf5ef40d-5e5c-4b2c-9570-63e552800b4e") : secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 27 08:05:59 crc kubenswrapper[4612]: I0227 08:05:59.860560 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-659dc6bbfc-vlczg"] Feb 27 08:05:59 crc kubenswrapper[4612]: I0227 08:05:59.870454 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-bccc79885-zdtzv"] Feb 27 08:05:59 crc kubenswrapper[4612]: W0227 08:05:59.875510 4612 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda4619f7b_7242_459b_8558_f84c2893ee55.slice/crio-f4e9e365f26c17e1f922a514a09c61c5cb034fbf6d08e45ba8f18c3a3b335d49 WatchSource:0}: Error finding container f4e9e365f26c17e1f922a514a09c61c5cb034fbf6d08e45ba8f18c3a3b335d49: Status 404 returned error can't find the container with id f4e9e365f26c17e1f922a514a09c61c5cb034fbf6d08e45ba8f18c3a3b335d49 Feb 27 08:05:59 crc kubenswrapper[4612]: I0227 08:05:59.893836 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-589c568786-qclqp"] Feb 27 08:05:59 crc kubenswrapper[4612]: I0227 08:05:59.918993 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-8s6n2"] Feb 27 08:05:59 crc kubenswrapper[4612]: I0227 08:05:59.928945 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-6994f66f48-qs4l7"] Feb 27 08:05:59 crc kubenswrapper[4612]: I0227 08:05:59.958286 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-68f46476f-r8lqs"] Feb 27 08:05:59 crc kubenswrapper[4612]: W0227 08:05:59.977315 4612 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf62fea1f_71de_4872_9baa_42ad2448d5f5.slice/crio-f00de76821c33a44c76016316f9148f5c2bb17315ad905fca59c63b7eab582f1 WatchSource:0}: Error finding container f00de76821c33a44c76016316f9148f5c2bb17315ad905fca59c63b7eab582f1: Status 404 returned error can't find the container with id f00de76821c33a44c76016316f9148f5c2bb17315ad905fca59c63b7eab582f1 Feb 27 08:05:59 crc kubenswrapper[4612]: I0227 08:05:59.978804 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5dc6794d5b-s2jnh"] Feb 27 08:05:59 crc kubenswrapper[4612]: E0227 08:05:59.981127 4612 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/mariadb-operator@sha256:a18f12497b7159b100fcfd72c7ba2273d0669a5c00600a9ff1333bca028f256a,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-hw2ws,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod mariadb-operator-controller-manager-6994f66f48-qs4l7_openstack-operators(f62fea1f-71de-4872-9baa-42ad2448d5f5): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Feb 27 08:05:59 crc kubenswrapper[4612]: E0227 08:05:59.982723 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/mariadb-operator-controller-manager-6994f66f48-qs4l7" podUID="f62fea1f-71de-4872-9baa-42ad2448d5f5" Feb 27 08:05:59 crc kubenswrapper[4612]: E0227 08:05:59.986090 4612 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/cinder-operator@sha256:7961c67cfc87de69055f8330771af625f73d857426c4bb17ebb888ead843fff3,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-b5897,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-operator-controller-manager-55d77d7b5c-462m7_openstack-operators(36fc734e-4b3e-46bb-b914-896b1ec5c2ee): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Feb 27 08:05:59 crc kubenswrapper[4612]: E0227 08:05:59.990077 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/cinder-operator-controller-manager-55d77d7b5c-462m7" podUID="36fc734e-4b3e-46bb-b914-896b1ec5c2ee" Feb 27 08:05:59 crc kubenswrapper[4612]: E0227 08:05:59.997383 4612 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/test-operator@sha256:38e6a5bd24ab1684f22a64186fe99a7cdc7897eb7feb715ec1704eea7596dd98,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-hnztn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-5dc6794d5b-s2jnh_openstack-operators(ad4fca59-6a87-4ebc-8681-d5a3a40f0b82): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Feb 27 08:05:59 crc kubenswrapper[4612]: E0227 08:05:59.999496 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/test-operator-controller-manager-5dc6794d5b-s2jnh" podUID="ad4fca59-6a87-4ebc-8681-d5a3a40f0b82" Feb 27 08:06:00 crc kubenswrapper[4612]: E0227 08:06:00.010607 4612 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/nova-operator@sha256:fe85dd595906fac0fe1e7a42215bb306a963cf87d55e07cd2573726b690b2838,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-8vwbr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod nova-operator-controller-manager-567668f5cf-tjqwp_openstack-operators(d84adbda-b53c-46c7-9196-682910a0009c): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Feb 27 08:06:00 crc kubenswrapper[4612]: E0227 08:06:00.012761 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/nova-operator-controller-manager-567668f5cf-tjqwp" podUID="d84adbda-b53c-46c7-9196-682910a0009c" Feb 27 08:06:00 crc kubenswrapper[4612]: I0227 08:06:00.014135 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-55d77d7b5c-462m7"] Feb 27 08:06:00 crc kubenswrapper[4612]: I0227 08:06:00.020890 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-5955d8c787-q74ht"] Feb 27 08:06:00 crc kubenswrapper[4612]: I0227 08:06:00.025395 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-567668f5cf-tjqwp"] Feb 27 08:06:00 crc kubenswrapper[4612]: W0227 08:06:00.035476 4612 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode0304d11_114b_4aa3_86ca_1747c60b5bfc.slice/crio-deb4506c643d4ae959a78ba0cdec88590debb5c7173f88da3944f9fd12985c04 WatchSource:0}: Error finding container deb4506c643d4ae959a78ba0cdec88590debb5c7173f88da3944f9fd12985c04: Status 404 returned error can't find the container with id deb4506c643d4ae959a78ba0cdec88590debb5c7173f88da3944f9fd12985c04 Feb 27 08:06:00 crc kubenswrapper[4612]: E0227 08:06:00.050210 4612 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/neutron-operator@sha256:14ae1fb8d065e2317959ce7490a878dc87731d27ebf40259f801ba1a83cfefcf,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-ws5gl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod neutron-operator-controller-manager-6bd4687957-2qmzb_openstack-operators(f292b2e2-2b32-4e4e-bfca-d5db60f0b358): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Feb 27 08:06:00 crc kubenswrapper[4612]: E0227 08:06:00.051328 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/neutron-operator-controller-manager-6bd4687957-2qmzb" podUID="f292b2e2-2b32-4e4e-bfca-d5db60f0b358" Feb 27 08:06:00 crc kubenswrapper[4612]: I0227 08:06:00.053138 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-6bd4687957-2qmzb"] Feb 27 08:06:00 crc kubenswrapper[4612]: I0227 08:06:00.127891 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29536326-pjlzf"] Feb 27 08:06:00 crc kubenswrapper[4612]: I0227 08:06:00.128714 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536326-pjlzf" Feb 27 08:06:00 crc kubenswrapper[4612]: I0227 08:06:00.135430 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 27 08:06:00 crc kubenswrapper[4612]: I0227 08:06:00.135816 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-5zwfx" Feb 27 08:06:00 crc kubenswrapper[4612]: I0227 08:06:00.138170 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 27 08:06:00 crc kubenswrapper[4612]: I0227 08:06:00.144754 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536326-pjlzf"] Feb 27 08:06:00 crc kubenswrapper[4612]: I0227 08:06:00.258810 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/41aed211-3a1f-4f7c-86ca-53e445dcd585-webhook-certs\") pod \"openstack-operator-controller-manager-69f9884c8f-v4kkz\" (UID: \"41aed211-3a1f-4f7c-86ca-53e445dcd585\") " pod="openstack-operators/openstack-operator-controller-manager-69f9884c8f-v4kkz" Feb 27 08:06:00 crc kubenswrapper[4612]: I0227 08:06:00.258898 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kdslx\" (UniqueName: \"kubernetes.io/projected/3aec7965-4aeb-432f-8b2a-f17d58427093-kube-api-access-kdslx\") pod \"auto-csr-approver-29536326-pjlzf\" (UID: \"3aec7965-4aeb-432f-8b2a-f17d58427093\") " pod="openshift-infra/auto-csr-approver-29536326-pjlzf" Feb 27 08:06:00 crc kubenswrapper[4612]: I0227 08:06:00.258937 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/41aed211-3a1f-4f7c-86ca-53e445dcd585-metrics-certs\") pod \"openstack-operator-controller-manager-69f9884c8f-v4kkz\" (UID: \"41aed211-3a1f-4f7c-86ca-53e445dcd585\") " pod="openstack-operators/openstack-operator-controller-manager-69f9884c8f-v4kkz" Feb 27 08:06:00 crc kubenswrapper[4612]: E0227 08:06:00.259009 4612 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Feb 27 08:06:00 crc kubenswrapper[4612]: E0227 08:06:00.259065 4612 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Feb 27 08:06:00 crc kubenswrapper[4612]: E0227 08:06:00.259091 4612 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/41aed211-3a1f-4f7c-86ca-53e445dcd585-webhook-certs podName:41aed211-3a1f-4f7c-86ca-53e445dcd585 nodeName:}" failed. No retries permitted until 2026-02-27 08:06:02.259072834 +0000 UTC m=+1020.113002832 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/41aed211-3a1f-4f7c-86ca-53e445dcd585-webhook-certs") pod "openstack-operator-controller-manager-69f9884c8f-v4kkz" (UID: "41aed211-3a1f-4f7c-86ca-53e445dcd585") : secret "webhook-server-cert" not found Feb 27 08:06:00 crc kubenswrapper[4612]: E0227 08:06:00.259113 4612 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/41aed211-3a1f-4f7c-86ca-53e445dcd585-metrics-certs podName:41aed211-3a1f-4f7c-86ca-53e445dcd585 nodeName:}" failed. No retries permitted until 2026-02-27 08:06:02.259098354 +0000 UTC m=+1020.113028352 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/41aed211-3a1f-4f7c-86ca-53e445dcd585-metrics-certs") pod "openstack-operator-controller-manager-69f9884c8f-v4kkz" (UID: "41aed211-3a1f-4f7c-86ca-53e445dcd585") : secret "metrics-server-cert" not found Feb 27 08:06:00 crc kubenswrapper[4612]: I0227 08:06:00.360891 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kdslx\" (UniqueName: \"kubernetes.io/projected/3aec7965-4aeb-432f-8b2a-f17d58427093-kube-api-access-kdslx\") pod \"auto-csr-approver-29536326-pjlzf\" (UID: \"3aec7965-4aeb-432f-8b2a-f17d58427093\") " pod="openshift-infra/auto-csr-approver-29536326-pjlzf" Feb 27 08:06:00 crc kubenswrapper[4612]: I0227 08:06:00.399360 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kdslx\" (UniqueName: \"kubernetes.io/projected/3aec7965-4aeb-432f-8b2a-f17d58427093-kube-api-access-kdslx\") pod \"auto-csr-approver-29536326-pjlzf\" (UID: \"3aec7965-4aeb-432f-8b2a-f17d58427093\") " pod="openshift-infra/auto-csr-approver-29536326-pjlzf" Feb 27 08:06:00 crc kubenswrapper[4612]: I0227 08:06:00.506141 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536326-pjlzf" Feb 27 08:06:00 crc kubenswrapper[4612]: I0227 08:06:00.798955 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-bccc79885-zdtzv" event={"ID":"a4619f7b-7242-459b-8558-f84c2893ee55","Type":"ContainerStarted","Data":"f4e9e365f26c17e1f922a514a09c61c5cb034fbf6d08e45ba8f18c3a3b335d49"} Feb 27 08:06:00 crc kubenswrapper[4612]: I0227 08:06:00.803461 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-55d77d7b5c-462m7" event={"ID":"36fc734e-4b3e-46bb-b914-896b1ec5c2ee","Type":"ContainerStarted","Data":"53a90452ccb91804cecf6a99752d454360268f2c2f8df83a7abdddf7d95f1c93"} Feb 27 08:06:00 crc kubenswrapper[4612]: E0227 08:06:00.805012 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/cinder-operator@sha256:7961c67cfc87de69055f8330771af625f73d857426c4bb17ebb888ead843fff3\\\"\"" pod="openstack-operators/cinder-operator-controller-manager-55d77d7b5c-462m7" podUID="36fc734e-4b3e-46bb-b914-896b1ec5c2ee" Feb 27 08:06:00 crc kubenswrapper[4612]: I0227 08:06:00.840037 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-68f46476f-r8lqs" event={"ID":"713a994e-490d-47b4-86ab-1203a6b806fc","Type":"ContainerStarted","Data":"6c69543b1c2654f7171cf62078e2961897124889baad626ed7f85ee27fb96183"} Feb 27 08:06:00 crc kubenswrapper[4612]: I0227 08:06:00.843020 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-567668f5cf-tjqwp" event={"ID":"d84adbda-b53c-46c7-9196-682910a0009c","Type":"ContainerStarted","Data":"2e0aec7ca4f5b2b3d13ab286bb118334a9741caf90f07ab0420d72ae05fb18cc"} Feb 27 08:06:00 crc kubenswrapper[4612]: I0227 08:06:00.845114 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-5955d8c787-q74ht" event={"ID":"e0304d11-114b-4aa3-86ca-1747c60b5bfc","Type":"ContainerStarted","Data":"deb4506c643d4ae959a78ba0cdec88590debb5c7173f88da3944f9fd12985c04"} Feb 27 08:06:00 crc kubenswrapper[4612]: E0227 08:06:00.846083 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/nova-operator@sha256:fe85dd595906fac0fe1e7a42215bb306a963cf87d55e07cd2573726b690b2838\\\"\"" pod="openstack-operators/nova-operator-controller-manager-567668f5cf-tjqwp" podUID="d84adbda-b53c-46c7-9196-682910a0009c" Feb 27 08:06:00 crc kubenswrapper[4612]: I0227 08:06:00.848452 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5dc6794d5b-s2jnh" event={"ID":"ad4fca59-6a87-4ebc-8681-d5a3a40f0b82","Type":"ContainerStarted","Data":"67b978183804425da0b29314d26709151e7c4b32ed183bc6244af27d4c9a14f5"} Feb 27 08:06:00 crc kubenswrapper[4612]: E0227 08:06:00.853660 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:38e6a5bd24ab1684f22a64186fe99a7cdc7897eb7feb715ec1704eea7596dd98\\\"\"" pod="openstack-operators/test-operator-controller-manager-5dc6794d5b-s2jnh" podUID="ad4fca59-6a87-4ebc-8681-d5a3a40f0b82" Feb 27 08:06:00 crc kubenswrapper[4612]: I0227 08:06:00.887750 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-8s6n2" event={"ID":"57ba67ae-3e41-4e2b-9ee6-30b2643805dd","Type":"ContainerStarted","Data":"c73763df49e0766c9e3a256e74257de2ec59dbac9cabf864850b236edcf248ca"} Feb 27 08:06:00 crc kubenswrapper[4612]: I0227 08:06:00.889428 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-589c568786-qclqp" event={"ID":"ee3f21f6-2f91-45c6-8c80-5fb8f89a41ce","Type":"ContainerStarted","Data":"aa056ee8bb7ccff5f70efe2981669a857e8cf6a42742413b80aeae5fa3dffd80"} Feb 27 08:06:00 crc kubenswrapper[4612]: I0227 08:06:00.890778 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-6994f66f48-qs4l7" event={"ID":"f62fea1f-71de-4872-9baa-42ad2448d5f5","Type":"ContainerStarted","Data":"f00de76821c33a44c76016316f9148f5c2bb17315ad905fca59c63b7eab582f1"} Feb 27 08:06:00 crc kubenswrapper[4612]: E0227 08:06:00.894530 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/mariadb-operator@sha256:a18f12497b7159b100fcfd72c7ba2273d0669a5c00600a9ff1333bca028f256a\\\"\"" pod="openstack-operators/mariadb-operator-controller-manager-6994f66f48-qs4l7" podUID="f62fea1f-71de-4872-9baa-42ad2448d5f5" Feb 27 08:06:00 crc kubenswrapper[4612]: I0227 08:06:00.903098 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-659dc6bbfc-vlczg" event={"ID":"bd923861-d2ad-4a06-b0c7-613605afefbb","Type":"ContainerStarted","Data":"1add4ce827caeb23e6f9adc44bcb052473cd30c8de7796433013caab1f932478"} Feb 27 08:06:00 crc kubenswrapper[4612]: I0227 08:06:00.926109 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-6bd4687957-2qmzb" event={"ID":"f292b2e2-2b32-4e4e-bfca-d5db60f0b358","Type":"ContainerStarted","Data":"99276c282e8ee90d362d59d61db79d5a6b82aa16c77604767d3ee3625c97be38"} Feb 27 08:06:00 crc kubenswrapper[4612]: E0227 08:06:00.929782 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/neutron-operator@sha256:14ae1fb8d065e2317959ce7490a878dc87731d27ebf40259f801ba1a83cfefcf\\\"\"" pod="openstack-operators/neutron-operator-controller-manager-6bd4687957-2qmzb" podUID="f292b2e2-2b32-4e4e-bfca-d5db60f0b358" Feb 27 08:06:01 crc kubenswrapper[4612]: I0227 08:06:01.020561 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536326-pjlzf"] Feb 27 08:06:01 crc kubenswrapper[4612]: W0227 08:06:01.057793 4612 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3aec7965_4aeb_432f_8b2a_f17d58427093.slice/crio-c3fce31c1331a7be792f6e0230f8814237f92ec0aa01baa28d206a188ff44377 WatchSource:0}: Error finding container c3fce31c1331a7be792f6e0230f8814237f92ec0aa01baa28d206a188ff44377: Status 404 returned error can't find the container with id c3fce31c1331a7be792f6e0230f8814237f92ec0aa01baa28d206a188ff44377 Feb 27 08:06:01 crc kubenswrapper[4612]: I0227 08:06:01.577474 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/98296d45-b487-47c3-a44a-fa63ba52426b-cert\") pod \"infra-operator-controller-manager-79d975b745-plm9k\" (UID: \"98296d45-b487-47c3-a44a-fa63ba52426b\") " pod="openstack-operators/infra-operator-controller-manager-79d975b745-plm9k" Feb 27 08:06:01 crc kubenswrapper[4612]: E0227 08:06:01.577809 4612 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Feb 27 08:06:01 crc kubenswrapper[4612]: E0227 08:06:01.577913 4612 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/98296d45-b487-47c3-a44a-fa63ba52426b-cert podName:98296d45-b487-47c3-a44a-fa63ba52426b nodeName:}" failed. No retries permitted until 2026-02-27 08:06:05.577895025 +0000 UTC m=+1023.431825023 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/98296d45-b487-47c3-a44a-fa63ba52426b-cert") pod "infra-operator-controller-manager-79d975b745-plm9k" (UID: "98296d45-b487-47c3-a44a-fa63ba52426b") : secret "infra-operator-webhook-server-cert" not found Feb 27 08:06:01 crc kubenswrapper[4612]: I0227 08:06:01.882733 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/cf5ef40d-5e5c-4b2c-9570-63e552800b4e-cert\") pod \"openstack-baremetal-operator-controller-manager-7c6767dc9clp9dx\" (UID: \"cf5ef40d-5e5c-4b2c-9570-63e552800b4e\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9clp9dx" Feb 27 08:06:01 crc kubenswrapper[4612]: E0227 08:06:01.882870 4612 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 27 08:06:01 crc kubenswrapper[4612]: E0227 08:06:01.882924 4612 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/cf5ef40d-5e5c-4b2c-9570-63e552800b4e-cert podName:cf5ef40d-5e5c-4b2c-9570-63e552800b4e nodeName:}" failed. No retries permitted until 2026-02-27 08:06:05.882909957 +0000 UTC m=+1023.736839955 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/cf5ef40d-5e5c-4b2c-9570-63e552800b4e-cert") pod "openstack-baremetal-operator-controller-manager-7c6767dc9clp9dx" (UID: "cf5ef40d-5e5c-4b2c-9570-63e552800b4e") : secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 27 08:06:01 crc kubenswrapper[4612]: I0227 08:06:01.938503 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536326-pjlzf" event={"ID":"3aec7965-4aeb-432f-8b2a-f17d58427093","Type":"ContainerStarted","Data":"c3fce31c1331a7be792f6e0230f8814237f92ec0aa01baa28d206a188ff44377"} Feb 27 08:06:01 crc kubenswrapper[4612]: E0227 08:06:01.952684 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/cinder-operator@sha256:7961c67cfc87de69055f8330771af625f73d857426c4bb17ebb888ead843fff3\\\"\"" pod="openstack-operators/cinder-operator-controller-manager-55d77d7b5c-462m7" podUID="36fc734e-4b3e-46bb-b914-896b1ec5c2ee" Feb 27 08:06:01 crc kubenswrapper[4612]: E0227 08:06:01.952724 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:38e6a5bd24ab1684f22a64186fe99a7cdc7897eb7feb715ec1704eea7596dd98\\\"\"" pod="openstack-operators/test-operator-controller-manager-5dc6794d5b-s2jnh" podUID="ad4fca59-6a87-4ebc-8681-d5a3a40f0b82" Feb 27 08:06:01 crc kubenswrapper[4612]: E0227 08:06:01.952725 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/neutron-operator@sha256:14ae1fb8d065e2317959ce7490a878dc87731d27ebf40259f801ba1a83cfefcf\\\"\"" pod="openstack-operators/neutron-operator-controller-manager-6bd4687957-2qmzb" podUID="f292b2e2-2b32-4e4e-bfca-d5db60f0b358" Feb 27 08:06:01 crc kubenswrapper[4612]: E0227 08:06:01.955155 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/mariadb-operator@sha256:a18f12497b7159b100fcfd72c7ba2273d0669a5c00600a9ff1333bca028f256a\\\"\"" pod="openstack-operators/mariadb-operator-controller-manager-6994f66f48-qs4l7" podUID="f62fea1f-71de-4872-9baa-42ad2448d5f5" Feb 27 08:06:01 crc kubenswrapper[4612]: E0227 08:06:01.955221 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/nova-operator@sha256:fe85dd595906fac0fe1e7a42215bb306a963cf87d55e07cd2573726b690b2838\\\"\"" pod="openstack-operators/nova-operator-controller-manager-567668f5cf-tjqwp" podUID="d84adbda-b53c-46c7-9196-682910a0009c" Feb 27 08:06:02 crc kubenswrapper[4612]: I0227 08:06:02.292569 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/41aed211-3a1f-4f7c-86ca-53e445dcd585-webhook-certs\") pod \"openstack-operator-controller-manager-69f9884c8f-v4kkz\" (UID: \"41aed211-3a1f-4f7c-86ca-53e445dcd585\") " pod="openstack-operators/openstack-operator-controller-manager-69f9884c8f-v4kkz" Feb 27 08:06:02 crc kubenswrapper[4612]: E0227 08:06:02.292793 4612 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Feb 27 08:06:02 crc kubenswrapper[4612]: E0227 08:06:02.293114 4612 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Feb 27 08:06:02 crc kubenswrapper[4612]: E0227 08:06:02.293127 4612 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/41aed211-3a1f-4f7c-86ca-53e445dcd585-webhook-certs podName:41aed211-3a1f-4f7c-86ca-53e445dcd585 nodeName:}" failed. No retries permitted until 2026-02-27 08:06:06.293091219 +0000 UTC m=+1024.147021397 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/41aed211-3a1f-4f7c-86ca-53e445dcd585-webhook-certs") pod "openstack-operator-controller-manager-69f9884c8f-v4kkz" (UID: "41aed211-3a1f-4f7c-86ca-53e445dcd585") : secret "webhook-server-cert" not found Feb 27 08:06:02 crc kubenswrapper[4612]: I0227 08:06:02.293006 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/41aed211-3a1f-4f7c-86ca-53e445dcd585-metrics-certs\") pod \"openstack-operator-controller-manager-69f9884c8f-v4kkz\" (UID: \"41aed211-3a1f-4f7c-86ca-53e445dcd585\") " pod="openstack-operators/openstack-operator-controller-manager-69f9884c8f-v4kkz" Feb 27 08:06:02 crc kubenswrapper[4612]: E0227 08:06:02.293174 4612 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/41aed211-3a1f-4f7c-86ca-53e445dcd585-metrics-certs podName:41aed211-3a1f-4f7c-86ca-53e445dcd585 nodeName:}" failed. No retries permitted until 2026-02-27 08:06:06.293154881 +0000 UTC m=+1024.147085089 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/41aed211-3a1f-4f7c-86ca-53e445dcd585-metrics-certs") pod "openstack-operator-controller-manager-69f9884c8f-v4kkz" (UID: "41aed211-3a1f-4f7c-86ca-53e445dcd585") : secret "metrics-server-cert" not found Feb 27 08:06:05 crc kubenswrapper[4612]: I0227 08:06:05.652364 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/98296d45-b487-47c3-a44a-fa63ba52426b-cert\") pod \"infra-operator-controller-manager-79d975b745-plm9k\" (UID: \"98296d45-b487-47c3-a44a-fa63ba52426b\") " pod="openstack-operators/infra-operator-controller-manager-79d975b745-plm9k" Feb 27 08:06:05 crc kubenswrapper[4612]: E0227 08:06:05.652516 4612 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Feb 27 08:06:05 crc kubenswrapper[4612]: E0227 08:06:05.653266 4612 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/98296d45-b487-47c3-a44a-fa63ba52426b-cert podName:98296d45-b487-47c3-a44a-fa63ba52426b nodeName:}" failed. No retries permitted until 2026-02-27 08:06:13.653240286 +0000 UTC m=+1031.507170304 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/98296d45-b487-47c3-a44a-fa63ba52426b-cert") pod "infra-operator-controller-manager-79d975b745-plm9k" (UID: "98296d45-b487-47c3-a44a-fa63ba52426b") : secret "infra-operator-webhook-server-cert" not found Feb 27 08:06:05 crc kubenswrapper[4612]: I0227 08:06:05.958151 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/cf5ef40d-5e5c-4b2c-9570-63e552800b4e-cert\") pod \"openstack-baremetal-operator-controller-manager-7c6767dc9clp9dx\" (UID: \"cf5ef40d-5e5c-4b2c-9570-63e552800b4e\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9clp9dx" Feb 27 08:06:05 crc kubenswrapper[4612]: E0227 08:06:05.958446 4612 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 27 08:06:05 crc kubenswrapper[4612]: E0227 08:06:05.959622 4612 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/cf5ef40d-5e5c-4b2c-9570-63e552800b4e-cert podName:cf5ef40d-5e5c-4b2c-9570-63e552800b4e nodeName:}" failed. No retries permitted until 2026-02-27 08:06:13.959601546 +0000 UTC m=+1031.813531544 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/cf5ef40d-5e5c-4b2c-9570-63e552800b4e-cert") pod "openstack-baremetal-operator-controller-manager-7c6767dc9clp9dx" (UID: "cf5ef40d-5e5c-4b2c-9570-63e552800b4e") : secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 27 08:06:06 crc kubenswrapper[4612]: I0227 08:06:06.366174 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/41aed211-3a1f-4f7c-86ca-53e445dcd585-metrics-certs\") pod \"openstack-operator-controller-manager-69f9884c8f-v4kkz\" (UID: \"41aed211-3a1f-4f7c-86ca-53e445dcd585\") " pod="openstack-operators/openstack-operator-controller-manager-69f9884c8f-v4kkz" Feb 27 08:06:06 crc kubenswrapper[4612]: I0227 08:06:06.366312 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/41aed211-3a1f-4f7c-86ca-53e445dcd585-webhook-certs\") pod \"openstack-operator-controller-manager-69f9884c8f-v4kkz\" (UID: \"41aed211-3a1f-4f7c-86ca-53e445dcd585\") " pod="openstack-operators/openstack-operator-controller-manager-69f9884c8f-v4kkz" Feb 27 08:06:06 crc kubenswrapper[4612]: E0227 08:06:06.366490 4612 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Feb 27 08:06:06 crc kubenswrapper[4612]: E0227 08:06:06.366551 4612 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Feb 27 08:06:06 crc kubenswrapper[4612]: E0227 08:06:06.366638 4612 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/41aed211-3a1f-4f7c-86ca-53e445dcd585-metrics-certs podName:41aed211-3a1f-4f7c-86ca-53e445dcd585 nodeName:}" failed. No retries permitted until 2026-02-27 08:06:14.366604157 +0000 UTC m=+1032.220534345 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/41aed211-3a1f-4f7c-86ca-53e445dcd585-metrics-certs") pod "openstack-operator-controller-manager-69f9884c8f-v4kkz" (UID: "41aed211-3a1f-4f7c-86ca-53e445dcd585") : secret "metrics-server-cert" not found Feb 27 08:06:06 crc kubenswrapper[4612]: E0227 08:06:06.366671 4612 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/41aed211-3a1f-4f7c-86ca-53e445dcd585-webhook-certs podName:41aed211-3a1f-4f7c-86ca-53e445dcd585 nodeName:}" failed. No retries permitted until 2026-02-27 08:06:14.366660389 +0000 UTC m=+1032.220590637 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/41aed211-3a1f-4f7c-86ca-53e445dcd585-webhook-certs") pod "openstack-operator-controller-manager-69f9884c8f-v4kkz" (UID: "41aed211-3a1f-4f7c-86ca-53e445dcd585") : secret "webhook-server-cert" not found Feb 27 08:06:11 crc kubenswrapper[4612]: E0227 08:06:11.891857 4612 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/barbican-operator@sha256:90ad8fd8c1889b6be77925016532218eb6149d2c1c8535a5f9f1775c776fa6cc" Feb 27 08:06:11 crc kubenswrapper[4612]: E0227 08:06:11.892432 4612 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/barbican-operator@sha256:90ad8fd8c1889b6be77925016532218eb6149d2c1c8535a5f9f1775c776fa6cc,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-br2gx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod barbican-operator-controller-manager-868647ff47-mlgz7_openstack-operators(2ddc15ea-b7c1-44e7-b795-392890e9c14e): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 27 08:06:11 crc kubenswrapper[4612]: E0227 08:06:11.894412 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/barbican-operator-controller-manager-868647ff47-mlgz7" podUID="2ddc15ea-b7c1-44e7-b795-392890e9c14e" Feb 27 08:06:12 crc kubenswrapper[4612]: E0227 08:06:12.030782 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/barbican-operator@sha256:90ad8fd8c1889b6be77925016532218eb6149d2c1c8535a5f9f1775c776fa6cc\\\"\"" pod="openstack-operators/barbican-operator-controller-manager-868647ff47-mlgz7" podUID="2ddc15ea-b7c1-44e7-b795-392890e9c14e" Feb 27 08:06:13 crc kubenswrapper[4612]: I0227 08:06:13.690579 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/98296d45-b487-47c3-a44a-fa63ba52426b-cert\") pod \"infra-operator-controller-manager-79d975b745-plm9k\" (UID: \"98296d45-b487-47c3-a44a-fa63ba52426b\") " pod="openstack-operators/infra-operator-controller-manager-79d975b745-plm9k" Feb 27 08:06:13 crc kubenswrapper[4612]: E0227 08:06:13.691169 4612 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Feb 27 08:06:13 crc kubenswrapper[4612]: E0227 08:06:13.691918 4612 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/98296d45-b487-47c3-a44a-fa63ba52426b-cert podName:98296d45-b487-47c3-a44a-fa63ba52426b nodeName:}" failed. No retries permitted until 2026-02-27 08:06:29.691897961 +0000 UTC m=+1047.545827979 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/98296d45-b487-47c3-a44a-fa63ba52426b-cert") pod "infra-operator-controller-manager-79d975b745-plm9k" (UID: "98296d45-b487-47c3-a44a-fa63ba52426b") : secret "infra-operator-webhook-server-cert" not found Feb 27 08:06:13 crc kubenswrapper[4612]: E0227 08:06:13.854333 4612 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/ovn-operator@sha256:f4143497c70c048a7733c284060347a0c74ef4e628aca22ee191e5bc9e4c7192" Feb 27 08:06:13 crc kubenswrapper[4612]: E0227 08:06:13.854631 4612 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ovn-operator@sha256:f4143497c70c048a7733c284060347a0c74ef4e628aca22ee191e5bc9e4c7192,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-4wp6l,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovn-operator-controller-manager-5955d8c787-q74ht_openstack-operators(e0304d11-114b-4aa3-86ca-1747c60b5bfc): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 27 08:06:13 crc kubenswrapper[4612]: E0227 08:06:13.855907 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/ovn-operator-controller-manager-5955d8c787-q74ht" podUID="e0304d11-114b-4aa3-86ca-1747c60b5bfc" Feb 27 08:06:13 crc kubenswrapper[4612]: I0227 08:06:13.996077 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/cf5ef40d-5e5c-4b2c-9570-63e552800b4e-cert\") pod \"openstack-baremetal-operator-controller-manager-7c6767dc9clp9dx\" (UID: \"cf5ef40d-5e5c-4b2c-9570-63e552800b4e\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9clp9dx" Feb 27 08:06:13 crc kubenswrapper[4612]: E0227 08:06:13.996260 4612 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 27 08:06:13 crc kubenswrapper[4612]: E0227 08:06:13.996326 4612 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/cf5ef40d-5e5c-4b2c-9570-63e552800b4e-cert podName:cf5ef40d-5e5c-4b2c-9570-63e552800b4e nodeName:}" failed. No retries permitted until 2026-02-27 08:06:29.996308605 +0000 UTC m=+1047.850238623 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/cf5ef40d-5e5c-4b2c-9570-63e552800b4e-cert") pod "openstack-baremetal-operator-controller-manager-7c6767dc9clp9dx" (UID: "cf5ef40d-5e5c-4b2c-9570-63e552800b4e") : secret "openstack-baremetal-operator-webhook-server-cert" not found Feb 27 08:06:14 crc kubenswrapper[4612]: E0227 08:06:14.045586 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ovn-operator@sha256:f4143497c70c048a7733c284060347a0c74ef4e628aca22ee191e5bc9e4c7192\\\"\"" pod="openstack-operators/ovn-operator-controller-manager-5955d8c787-q74ht" podUID="e0304d11-114b-4aa3-86ca-1747c60b5bfc" Feb 27 08:06:14 crc kubenswrapper[4612]: I0227 08:06:14.402755 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/41aed211-3a1f-4f7c-86ca-53e445dcd585-webhook-certs\") pod \"openstack-operator-controller-manager-69f9884c8f-v4kkz\" (UID: \"41aed211-3a1f-4f7c-86ca-53e445dcd585\") " pod="openstack-operators/openstack-operator-controller-manager-69f9884c8f-v4kkz" Feb 27 08:06:14 crc kubenswrapper[4612]: I0227 08:06:14.402847 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/41aed211-3a1f-4f7c-86ca-53e445dcd585-metrics-certs\") pod \"openstack-operator-controller-manager-69f9884c8f-v4kkz\" (UID: \"41aed211-3a1f-4f7c-86ca-53e445dcd585\") " pod="openstack-operators/openstack-operator-controller-manager-69f9884c8f-v4kkz" Feb 27 08:06:14 crc kubenswrapper[4612]: E0227 08:06:14.402879 4612 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Feb 27 08:06:14 crc kubenswrapper[4612]: E0227 08:06:14.402926 4612 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Feb 27 08:06:14 crc kubenswrapper[4612]: E0227 08:06:14.402932 4612 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/41aed211-3a1f-4f7c-86ca-53e445dcd585-webhook-certs podName:41aed211-3a1f-4f7c-86ca-53e445dcd585 nodeName:}" failed. No retries permitted until 2026-02-27 08:06:30.402916335 +0000 UTC m=+1048.256846333 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/41aed211-3a1f-4f7c-86ca-53e445dcd585-webhook-certs") pod "openstack-operator-controller-manager-69f9884c8f-v4kkz" (UID: "41aed211-3a1f-4f7c-86ca-53e445dcd585") : secret "webhook-server-cert" not found Feb 27 08:06:14 crc kubenswrapper[4612]: E0227 08:06:14.402953 4612 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/41aed211-3a1f-4f7c-86ca-53e445dcd585-metrics-certs podName:41aed211-3a1f-4f7c-86ca-53e445dcd585 nodeName:}" failed. No retries permitted until 2026-02-27 08:06:30.402945276 +0000 UTC m=+1048.256875274 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/41aed211-3a1f-4f7c-86ca-53e445dcd585-metrics-certs") pod "openstack-operator-controller-manager-69f9884c8f-v4kkz" (UID: "41aed211-3a1f-4f7c-86ca-53e445dcd585") : secret "metrics-server-cert" not found Feb 27 08:06:14 crc kubenswrapper[4612]: E0227 08:06:14.496220 4612 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/manila-operator@sha256:f1158ec4d879c4646eee4323bc501eba4d377beb2ad6fbe08ed30070c441ac26" Feb 27 08:06:14 crc kubenswrapper[4612]: E0227 08:06:14.496411 4612 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/manila-operator@sha256:f1158ec4d879c4646eee4323bc501eba4d377beb2ad6fbe08ed30070c441ac26,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-fqgp9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod manila-operator-controller-manager-67d996989d-4dhhw_openstack-operators(b60d4209-cd53-4cfe-99f7-7e7351041857): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 27 08:06:14 crc kubenswrapper[4612]: E0227 08:06:14.497808 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/manila-operator-controller-manager-67d996989d-4dhhw" podUID="b60d4209-cd53-4cfe-99f7-7e7351041857" Feb 27 08:06:15 crc kubenswrapper[4612]: E0227 08:06:15.072854 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/manila-operator@sha256:f1158ec4d879c4646eee4323bc501eba4d377beb2ad6fbe08ed30070c441ac26\\\"\"" pod="openstack-operators/manila-operator-controller-manager-67d996989d-4dhhw" podUID="b60d4209-cd53-4cfe-99f7-7e7351041857" Feb 27 08:06:15 crc kubenswrapper[4612]: E0227 08:06:15.074837 4612 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/octavia-operator@sha256:c7c7d4228994efb8b93cfabe4d78b40b085d91848dc49db247b7bbca689dae06" Feb 27 08:06:15 crc kubenswrapper[4612]: E0227 08:06:15.075014 4612 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/octavia-operator@sha256:c7c7d4228994efb8b93cfabe4d78b40b085d91848dc49db247b7bbca689dae06,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-2ltcz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod octavia-operator-controller-manager-659dc6bbfc-vlczg_openstack-operators(bd923861-d2ad-4a06-b0c7-613605afefbb): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 27 08:06:15 crc kubenswrapper[4612]: E0227 08:06:15.076169 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/octavia-operator-controller-manager-659dc6bbfc-vlczg" podUID="bd923861-d2ad-4a06-b0c7-613605afefbb" Feb 27 08:06:15 crc kubenswrapper[4612]: E0227 08:06:15.544250 4612 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/swift-operator@sha256:3d676f1281e24ef07de617570d2f7fbf625032e41866d1551a856c052248bb04" Feb 27 08:06:15 crc kubenswrapper[4612]: E0227 08:06:15.544639 4612 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/swift-operator@sha256:3d676f1281e24ef07de617570d2f7fbf625032e41866d1551a856c052248bb04,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-k5p4x,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-68f46476f-r8lqs_openstack-operators(713a994e-490d-47b4-86ab-1203a6b806fc): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 27 08:06:15 crc kubenswrapper[4612]: E0227 08:06:15.545861 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/swift-operator-controller-manager-68f46476f-r8lqs" podUID="713a994e-490d-47b4-86ab-1203a6b806fc" Feb 27 08:06:16 crc kubenswrapper[4612]: E0227 08:06:16.089005 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/octavia-operator@sha256:c7c7d4228994efb8b93cfabe4d78b40b085d91848dc49db247b7bbca689dae06\\\"\"" pod="openstack-operators/octavia-operator-controller-manager-659dc6bbfc-vlczg" podUID="bd923861-d2ad-4a06-b0c7-613605afefbb" Feb 27 08:06:16 crc kubenswrapper[4612]: E0227 08:06:16.089191 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:3d676f1281e24ef07de617570d2f7fbf625032e41866d1551a856c052248bb04\\\"\"" pod="openstack-operators/swift-operator-controller-manager-68f46476f-r8lqs" podUID="713a994e-490d-47b4-86ab-1203a6b806fc" Feb 27 08:06:17 crc kubenswrapper[4612]: E0227 08:06:17.556543 4612 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/designate-operator@sha256:c1e33e962043cd6e3d09ebd225cb72781451dba7af2d57522e5c6eedbdc91642" Feb 27 08:06:17 crc kubenswrapper[4612]: E0227 08:06:17.557177 4612 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/designate-operator@sha256:c1e33e962043cd6e3d09ebd225cb72781451dba7af2d57522e5c6eedbdc91642,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-f5pcx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod designate-operator-controller-manager-6d8bf5c495-t95hp_openstack-operators(63d519c2-d2f9-4349-ad82-0399e472d78a): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 27 08:06:17 crc kubenswrapper[4612]: E0227 08:06:17.558351 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/designate-operator-controller-manager-6d8bf5c495-t95hp" podUID="63d519c2-d2f9-4349-ad82-0399e472d78a" Feb 27 08:06:18 crc kubenswrapper[4612]: E0227 08:06:18.098941 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/designate-operator@sha256:c1e33e962043cd6e3d09ebd225cb72781451dba7af2d57522e5c6eedbdc91642\\\"\"" pod="openstack-operators/designate-operator-controller-manager-6d8bf5c495-t95hp" podUID="63d519c2-d2f9-4349-ad82-0399e472d78a" Feb 27 08:06:22 crc kubenswrapper[4612]: E0227 08:06:22.755280 4612 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/placement-operator@sha256:a57336b9f95b703f80453db87e43a2834ca1bdc89480796d28ebbe0a9702ecfd" Feb 27 08:06:22 crc kubenswrapper[4612]: E0227 08:06:22.755779 4612 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/placement-operator@sha256:a57336b9f95b703f80453db87e43a2834ca1bdc89480796d28ebbe0a9702ecfd,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-sfh76,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-8497b45c89-dvq8j_openstack-operators(8414b96c-7e5b-4824-998e-549a0c43d9d5): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 27 08:06:22 crc kubenswrapper[4612]: E0227 08:06:22.756991 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/placement-operator-controller-manager-8497b45c89-dvq8j" podUID="8414b96c-7e5b-4824-998e-549a0c43d9d5" Feb 27 08:06:23 crc kubenswrapper[4612]: E0227 08:06:23.134477 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:a57336b9f95b703f80453db87e43a2834ca1bdc89480796d28ebbe0a9702ecfd\\\"\"" pod="openstack-operators/placement-operator-controller-manager-8497b45c89-dvq8j" podUID="8414b96c-7e5b-4824-998e-549a0c43d9d5" Feb 27 08:06:24 crc kubenswrapper[4612]: E0227 08:06:24.449818 4612 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/ironic-operator@sha256:7e1b0b7b172ad0d707ab80dd72d609e1d0f5bbd38a22c24a28ed0f17a960c867" Feb 27 08:06:24 crc kubenswrapper[4612]: E0227 08:06:24.450441 4612 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ironic-operator@sha256:7e1b0b7b172ad0d707ab80dd72d609e1d0f5bbd38a22c24a28ed0f17a960c867,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-j284h,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ironic-operator-controller-manager-554564d7fc-8kpwf_openstack-operators(bd687bfa-c843-4d11-80f9-cedcdce41e3d): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 27 08:06:24 crc kubenswrapper[4612]: E0227 08:06:24.451660 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/ironic-operator-controller-manager-554564d7fc-8kpwf" podUID="bd687bfa-c843-4d11-80f9-cedcdce41e3d" Feb 27 08:06:24 crc kubenswrapper[4612]: E0227 08:06:24.924480 4612 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/horizon-operator@sha256:9f2e1299d908411457e53b49e1062265d2b9d76f6719db24d1be9347c388e4da" Feb 27 08:06:24 crc kubenswrapper[4612]: E0227 08:06:24.924677 4612 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/horizon-operator@sha256:9f2e1299d908411457e53b49e1062265d2b9d76f6719db24d1be9347c388e4da,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-tmdcx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-operator-controller-manager-5b9b8895d5-t4rct_openstack-operators(1903598c-2428-4e90-ab81-9a47fa5af9ef): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 27 08:06:24 crc kubenswrapper[4612]: E0227 08:06:24.925903 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/horizon-operator-controller-manager-5b9b8895d5-t4rct" podUID="1903598c-2428-4e90-ab81-9a47fa5af9ef" Feb 27 08:06:25 crc kubenswrapper[4612]: E0227 08:06:25.147536 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/horizon-operator@sha256:9f2e1299d908411457e53b49e1062265d2b9d76f6719db24d1be9347c388e4da\\\"\"" pod="openstack-operators/horizon-operator-controller-manager-5b9b8895d5-t4rct" podUID="1903598c-2428-4e90-ab81-9a47fa5af9ef" Feb 27 08:06:25 crc kubenswrapper[4612]: E0227 08:06:25.148747 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ironic-operator@sha256:7e1b0b7b172ad0d707ab80dd72d609e1d0f5bbd38a22c24a28ed0f17a960c867\\\"\"" pod="openstack-operators/ironic-operator-controller-manager-554564d7fc-8kpwf" podUID="bd687bfa-c843-4d11-80f9-cedcdce41e3d" Feb 27 08:06:26 crc kubenswrapper[4612]: E0227 08:06:26.379917 4612 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/heat-operator@sha256:e8a675284ff97a1d3f0f07583863be20b20b4aa48ebb34dbc80d83fe39d757b2" Feb 27 08:06:26 crc kubenswrapper[4612]: E0227 08:06:26.380498 4612 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/heat-operator@sha256:e8a675284ff97a1d3f0f07583863be20b20b4aa48ebb34dbc80d83fe39d757b2,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-gmvbz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod heat-operator-controller-manager-69f49c598c-7f9bj_openstack-operators(6065186a-5435-4db3-8adc-698bec4358ef): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 27 08:06:26 crc kubenswrapper[4612]: E0227 08:06:26.382529 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/heat-operator-controller-manager-69f49c598c-7f9bj" podUID="6065186a-5435-4db3-8adc-698bec4358ef" Feb 27 08:06:26 crc kubenswrapper[4612]: E0227 08:06:26.891253 4612 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/watcher-operator@sha256:06311600a491c689493552e7ff26e36df740fa4e7c143fca874bef19f24afb97" Feb 27 08:06:26 crc kubenswrapper[4612]: E0227 08:06:26.891436 4612 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/watcher-operator@sha256:06311600a491c689493552e7ff26e36df740fa4e7c143fca874bef19f24afb97,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-wn2qx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-bccc79885-zdtzv_openstack-operators(a4619f7b-7242-459b-8558-f84c2893ee55): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 27 08:06:26 crc kubenswrapper[4612]: E0227 08:06:26.892589 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/watcher-operator-controller-manager-bccc79885-zdtzv" podUID="a4619f7b-7242-459b-8558-f84c2893ee55" Feb 27 08:06:27 crc kubenswrapper[4612]: E0227 08:06:27.162180 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/heat-operator@sha256:e8a675284ff97a1d3f0f07583863be20b20b4aa48ebb34dbc80d83fe39d757b2\\\"\"" pod="openstack-operators/heat-operator-controller-manager-69f49c598c-7f9bj" podUID="6065186a-5435-4db3-8adc-698bec4358ef" Feb 27 08:06:27 crc kubenswrapper[4612]: E0227 08:06:27.165643 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:06311600a491c689493552e7ff26e36df740fa4e7c143fca874bef19f24afb97\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-bccc79885-zdtzv" podUID="a4619f7b-7242-459b-8558-f84c2893ee55" Feb 27 08:06:27 crc kubenswrapper[4612]: E0227 08:06:27.946728 4612 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/keystone-operator@sha256:c6ad383f55f955902b074d1ee947a2233a5fcbf40698479ae693ce056c80dcc1" Feb 27 08:06:27 crc kubenswrapper[4612]: E0227 08:06:27.946934 4612 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/keystone-operator@sha256:c6ad383f55f955902b074d1ee947a2233a5fcbf40698479ae693ce056c80dcc1,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-lflq8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod keystone-operator-controller-manager-b4d948c87-2hm5l_openstack-operators(259df4bd-af1b-4c61-a5bd-0ca28ac531d4): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 27 08:06:27 crc kubenswrapper[4612]: E0227 08:06:27.948179 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/keystone-operator-controller-manager-b4d948c87-2hm5l" podUID="259df4bd-af1b-4c61-a5bd-0ca28ac531d4" Feb 27 08:06:28 crc kubenswrapper[4612]: E0227 08:06:28.167491 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/keystone-operator@sha256:c6ad383f55f955902b074d1ee947a2233a5fcbf40698479ae693ce056c80dcc1\\\"\"" pod="openstack-operators/keystone-operator-controller-manager-b4d948c87-2hm5l" podUID="259df4bd-af1b-4c61-a5bd-0ca28ac531d4" Feb 27 08:06:29 crc kubenswrapper[4612]: I0227 08:06:29.730669 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/98296d45-b487-47c3-a44a-fa63ba52426b-cert\") pod \"infra-operator-controller-manager-79d975b745-plm9k\" (UID: \"98296d45-b487-47c3-a44a-fa63ba52426b\") " pod="openstack-operators/infra-operator-controller-manager-79d975b745-plm9k" Feb 27 08:06:29 crc kubenswrapper[4612]: I0227 08:06:29.737130 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/98296d45-b487-47c3-a44a-fa63ba52426b-cert\") pod \"infra-operator-controller-manager-79d975b745-plm9k\" (UID: \"98296d45-b487-47c3-a44a-fa63ba52426b\") " pod="openstack-operators/infra-operator-controller-manager-79d975b745-plm9k" Feb 27 08:06:29 crc kubenswrapper[4612]: I0227 08:06:29.817303 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-g2x8p" Feb 27 08:06:29 crc kubenswrapper[4612]: I0227 08:06:29.825545 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-79d975b745-plm9k" Feb 27 08:06:30 crc kubenswrapper[4612]: I0227 08:06:30.036205 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/cf5ef40d-5e5c-4b2c-9570-63e552800b4e-cert\") pod \"openstack-baremetal-operator-controller-manager-7c6767dc9clp9dx\" (UID: \"cf5ef40d-5e5c-4b2c-9570-63e552800b4e\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9clp9dx" Feb 27 08:06:30 crc kubenswrapper[4612]: I0227 08:06:30.040351 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/cf5ef40d-5e5c-4b2c-9570-63e552800b4e-cert\") pod \"openstack-baremetal-operator-controller-manager-7c6767dc9clp9dx\" (UID: \"cf5ef40d-5e5c-4b2c-9570-63e552800b4e\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9clp9dx" Feb 27 08:06:30 crc kubenswrapper[4612]: I0227 08:06:30.209764 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-nhbw6" Feb 27 08:06:30 crc kubenswrapper[4612]: I0227 08:06:30.218011 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9clp9dx" Feb 27 08:06:30 crc kubenswrapper[4612]: I0227 08:06:30.443217 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/41aed211-3a1f-4f7c-86ca-53e445dcd585-webhook-certs\") pod \"openstack-operator-controller-manager-69f9884c8f-v4kkz\" (UID: \"41aed211-3a1f-4f7c-86ca-53e445dcd585\") " pod="openstack-operators/openstack-operator-controller-manager-69f9884c8f-v4kkz" Feb 27 08:06:30 crc kubenswrapper[4612]: I0227 08:06:30.443359 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/41aed211-3a1f-4f7c-86ca-53e445dcd585-metrics-certs\") pod \"openstack-operator-controller-manager-69f9884c8f-v4kkz\" (UID: \"41aed211-3a1f-4f7c-86ca-53e445dcd585\") " pod="openstack-operators/openstack-operator-controller-manager-69f9884c8f-v4kkz" Feb 27 08:06:30 crc kubenswrapper[4612]: I0227 08:06:30.449294 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/41aed211-3a1f-4f7c-86ca-53e445dcd585-webhook-certs\") pod \"openstack-operator-controller-manager-69f9884c8f-v4kkz\" (UID: \"41aed211-3a1f-4f7c-86ca-53e445dcd585\") " pod="openstack-operators/openstack-operator-controller-manager-69f9884c8f-v4kkz" Feb 27 08:06:30 crc kubenswrapper[4612]: I0227 08:06:30.449681 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/41aed211-3a1f-4f7c-86ca-53e445dcd585-metrics-certs\") pod \"openstack-operator-controller-manager-69f9884c8f-v4kkz\" (UID: \"41aed211-3a1f-4f7c-86ca-53e445dcd585\") " pod="openstack-operators/openstack-operator-controller-manager-69f9884c8f-v4kkz" Feb 27 08:06:30 crc kubenswrapper[4612]: I0227 08:06:30.672485 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-kmvsq" Feb 27 08:06:30 crc kubenswrapper[4612]: I0227 08:06:30.681078 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-69f9884c8f-v4kkz" Feb 27 08:06:34 crc kubenswrapper[4612]: I0227 08:06:34.130009 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-79d975b745-plm9k"] Feb 27 08:06:34 crc kubenswrapper[4612]: I0227 08:06:34.189364 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-69f9884c8f-v4kkz"] Feb 27 08:06:34 crc kubenswrapper[4612]: W0227 08:06:34.204892 4612 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod98296d45_b487_47c3_a44a_fa63ba52426b.slice/crio-2a678e8afcc4518f3d05e2939dbc4eeb89d61c9be1e64195e6f15c748f9b41c9 WatchSource:0}: Error finding container 2a678e8afcc4518f3d05e2939dbc4eeb89d61c9be1e64195e6f15c748f9b41c9: Status 404 returned error can't find the container with id 2a678e8afcc4518f3d05e2939dbc4eeb89d61c9be1e64195e6f15c748f9b41c9 Feb 27 08:06:34 crc kubenswrapper[4612]: I0227 08:06:34.215339 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-589c568786-qclqp" event={"ID":"ee3f21f6-2f91-45c6-8c80-5fb8f89a41ce","Type":"ContainerStarted","Data":"08fec9a5723580612cee0111ababe9a4e67c6a284d6b04589f52e388d2a8466e"} Feb 27 08:06:34 crc kubenswrapper[4612]: I0227 08:06:34.215422 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-589c568786-qclqp" Feb 27 08:06:34 crc kubenswrapper[4612]: I0227 08:06:34.223882 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-784b5bb6c5-g4p5v" event={"ID":"71c482bc-7818-4193-a80b-323682cdcfb3","Type":"ContainerStarted","Data":"ede266f309d08d306fbcdd99c1e74b8839c5f36d19924957ada40e16b0778386"} Feb 27 08:06:34 crc kubenswrapper[4612]: I0227 08:06:34.224586 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-784b5bb6c5-g4p5v" Feb 27 08:06:34 crc kubenswrapper[4612]: I0227 08:06:34.232235 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-868647ff47-mlgz7" event={"ID":"2ddc15ea-b7c1-44e7-b795-392890e9c14e","Type":"ContainerStarted","Data":"e3ba1dab1e97a051c61b7c0bd15736192b5555ba9fe281da52728794e670b559"} Feb 27 08:06:34 crc kubenswrapper[4612]: I0227 08:06:34.233388 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-868647ff47-mlgz7" Feb 27 08:06:34 crc kubenswrapper[4612]: I0227 08:06:34.237437 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-589c568786-qclqp" podStartSLOduration=8.23279186 podStartE2EDuration="37.237421656s" podCreationTimestamp="2026-02-27 08:05:57 +0000 UTC" firstStartedPulling="2026-02-27 08:05:59.936373656 +0000 UTC m=+1017.790303654" lastFinishedPulling="2026-02-27 08:06:28.941003452 +0000 UTC m=+1046.794933450" observedRunningTime="2026-02-27 08:06:34.236608923 +0000 UTC m=+1052.090538921" watchObservedRunningTime="2026-02-27 08:06:34.237421656 +0000 UTC m=+1052.091351654" Feb 27 08:06:34 crc kubenswrapper[4612]: I0227 08:06:34.262303 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9clp9dx"] Feb 27 08:06:34 crc kubenswrapper[4612]: I0227 08:06:34.263663 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-784b5bb6c5-g4p5v" podStartSLOduration=7.425744279 podStartE2EDuration="37.263647968s" podCreationTimestamp="2026-02-27 08:05:57 +0000 UTC" firstStartedPulling="2026-02-27 08:05:59.103107603 +0000 UTC m=+1016.957037601" lastFinishedPulling="2026-02-27 08:06:28.941011292 +0000 UTC m=+1046.794941290" observedRunningTime="2026-02-27 08:06:34.25360397 +0000 UTC m=+1052.107533968" watchObservedRunningTime="2026-02-27 08:06:34.263647968 +0000 UTC m=+1052.117577966" Feb 27 08:06:34 crc kubenswrapper[4612]: I0227 08:06:34.273305 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-868647ff47-mlgz7" podStartSLOduration=2.578484696 podStartE2EDuration="37.273287654s" podCreationTimestamp="2026-02-27 08:05:57 +0000 UTC" firstStartedPulling="2026-02-27 08:05:59.024711558 +0000 UTC m=+1016.878641556" lastFinishedPulling="2026-02-27 08:06:33.719514506 +0000 UTC m=+1051.573444514" observedRunningTime="2026-02-27 08:06:34.266986114 +0000 UTC m=+1052.120916112" watchObservedRunningTime="2026-02-27 08:06:34.273287654 +0000 UTC m=+1052.127217652" Feb 27 08:06:34 crc kubenswrapper[4612]: W0227 08:06:34.277275 4612 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod41aed211_3a1f_4f7c_86ca_53e445dcd585.slice/crio-f53f1d6aec7e7aa911c84c70e134e75b852139bfea0944cdf4f0d5c2ab16b20d WatchSource:0}: Error finding container f53f1d6aec7e7aa911c84c70e134e75b852139bfea0944cdf4f0d5c2ab16b20d: Status 404 returned error can't find the container with id f53f1d6aec7e7aa911c84c70e134e75b852139bfea0944cdf4f0d5c2ab16b20d Feb 27 08:06:34 crc kubenswrapper[4612]: W0227 08:06:34.361186 4612 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcf5ef40d_5e5c_4b2c_9570_63e552800b4e.slice/crio-f4552a46bb97308999058b074a3945fd02489b4d501b443ee71b83c45f12456e WatchSource:0}: Error finding container f4552a46bb97308999058b074a3945fd02489b4d501b443ee71b83c45f12456e: Status 404 returned error can't find the container with id f4552a46bb97308999058b074a3945fd02489b4d501b443ee71b83c45f12456e Feb 27 08:06:35 crc kubenswrapper[4612]: I0227 08:06:35.275262 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-68f46476f-r8lqs" event={"ID":"713a994e-490d-47b4-86ab-1203a6b806fc","Type":"ContainerStarted","Data":"dc49bb94571ca911e0629e1b7cfc9e11d1864e4e1922cd036d541d4ee83ce8f9"} Feb 27 08:06:35 crc kubenswrapper[4612]: I0227 08:06:35.276531 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-68f46476f-r8lqs" Feb 27 08:06:35 crc kubenswrapper[4612]: I0227 08:06:35.293997 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5dc6794d5b-s2jnh" event={"ID":"ad4fca59-6a87-4ebc-8681-d5a3a40f0b82","Type":"ContainerStarted","Data":"efa7bc89d47affc760f7ce9d08972c13de6347a47f8a46d6c48534ceaeef0693"} Feb 27 08:06:35 crc kubenswrapper[4612]: I0227 08:06:35.294826 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-5dc6794d5b-s2jnh" Feb 27 08:06:35 crc kubenswrapper[4612]: I0227 08:06:35.303196 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-6bd4687957-2qmzb" event={"ID":"f292b2e2-2b32-4e4e-bfca-d5db60f0b358","Type":"ContainerStarted","Data":"02094d143cac9998a02f5c79d92e5b0778c1a4cdc80b419dcab4bf6334becf74"} Feb 27 08:06:35 crc kubenswrapper[4612]: I0227 08:06:35.303883 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-6bd4687957-2qmzb" Feb 27 08:06:35 crc kubenswrapper[4612]: I0227 08:06:35.305525 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-6994f66f48-qs4l7" event={"ID":"f62fea1f-71de-4872-9baa-42ad2448d5f5","Type":"ContainerStarted","Data":"bef18ee613b84dc0b173cf13d43bfd1910f2412d8af7c765b8e53a95674b6a8d"} Feb 27 08:06:35 crc kubenswrapper[4612]: I0227 08:06:35.305966 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-6994f66f48-qs4l7" Feb 27 08:06:35 crc kubenswrapper[4612]: I0227 08:06:35.310838 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-55d77d7b5c-462m7" event={"ID":"36fc734e-4b3e-46bb-b914-896b1ec5c2ee","Type":"ContainerStarted","Data":"2dbdb1a79d57c7ebc3139d1c220af601887ba3f05f992fcf840e90b7f18dd584"} Feb 27 08:06:35 crc kubenswrapper[4612]: I0227 08:06:35.311145 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-55d77d7b5c-462m7" Feb 27 08:06:35 crc kubenswrapper[4612]: I0227 08:06:35.322001 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-6d8bf5c495-t95hp" event={"ID":"63d519c2-d2f9-4349-ad82-0399e472d78a","Type":"ContainerStarted","Data":"f138be5939cb4791bce50bb3e92ad87a7fa56260917cb4578a217e6552138b13"} Feb 27 08:06:35 crc kubenswrapper[4612]: I0227 08:06:35.322566 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-6d8bf5c495-t95hp" Feb 27 08:06:35 crc kubenswrapper[4612]: I0227 08:06:35.330863 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-567668f5cf-tjqwp" event={"ID":"d84adbda-b53c-46c7-9196-682910a0009c","Type":"ContainerStarted","Data":"09225895c8c3acab582da74fd33a9b032e8f02e39472466433bb07204835aec0"} Feb 27 08:06:35 crc kubenswrapper[4612]: I0227 08:06:35.331641 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-567668f5cf-tjqwp" Feb 27 08:06:35 crc kubenswrapper[4612]: I0227 08:06:35.361759 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-69f9884c8f-v4kkz" event={"ID":"41aed211-3a1f-4f7c-86ca-53e445dcd585","Type":"ContainerStarted","Data":"101d91f0f5c96291996615d665b85d81472f141f23bc491eddb4037494a19b94"} Feb 27 08:06:35 crc kubenswrapper[4612]: I0227 08:06:35.361803 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-69f9884c8f-v4kkz" event={"ID":"41aed211-3a1f-4f7c-86ca-53e445dcd585","Type":"ContainerStarted","Data":"f53f1d6aec7e7aa911c84c70e134e75b852139bfea0944cdf4f0d5c2ab16b20d"} Feb 27 08:06:35 crc kubenswrapper[4612]: I0227 08:06:35.362383 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-69f9884c8f-v4kkz" Feb 27 08:06:35 crc kubenswrapper[4612]: I0227 08:06:35.373536 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-68f46476f-r8lqs" podStartSLOduration=4.636195812 podStartE2EDuration="38.37352136s" podCreationTimestamp="2026-02-27 08:05:57 +0000 UTC" firstStartedPulling="2026-02-27 08:05:59.978290596 +0000 UTC m=+1017.832220594" lastFinishedPulling="2026-02-27 08:06:33.715616134 +0000 UTC m=+1051.569546142" observedRunningTime="2026-02-27 08:06:35.343082188 +0000 UTC m=+1053.197012196" watchObservedRunningTime="2026-02-27 08:06:35.37352136 +0000 UTC m=+1053.227451348" Feb 27 08:06:35 crc kubenswrapper[4612]: I0227 08:06:35.375809 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-6994f66f48-qs4l7" podStartSLOduration=4.641113503 podStartE2EDuration="38.375800066s" podCreationTimestamp="2026-02-27 08:05:57 +0000 UTC" firstStartedPulling="2026-02-27 08:05:59.980944862 +0000 UTC m=+1017.834874870" lastFinishedPulling="2026-02-27 08:06:33.715631435 +0000 UTC m=+1051.569561433" observedRunningTime="2026-02-27 08:06:35.365985724 +0000 UTC m=+1053.219915722" watchObservedRunningTime="2026-02-27 08:06:35.375800066 +0000 UTC m=+1053.229730064" Feb 27 08:06:35 crc kubenswrapper[4612]: I0227 08:06:35.381941 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-8s6n2" event={"ID":"57ba67ae-3e41-4e2b-9ee6-30b2643805dd","Type":"ContainerStarted","Data":"cd72879c385fc1b612475cbcd7aff5fd0f5de3dfa7a06625d7106e3e74a49d0e"} Feb 27 08:06:35 crc kubenswrapper[4612]: I0227 08:06:35.388813 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9clp9dx" event={"ID":"cf5ef40d-5e5c-4b2c-9570-63e552800b4e","Type":"ContainerStarted","Data":"f4552a46bb97308999058b074a3945fd02489b4d501b443ee71b83c45f12456e"} Feb 27 08:06:35 crc kubenswrapper[4612]: I0227 08:06:35.398498 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-5955d8c787-q74ht" event={"ID":"e0304d11-114b-4aa3-86ca-1747c60b5bfc","Type":"ContainerStarted","Data":"a31c1dcf3a8c87f7849cd23fc9e10e59126f6557717031f029bfdaf16dc3512a"} Feb 27 08:06:35 crc kubenswrapper[4612]: I0227 08:06:35.399301 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-5955d8c787-q74ht" Feb 27 08:06:35 crc kubenswrapper[4612]: I0227 08:06:35.427535 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536326-pjlzf" event={"ID":"3aec7965-4aeb-432f-8b2a-f17d58427093","Type":"ContainerStarted","Data":"a0cfe92dc9f1a9bddbcb67a210c2a83330e1fdb4e95bccc2b9d8a7c4b47dd5aa"} Feb 27 08:06:35 crc kubenswrapper[4612]: I0227 08:06:35.459892 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-67d996989d-4dhhw" event={"ID":"b60d4209-cd53-4cfe-99f7-7e7351041857","Type":"ContainerStarted","Data":"bfee9a26a32bb51de08d0e41a481f3ce78724a7d6cb71e00711254f3825ab4ac"} Feb 27 08:06:35 crc kubenswrapper[4612]: I0227 08:06:35.460454 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-67d996989d-4dhhw" Feb 27 08:06:35 crc kubenswrapper[4612]: I0227 08:06:35.470039 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-79d975b745-plm9k" event={"ID":"98296d45-b487-47c3-a44a-fa63ba52426b","Type":"ContainerStarted","Data":"2a678e8afcc4518f3d05e2939dbc4eeb89d61c9be1e64195e6f15c748f9b41c9"} Feb 27 08:06:35 crc kubenswrapper[4612]: I0227 08:06:35.479113 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-6bd4687957-2qmzb" podStartSLOduration=4.800510758 podStartE2EDuration="38.479095015s" podCreationTimestamp="2026-02-27 08:05:57 +0000 UTC" firstStartedPulling="2026-02-27 08:06:00.0500544 +0000 UTC m=+1017.903984398" lastFinishedPulling="2026-02-27 08:06:33.728638657 +0000 UTC m=+1051.582568655" observedRunningTime="2026-02-27 08:06:35.426208 +0000 UTC m=+1053.280138008" watchObservedRunningTime="2026-02-27 08:06:35.479095015 +0000 UTC m=+1053.333025013" Feb 27 08:06:35 crc kubenswrapper[4612]: I0227 08:06:35.524129 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-6d8bf5c495-t95hp" podStartSLOduration=4.0450004 podStartE2EDuration="38.524111855s" podCreationTimestamp="2026-02-27 08:05:57 +0000 UTC" firstStartedPulling="2026-02-27 08:05:59.322251976 +0000 UTC m=+1017.176181974" lastFinishedPulling="2026-02-27 08:06:33.801363431 +0000 UTC m=+1051.655293429" observedRunningTime="2026-02-27 08:06:35.480055163 +0000 UTC m=+1053.333985161" watchObservedRunningTime="2026-02-27 08:06:35.524111855 +0000 UTC m=+1053.378041853" Feb 27 08:06:35 crc kubenswrapper[4612]: I0227 08:06:35.556803 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-5dc6794d5b-s2jnh" podStartSLOduration=3.837112459 podStartE2EDuration="37.556787412s" podCreationTimestamp="2026-02-27 08:05:58 +0000 UTC" firstStartedPulling="2026-02-27 08:05:59.996487147 +0000 UTC m=+1017.850417145" lastFinishedPulling="2026-02-27 08:06:33.71616209 +0000 UTC m=+1051.570092098" observedRunningTime="2026-02-27 08:06:35.552009465 +0000 UTC m=+1053.405939463" watchObservedRunningTime="2026-02-27 08:06:35.556787412 +0000 UTC m=+1053.410717410" Feb 27 08:06:35 crc kubenswrapper[4612]: I0227 08:06:35.558348 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-567668f5cf-tjqwp" podStartSLOduration=4.849285227 podStartE2EDuration="38.558340096s" podCreationTimestamp="2026-02-27 08:05:57 +0000 UTC" firstStartedPulling="2026-02-27 08:06:00.010463017 +0000 UTC m=+1017.864393015" lastFinishedPulling="2026-02-27 08:06:33.719517886 +0000 UTC m=+1051.573447884" observedRunningTime="2026-02-27 08:06:35.522073787 +0000 UTC m=+1053.376003785" watchObservedRunningTime="2026-02-27 08:06:35.558340096 +0000 UTC m=+1053.412270094" Feb 27 08:06:35 crc kubenswrapper[4612]: I0227 08:06:35.635042 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-55d77d7b5c-462m7" podStartSLOduration=4.904666334 podStartE2EDuration="38.635024034s" podCreationTimestamp="2026-02-27 08:05:57 +0000 UTC" firstStartedPulling="2026-02-27 08:05:59.985716678 +0000 UTC m=+1017.839646676" lastFinishedPulling="2026-02-27 08:06:33.716074378 +0000 UTC m=+1051.570004376" observedRunningTime="2026-02-27 08:06:35.624327227 +0000 UTC m=+1053.478257215" watchObservedRunningTime="2026-02-27 08:06:35.635024034 +0000 UTC m=+1053.488954032" Feb 27 08:06:35 crc kubenswrapper[4612]: I0227 08:06:35.672380 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-8s6n2" podStartSLOduration=8.690250512 podStartE2EDuration="37.672359823s" podCreationTimestamp="2026-02-27 08:05:58 +0000 UTC" firstStartedPulling="2026-02-27 08:05:59.958895361 +0000 UTC m=+1017.812825359" lastFinishedPulling="2026-02-27 08:06:28.941004672 +0000 UTC m=+1046.794934670" observedRunningTime="2026-02-27 08:06:35.665293771 +0000 UTC m=+1053.519223769" watchObservedRunningTime="2026-02-27 08:06:35.672359823 +0000 UTC m=+1053.526289821" Feb 27 08:06:35 crc kubenswrapper[4612]: I0227 08:06:35.710290 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-69f9884c8f-v4kkz" podStartSLOduration=37.71027137 podStartE2EDuration="37.71027137s" podCreationTimestamp="2026-02-27 08:05:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 08:06:35.699305635 +0000 UTC m=+1053.553235633" watchObservedRunningTime="2026-02-27 08:06:35.71027137 +0000 UTC m=+1053.564201368" Feb 27 08:06:35 crc kubenswrapper[4612]: I0227 08:06:35.736198 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-67d996989d-4dhhw" podStartSLOduration=4.514506248 podStartE2EDuration="38.736178112s" podCreationTimestamp="2026-02-27 08:05:57 +0000 UTC" firstStartedPulling="2026-02-27 08:05:59.496605327 +0000 UTC m=+1017.350535325" lastFinishedPulling="2026-02-27 08:06:33.718277171 +0000 UTC m=+1051.572207189" observedRunningTime="2026-02-27 08:06:35.734857874 +0000 UTC m=+1053.588787882" watchObservedRunningTime="2026-02-27 08:06:35.736178112 +0000 UTC m=+1053.590108110" Feb 27 08:06:35 crc kubenswrapper[4612]: I0227 08:06:35.769392 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-5955d8c787-q74ht" podStartSLOduration=5.0916877209999996 podStartE2EDuration="38.769373343s" podCreationTimestamp="2026-02-27 08:05:57 +0000 UTC" firstStartedPulling="2026-02-27 08:06:00.040761664 +0000 UTC m=+1017.894691662" lastFinishedPulling="2026-02-27 08:06:33.718447266 +0000 UTC m=+1051.572377284" observedRunningTime="2026-02-27 08:06:35.764920886 +0000 UTC m=+1053.618850884" watchObservedRunningTime="2026-02-27 08:06:35.769373343 +0000 UTC m=+1053.623303351" Feb 27 08:06:35 crc kubenswrapper[4612]: I0227 08:06:35.808425 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29536326-pjlzf" podStartSLOduration=3.320374589 podStartE2EDuration="35.808409912s" podCreationTimestamp="2026-02-27 08:06:00 +0000 UTC" firstStartedPulling="2026-02-27 08:06:01.066357972 +0000 UTC m=+1018.920287970" lastFinishedPulling="2026-02-27 08:06:33.554393255 +0000 UTC m=+1051.408323293" observedRunningTime="2026-02-27 08:06:35.804650724 +0000 UTC m=+1053.658580732" watchObservedRunningTime="2026-02-27 08:06:35.808409912 +0000 UTC m=+1053.662339910" Feb 27 08:06:36 crc kubenswrapper[4612]: I0227 08:06:36.481064 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-659dc6bbfc-vlczg" event={"ID":"bd923861-d2ad-4a06-b0c7-613605afefbb","Type":"ContainerStarted","Data":"6be85674cd764def7e9fbc1b87fb84a13079bae459ac4a6ad589d4a43d2240ae"} Feb 27 08:06:36 crc kubenswrapper[4612]: I0227 08:06:36.482584 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-659dc6bbfc-vlczg" Feb 27 08:06:36 crc kubenswrapper[4612]: I0227 08:06:36.514253 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-659dc6bbfc-vlczg" podStartSLOduration=5.701742177 podStartE2EDuration="39.514234657s" podCreationTimestamp="2026-02-27 08:05:57 +0000 UTC" firstStartedPulling="2026-02-27 08:05:59.903962228 +0000 UTC m=+1017.757892226" lastFinishedPulling="2026-02-27 08:06:33.716454708 +0000 UTC m=+1051.570384706" observedRunningTime="2026-02-27 08:06:36.51156474 +0000 UTC m=+1054.365494748" watchObservedRunningTime="2026-02-27 08:06:36.514234657 +0000 UTC m=+1054.368164655" Feb 27 08:06:37 crc kubenswrapper[4612]: I0227 08:06:37.508614 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-554564d7fc-8kpwf" event={"ID":"bd687bfa-c843-4d11-80f9-cedcdce41e3d","Type":"ContainerStarted","Data":"a9af7303d76a1891f9069bc6fc24371cf4ffe6d60fb910226a1a7881857d7a25"} Feb 27 08:06:37 crc kubenswrapper[4612]: I0227 08:06:37.510139 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-554564d7fc-8kpwf" Feb 27 08:06:37 crc kubenswrapper[4612]: I0227 08:06:37.515479 4612 generic.go:334] "Generic (PLEG): container finished" podID="3aec7965-4aeb-432f-8b2a-f17d58427093" containerID="a0cfe92dc9f1a9bddbcb67a210c2a83330e1fdb4e95bccc2b9d8a7c4b47dd5aa" exitCode=0 Feb 27 08:06:37 crc kubenswrapper[4612]: I0227 08:06:37.516446 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536326-pjlzf" event={"ID":"3aec7965-4aeb-432f-8b2a-f17d58427093","Type":"ContainerDied","Data":"a0cfe92dc9f1a9bddbcb67a210c2a83330e1fdb4e95bccc2b9d8a7c4b47dd5aa"} Feb 27 08:06:37 crc kubenswrapper[4612]: I0227 08:06:37.533355 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-554564d7fc-8kpwf" podStartSLOduration=3.504497858 podStartE2EDuration="40.533337067s" podCreationTimestamp="2026-02-27 08:05:57 +0000 UTC" firstStartedPulling="2026-02-27 08:05:59.461308337 +0000 UTC m=+1017.315238325" lastFinishedPulling="2026-02-27 08:06:36.490147536 +0000 UTC m=+1054.344077534" observedRunningTime="2026-02-27 08:06:37.529226019 +0000 UTC m=+1055.383156057" watchObservedRunningTime="2026-02-27 08:06:37.533337067 +0000 UTC m=+1055.387267065" Feb 27 08:06:38 crc kubenswrapper[4612]: I0227 08:06:38.931260 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536326-pjlzf" Feb 27 08:06:39 crc kubenswrapper[4612]: I0227 08:06:39.100403 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kdslx\" (UniqueName: \"kubernetes.io/projected/3aec7965-4aeb-432f-8b2a-f17d58427093-kube-api-access-kdslx\") pod \"3aec7965-4aeb-432f-8b2a-f17d58427093\" (UID: \"3aec7965-4aeb-432f-8b2a-f17d58427093\") " Feb 27 08:06:39 crc kubenswrapper[4612]: I0227 08:06:39.120973 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3aec7965-4aeb-432f-8b2a-f17d58427093-kube-api-access-kdslx" (OuterVolumeSpecName: "kube-api-access-kdslx") pod "3aec7965-4aeb-432f-8b2a-f17d58427093" (UID: "3aec7965-4aeb-432f-8b2a-f17d58427093"). InnerVolumeSpecName "kube-api-access-kdslx". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 08:06:39 crc kubenswrapper[4612]: I0227 08:06:39.202085 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kdslx\" (UniqueName: \"kubernetes.io/projected/3aec7965-4aeb-432f-8b2a-f17d58427093-kube-api-access-kdslx\") on node \"crc\" DevicePath \"\"" Feb 27 08:06:39 crc kubenswrapper[4612]: I0227 08:06:39.530769 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-8497b45c89-dvq8j" event={"ID":"8414b96c-7e5b-4824-998e-549a0c43d9d5","Type":"ContainerStarted","Data":"300337eeeb8664d80b399123b95367aa6d132f61cbf17a372badb47fea1536cb"} Feb 27 08:06:39 crc kubenswrapper[4612]: I0227 08:06:39.531456 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-8497b45c89-dvq8j" Feb 27 08:06:39 crc kubenswrapper[4612]: I0227 08:06:39.532802 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536326-pjlzf" event={"ID":"3aec7965-4aeb-432f-8b2a-f17d58427093","Type":"ContainerDied","Data":"c3fce31c1331a7be792f6e0230f8814237f92ec0aa01baa28d206a188ff44377"} Feb 27 08:06:39 crc kubenswrapper[4612]: I0227 08:06:39.532840 4612 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c3fce31c1331a7be792f6e0230f8814237f92ec0aa01baa28d206a188ff44377" Feb 27 08:06:39 crc kubenswrapper[4612]: I0227 08:06:39.532892 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536326-pjlzf" Feb 27 08:06:39 crc kubenswrapper[4612]: I0227 08:06:39.535383 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-79d975b745-plm9k" event={"ID":"98296d45-b487-47c3-a44a-fa63ba52426b","Type":"ContainerStarted","Data":"3b328dbc70500ef499dfbf852bc78c6a47b0416ed23e8f8aa65dac39e4ac1503"} Feb 27 08:06:39 crc kubenswrapper[4612]: I0227 08:06:39.535503 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-79d975b745-plm9k" Feb 27 08:06:39 crc kubenswrapper[4612]: I0227 08:06:39.536769 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9clp9dx" event={"ID":"cf5ef40d-5e5c-4b2c-9570-63e552800b4e","Type":"ContainerStarted","Data":"362ac5b4c4d7515dbed3a2dd328544c29a4234190324df37f764f5fef4d40b14"} Feb 27 08:06:39 crc kubenswrapper[4612]: I0227 08:06:39.536942 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9clp9dx" Feb 27 08:06:39 crc kubenswrapper[4612]: I0227 08:06:39.546353 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-8497b45c89-dvq8j" podStartSLOduration=2.904894521 podStartE2EDuration="42.546334938s" podCreationTimestamp="2026-02-27 08:05:57 +0000 UTC" firstStartedPulling="2026-02-27 08:05:59.676342292 +0000 UTC m=+1017.530272290" lastFinishedPulling="2026-02-27 08:06:39.317782719 +0000 UTC m=+1057.171712707" observedRunningTime="2026-02-27 08:06:39.545889655 +0000 UTC m=+1057.399819653" watchObservedRunningTime="2026-02-27 08:06:39.546334938 +0000 UTC m=+1057.400264936" Feb 27 08:06:39 crc kubenswrapper[4612]: I0227 08:06:39.562793 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-79d975b745-plm9k" podStartSLOduration=37.862308391 podStartE2EDuration="42.562770369s" podCreationTimestamp="2026-02-27 08:05:57 +0000 UTC" firstStartedPulling="2026-02-27 08:06:34.208147137 +0000 UTC m=+1052.062077135" lastFinishedPulling="2026-02-27 08:06:38.908609115 +0000 UTC m=+1056.762539113" observedRunningTime="2026-02-27 08:06:39.561123172 +0000 UTC m=+1057.415053190" watchObservedRunningTime="2026-02-27 08:06:39.562770369 +0000 UTC m=+1057.416700367" Feb 27 08:06:39 crc kubenswrapper[4612]: I0227 08:06:39.608407 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9clp9dx" podStartSLOduration=38.059688817 podStartE2EDuration="42.608381516s" podCreationTimestamp="2026-02-27 08:05:57 +0000 UTC" firstStartedPulling="2026-02-27 08:06:34.364449506 +0000 UTC m=+1052.218379504" lastFinishedPulling="2026-02-27 08:06:38.913142205 +0000 UTC m=+1056.767072203" observedRunningTime="2026-02-27 08:06:39.600443089 +0000 UTC m=+1057.454373087" watchObservedRunningTime="2026-02-27 08:06:39.608381516 +0000 UTC m=+1057.462311514" Feb 27 08:06:39 crc kubenswrapper[4612]: I0227 08:06:39.987321 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29536320-n4vgt"] Feb 27 08:06:39 crc kubenswrapper[4612]: I0227 08:06:39.991572 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29536320-n4vgt"] Feb 27 08:06:40 crc kubenswrapper[4612]: I0227 08:06:40.688944 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-69f9884c8f-v4kkz" Feb 27 08:06:40 crc kubenswrapper[4612]: I0227 08:06:40.862976 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fa6e044a-296c-448c-909f-9b81a8d01f15" path="/var/lib/kubelet/pods/fa6e044a-296c-448c-909f-9b81a8d01f15/volumes" Feb 27 08:06:41 crc kubenswrapper[4612]: I0227 08:06:41.552390 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-bccc79885-zdtzv" event={"ID":"a4619f7b-7242-459b-8558-f84c2893ee55","Type":"ContainerStarted","Data":"e5342aa72312445fe67eea8fadf721b1ed2dc1dd986de5a9e49bb212b6cac13e"} Feb 27 08:06:41 crc kubenswrapper[4612]: I0227 08:06:41.552610 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-bccc79885-zdtzv" Feb 27 08:06:41 crc kubenswrapper[4612]: I0227 08:06:41.554084 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-5b9b8895d5-t4rct" event={"ID":"1903598c-2428-4e90-ab81-9a47fa5af9ef","Type":"ContainerStarted","Data":"b9af04e2ba7a82137d544d01cf7c09c1071682caa7bda93cf0d5827be1f540de"} Feb 27 08:06:41 crc kubenswrapper[4612]: I0227 08:06:41.554260 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-5b9b8895d5-t4rct" Feb 27 08:06:41 crc kubenswrapper[4612]: I0227 08:06:41.555804 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-b4d948c87-2hm5l" event={"ID":"259df4bd-af1b-4c61-a5bd-0ca28ac531d4","Type":"ContainerStarted","Data":"29b5b5fdc937ab6ee0e81a5a498e28c56a0d897be6771147beb8137ffea9d86f"} Feb 27 08:06:41 crc kubenswrapper[4612]: I0227 08:06:41.555967 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-b4d948c87-2hm5l" Feb 27 08:06:41 crc kubenswrapper[4612]: I0227 08:06:41.568091 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-bccc79885-zdtzv" podStartSLOduration=3.107486793 podStartE2EDuration="43.568074308s" podCreationTimestamp="2026-02-27 08:05:58 +0000 UTC" firstStartedPulling="2026-02-27 08:05:59.903485344 +0000 UTC m=+1017.757415342" lastFinishedPulling="2026-02-27 08:06:40.364072839 +0000 UTC m=+1058.218002857" observedRunningTime="2026-02-27 08:06:41.565474784 +0000 UTC m=+1059.419404792" watchObservedRunningTime="2026-02-27 08:06:41.568074308 +0000 UTC m=+1059.422004306" Feb 27 08:06:41 crc kubenswrapper[4612]: I0227 08:06:41.610919 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-b4d948c87-2hm5l" podStartSLOduration=3.761373207 podStartE2EDuration="44.610900326s" podCreationTimestamp="2026-02-27 08:05:57 +0000 UTC" firstStartedPulling="2026-02-27 08:05:59.515521688 +0000 UTC m=+1017.369451676" lastFinishedPulling="2026-02-27 08:06:40.365048797 +0000 UTC m=+1058.218978795" observedRunningTime="2026-02-27 08:06:41.590830581 +0000 UTC m=+1059.444760579" watchObservedRunningTime="2026-02-27 08:06:41.610900326 +0000 UTC m=+1059.464830324" Feb 27 08:06:41 crc kubenswrapper[4612]: I0227 08:06:41.612173 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-5b9b8895d5-t4rct" podStartSLOduration=3.866473845 podStartE2EDuration="44.612163792s" podCreationTimestamp="2026-02-27 08:05:57 +0000 UTC" firstStartedPulling="2026-02-27 08:05:59.676080925 +0000 UTC m=+1017.530010923" lastFinishedPulling="2026-02-27 08:06:40.421770862 +0000 UTC m=+1058.275700870" observedRunningTime="2026-02-27 08:06:41.607700374 +0000 UTC m=+1059.461630392" watchObservedRunningTime="2026-02-27 08:06:41.612163792 +0000 UTC m=+1059.466093790" Feb 27 08:06:42 crc kubenswrapper[4612]: I0227 08:06:42.564906 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-69f49c598c-7f9bj" event={"ID":"6065186a-5435-4db3-8adc-698bec4358ef","Type":"ContainerStarted","Data":"f9da5035f5695d14e03b791a1095d5837b4ced00b8bad65e4d4dca4ea3c313be"} Feb 27 08:06:42 crc kubenswrapper[4612]: I0227 08:06:42.590129 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-69f49c598c-7f9bj" podStartSLOduration=3.157876134 podStartE2EDuration="45.590104414s" podCreationTimestamp="2026-02-27 08:05:57 +0000 UTC" firstStartedPulling="2026-02-27 08:05:59.510138974 +0000 UTC m=+1017.364068962" lastFinishedPulling="2026-02-27 08:06:41.942367244 +0000 UTC m=+1059.796297242" observedRunningTime="2026-02-27 08:06:42.587325454 +0000 UTC m=+1060.441255462" watchObservedRunningTime="2026-02-27 08:06:42.590104414 +0000 UTC m=+1060.444034452" Feb 27 08:06:47 crc kubenswrapper[4612]: I0227 08:06:47.855986 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-868647ff47-mlgz7" Feb 27 08:06:47 crc kubenswrapper[4612]: I0227 08:06:47.899752 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-784b5bb6c5-g4p5v" Feb 27 08:06:47 crc kubenswrapper[4612]: I0227 08:06:47.918586 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-6d8bf5c495-t95hp" Feb 27 08:06:48 crc kubenswrapper[4612]: I0227 08:06:48.140200 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-55d77d7b5c-462m7" Feb 27 08:06:48 crc kubenswrapper[4612]: I0227 08:06:48.145373 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-b4d948c87-2hm5l" Feb 27 08:06:48 crc kubenswrapper[4612]: I0227 08:06:48.245572 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-69f49c598c-7f9bj" Feb 27 08:06:48 crc kubenswrapper[4612]: I0227 08:06:48.248505 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-69f49c598c-7f9bj" Feb 27 08:06:48 crc kubenswrapper[4612]: I0227 08:06:48.251614 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-554564d7fc-8kpwf" Feb 27 08:06:48 crc kubenswrapper[4612]: I0227 08:06:48.285308 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-5b9b8895d5-t4rct" Feb 27 08:06:48 crc kubenswrapper[4612]: I0227 08:06:48.334011 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-67d996989d-4dhhw" Feb 27 08:06:48 crc kubenswrapper[4612]: I0227 08:06:48.354961 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-6994f66f48-qs4l7" Feb 27 08:06:48 crc kubenswrapper[4612]: I0227 08:06:48.376734 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-6bd4687957-2qmzb" Feb 27 08:06:48 crc kubenswrapper[4612]: I0227 08:06:48.404002 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-567668f5cf-tjqwp" Feb 27 08:06:48 crc kubenswrapper[4612]: I0227 08:06:48.405547 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-659dc6bbfc-vlczg" Feb 27 08:06:48 crc kubenswrapper[4612]: I0227 08:06:48.437614 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-5955d8c787-q74ht" Feb 27 08:06:48 crc kubenswrapper[4612]: I0227 08:06:48.464391 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-589c568786-qclqp" Feb 27 08:06:48 crc kubenswrapper[4612]: I0227 08:06:48.506639 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-8497b45c89-dvq8j" Feb 27 08:06:48 crc kubenswrapper[4612]: I0227 08:06:48.716569 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-68f46476f-r8lqs" Feb 27 08:06:48 crc kubenswrapper[4612]: I0227 08:06:48.782189 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-5dc6794d5b-s2jnh" Feb 27 08:06:48 crc kubenswrapper[4612]: I0227 08:06:48.828760 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-bccc79885-zdtzv" Feb 27 08:06:49 crc kubenswrapper[4612]: I0227 08:06:49.834618 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-79d975b745-plm9k" Feb 27 08:06:50 crc kubenswrapper[4612]: I0227 08:06:50.228116 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-7c6767dc9clp9dx" Feb 27 08:07:03 crc kubenswrapper[4612]: I0227 08:07:03.876179 4612 scope.go:117] "RemoveContainer" containerID="0f22ee3abd5aaa17b4be1aeea0a2da15f2f6708489f9b954cfb0662bf9fb8a7b" Feb 27 08:07:10 crc kubenswrapper[4612]: I0227 08:07:10.535906 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-7smgg"] Feb 27 08:07:10 crc kubenswrapper[4612]: E0227 08:07:10.536906 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3aec7965-4aeb-432f-8b2a-f17d58427093" containerName="oc" Feb 27 08:07:10 crc kubenswrapper[4612]: I0227 08:07:10.536919 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="3aec7965-4aeb-432f-8b2a-f17d58427093" containerName="oc" Feb 27 08:07:10 crc kubenswrapper[4612]: I0227 08:07:10.537052 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="3aec7965-4aeb-432f-8b2a-f17d58427093" containerName="oc" Feb 27 08:07:10 crc kubenswrapper[4612]: I0227 08:07:10.537899 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-7smgg" Feb 27 08:07:10 crc kubenswrapper[4612]: I0227 08:07:10.539963 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-mw48z" Feb 27 08:07:10 crc kubenswrapper[4612]: I0227 08:07:10.540290 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Feb 27 08:07:10 crc kubenswrapper[4612]: I0227 08:07:10.540818 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Feb 27 08:07:10 crc kubenswrapper[4612]: I0227 08:07:10.547001 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Feb 27 08:07:10 crc kubenswrapper[4612]: I0227 08:07:10.563718 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-7smgg"] Feb 27 08:07:10 crc kubenswrapper[4612]: I0227 08:07:10.678840 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-579pm"] Feb 27 08:07:10 crc kubenswrapper[4612]: I0227 08:07:10.680642 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-579pm" Feb 27 08:07:10 crc kubenswrapper[4612]: I0227 08:07:10.685035 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Feb 27 08:07:10 crc kubenswrapper[4612]: I0227 08:07:10.714332 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pzl2m\" (UniqueName: \"kubernetes.io/projected/e2653a32-b431-40b0-b9de-6a431d7f0644-kube-api-access-pzl2m\") pod \"dnsmasq-dns-675f4bcbfc-7smgg\" (UID: \"e2653a32-b431-40b0-b9de-6a431d7f0644\") " pod="openstack/dnsmasq-dns-675f4bcbfc-7smgg" Feb 27 08:07:10 crc kubenswrapper[4612]: I0227 08:07:10.714390 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/93fcf290-888e-43d5-b46b-0a235776bdbc-config\") pod \"dnsmasq-dns-78dd6ddcc-579pm\" (UID: \"93fcf290-888e-43d5-b46b-0a235776bdbc\") " pod="openstack/dnsmasq-dns-78dd6ddcc-579pm" Feb 27 08:07:10 crc kubenswrapper[4612]: I0227 08:07:10.714430 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/93fcf290-888e-43d5-b46b-0a235776bdbc-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-579pm\" (UID: \"93fcf290-888e-43d5-b46b-0a235776bdbc\") " pod="openstack/dnsmasq-dns-78dd6ddcc-579pm" Feb 27 08:07:10 crc kubenswrapper[4612]: I0227 08:07:10.714472 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2hwtr\" (UniqueName: \"kubernetes.io/projected/93fcf290-888e-43d5-b46b-0a235776bdbc-kube-api-access-2hwtr\") pod \"dnsmasq-dns-78dd6ddcc-579pm\" (UID: \"93fcf290-888e-43d5-b46b-0a235776bdbc\") " pod="openstack/dnsmasq-dns-78dd6ddcc-579pm" Feb 27 08:07:10 crc kubenswrapper[4612]: I0227 08:07:10.714523 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e2653a32-b431-40b0-b9de-6a431d7f0644-config\") pod \"dnsmasq-dns-675f4bcbfc-7smgg\" (UID: \"e2653a32-b431-40b0-b9de-6a431d7f0644\") " pod="openstack/dnsmasq-dns-675f4bcbfc-7smgg" Feb 27 08:07:10 crc kubenswrapper[4612]: I0227 08:07:10.745057 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-579pm"] Feb 27 08:07:10 crc kubenswrapper[4612]: I0227 08:07:10.815163 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e2653a32-b431-40b0-b9de-6a431d7f0644-config\") pod \"dnsmasq-dns-675f4bcbfc-7smgg\" (UID: \"e2653a32-b431-40b0-b9de-6a431d7f0644\") " pod="openstack/dnsmasq-dns-675f4bcbfc-7smgg" Feb 27 08:07:10 crc kubenswrapper[4612]: I0227 08:07:10.815219 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pzl2m\" (UniqueName: \"kubernetes.io/projected/e2653a32-b431-40b0-b9de-6a431d7f0644-kube-api-access-pzl2m\") pod \"dnsmasq-dns-675f4bcbfc-7smgg\" (UID: \"e2653a32-b431-40b0-b9de-6a431d7f0644\") " pod="openstack/dnsmasq-dns-675f4bcbfc-7smgg" Feb 27 08:07:10 crc kubenswrapper[4612]: I0227 08:07:10.815252 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/93fcf290-888e-43d5-b46b-0a235776bdbc-config\") pod \"dnsmasq-dns-78dd6ddcc-579pm\" (UID: \"93fcf290-888e-43d5-b46b-0a235776bdbc\") " pod="openstack/dnsmasq-dns-78dd6ddcc-579pm" Feb 27 08:07:10 crc kubenswrapper[4612]: I0227 08:07:10.815273 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/93fcf290-888e-43d5-b46b-0a235776bdbc-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-579pm\" (UID: \"93fcf290-888e-43d5-b46b-0a235776bdbc\") " pod="openstack/dnsmasq-dns-78dd6ddcc-579pm" Feb 27 08:07:10 crc kubenswrapper[4612]: I0227 08:07:10.815309 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2hwtr\" (UniqueName: \"kubernetes.io/projected/93fcf290-888e-43d5-b46b-0a235776bdbc-kube-api-access-2hwtr\") pod \"dnsmasq-dns-78dd6ddcc-579pm\" (UID: \"93fcf290-888e-43d5-b46b-0a235776bdbc\") " pod="openstack/dnsmasq-dns-78dd6ddcc-579pm" Feb 27 08:07:10 crc kubenswrapper[4612]: I0227 08:07:10.816737 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e2653a32-b431-40b0-b9de-6a431d7f0644-config\") pod \"dnsmasq-dns-675f4bcbfc-7smgg\" (UID: \"e2653a32-b431-40b0-b9de-6a431d7f0644\") " pod="openstack/dnsmasq-dns-675f4bcbfc-7smgg" Feb 27 08:07:10 crc kubenswrapper[4612]: I0227 08:07:10.817266 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/93fcf290-888e-43d5-b46b-0a235776bdbc-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-579pm\" (UID: \"93fcf290-888e-43d5-b46b-0a235776bdbc\") " pod="openstack/dnsmasq-dns-78dd6ddcc-579pm" Feb 27 08:07:10 crc kubenswrapper[4612]: I0227 08:07:10.842950 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2hwtr\" (UniqueName: \"kubernetes.io/projected/93fcf290-888e-43d5-b46b-0a235776bdbc-kube-api-access-2hwtr\") pod \"dnsmasq-dns-78dd6ddcc-579pm\" (UID: \"93fcf290-888e-43d5-b46b-0a235776bdbc\") " pod="openstack/dnsmasq-dns-78dd6ddcc-579pm" Feb 27 08:07:10 crc kubenswrapper[4612]: I0227 08:07:10.860406 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pzl2m\" (UniqueName: \"kubernetes.io/projected/e2653a32-b431-40b0-b9de-6a431d7f0644-kube-api-access-pzl2m\") pod \"dnsmasq-dns-675f4bcbfc-7smgg\" (UID: \"e2653a32-b431-40b0-b9de-6a431d7f0644\") " pod="openstack/dnsmasq-dns-675f4bcbfc-7smgg" Feb 27 08:07:10 crc kubenswrapper[4612]: I0227 08:07:10.864226 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/93fcf290-888e-43d5-b46b-0a235776bdbc-config\") pod \"dnsmasq-dns-78dd6ddcc-579pm\" (UID: \"93fcf290-888e-43d5-b46b-0a235776bdbc\") " pod="openstack/dnsmasq-dns-78dd6ddcc-579pm" Feb 27 08:07:11 crc kubenswrapper[4612]: I0227 08:07:11.016048 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-579pm" Feb 27 08:07:11 crc kubenswrapper[4612]: I0227 08:07:11.157135 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-7smgg" Feb 27 08:07:11 crc kubenswrapper[4612]: I0227 08:07:11.493790 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-579pm"] Feb 27 08:07:11 crc kubenswrapper[4612]: I0227 08:07:11.678257 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-7smgg"] Feb 27 08:07:11 crc kubenswrapper[4612]: W0227 08:07:11.679786 4612 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode2653a32_b431_40b0_b9de_6a431d7f0644.slice/crio-63f5baf13892f51a4eba21cfaa2db5067796cd93aea442bf49b7482c73cf2162 WatchSource:0}: Error finding container 63f5baf13892f51a4eba21cfaa2db5067796cd93aea442bf49b7482c73cf2162: Status 404 returned error can't find the container with id 63f5baf13892f51a4eba21cfaa2db5067796cd93aea442bf49b7482c73cf2162 Feb 27 08:07:11 crc kubenswrapper[4612]: I0227 08:07:11.797433 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-7smgg" event={"ID":"e2653a32-b431-40b0-b9de-6a431d7f0644","Type":"ContainerStarted","Data":"63f5baf13892f51a4eba21cfaa2db5067796cd93aea442bf49b7482c73cf2162"} Feb 27 08:07:11 crc kubenswrapper[4612]: I0227 08:07:11.798800 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-579pm" event={"ID":"93fcf290-888e-43d5-b46b-0a235776bdbc","Type":"ContainerStarted","Data":"bef2679160619fc74b69b6fb9240cff0be2dd3e92d67f536c29de144de780ec3"} Feb 27 08:07:13 crc kubenswrapper[4612]: I0227 08:07:13.369321 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-7smgg"] Feb 27 08:07:13 crc kubenswrapper[4612]: I0227 08:07:13.401492 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-jghh4"] Feb 27 08:07:13 crc kubenswrapper[4612]: I0227 08:07:13.402830 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-jghh4" Feb 27 08:07:13 crc kubenswrapper[4612]: I0227 08:07:13.418310 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-jghh4"] Feb 27 08:07:13 crc kubenswrapper[4612]: I0227 08:07:13.584503 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/57036ec4-7696-4e30-b365-44176b0afdc2-dns-svc\") pod \"dnsmasq-dns-666b6646f7-jghh4\" (UID: \"57036ec4-7696-4e30-b365-44176b0afdc2\") " pod="openstack/dnsmasq-dns-666b6646f7-jghh4" Feb 27 08:07:13 crc kubenswrapper[4612]: I0227 08:07:13.584534 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jssxq\" (UniqueName: \"kubernetes.io/projected/57036ec4-7696-4e30-b365-44176b0afdc2-kube-api-access-jssxq\") pod \"dnsmasq-dns-666b6646f7-jghh4\" (UID: \"57036ec4-7696-4e30-b365-44176b0afdc2\") " pod="openstack/dnsmasq-dns-666b6646f7-jghh4" Feb 27 08:07:13 crc kubenswrapper[4612]: I0227 08:07:13.584577 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/57036ec4-7696-4e30-b365-44176b0afdc2-config\") pod \"dnsmasq-dns-666b6646f7-jghh4\" (UID: \"57036ec4-7696-4e30-b365-44176b0afdc2\") " pod="openstack/dnsmasq-dns-666b6646f7-jghh4" Feb 27 08:07:13 crc kubenswrapper[4612]: I0227 08:07:13.685828 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jssxq\" (UniqueName: \"kubernetes.io/projected/57036ec4-7696-4e30-b365-44176b0afdc2-kube-api-access-jssxq\") pod \"dnsmasq-dns-666b6646f7-jghh4\" (UID: \"57036ec4-7696-4e30-b365-44176b0afdc2\") " pod="openstack/dnsmasq-dns-666b6646f7-jghh4" Feb 27 08:07:13 crc kubenswrapper[4612]: I0227 08:07:13.685869 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/57036ec4-7696-4e30-b365-44176b0afdc2-dns-svc\") pod \"dnsmasq-dns-666b6646f7-jghh4\" (UID: \"57036ec4-7696-4e30-b365-44176b0afdc2\") " pod="openstack/dnsmasq-dns-666b6646f7-jghh4" Feb 27 08:07:13 crc kubenswrapper[4612]: I0227 08:07:13.685909 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/57036ec4-7696-4e30-b365-44176b0afdc2-config\") pod \"dnsmasq-dns-666b6646f7-jghh4\" (UID: \"57036ec4-7696-4e30-b365-44176b0afdc2\") " pod="openstack/dnsmasq-dns-666b6646f7-jghh4" Feb 27 08:07:13 crc kubenswrapper[4612]: I0227 08:07:13.686834 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/57036ec4-7696-4e30-b365-44176b0afdc2-config\") pod \"dnsmasq-dns-666b6646f7-jghh4\" (UID: \"57036ec4-7696-4e30-b365-44176b0afdc2\") " pod="openstack/dnsmasq-dns-666b6646f7-jghh4" Feb 27 08:07:13 crc kubenswrapper[4612]: I0227 08:07:13.687217 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/57036ec4-7696-4e30-b365-44176b0afdc2-dns-svc\") pod \"dnsmasq-dns-666b6646f7-jghh4\" (UID: \"57036ec4-7696-4e30-b365-44176b0afdc2\") " pod="openstack/dnsmasq-dns-666b6646f7-jghh4" Feb 27 08:07:13 crc kubenswrapper[4612]: I0227 08:07:13.724670 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jssxq\" (UniqueName: \"kubernetes.io/projected/57036ec4-7696-4e30-b365-44176b0afdc2-kube-api-access-jssxq\") pod \"dnsmasq-dns-666b6646f7-jghh4\" (UID: \"57036ec4-7696-4e30-b365-44176b0afdc2\") " pod="openstack/dnsmasq-dns-666b6646f7-jghh4" Feb 27 08:07:13 crc kubenswrapper[4612]: I0227 08:07:13.734005 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-jghh4" Feb 27 08:07:13 crc kubenswrapper[4612]: I0227 08:07:13.806591 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-579pm"] Feb 27 08:07:13 crc kubenswrapper[4612]: I0227 08:07:13.861029 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-bz4dn"] Feb 27 08:07:13 crc kubenswrapper[4612]: I0227 08:07:13.877175 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-bz4dn" Feb 27 08:07:13 crc kubenswrapper[4612]: I0227 08:07:13.895500 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/72e3b749-ce9a-4480-8873-d8c9b6495cbc-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-bz4dn\" (UID: \"72e3b749-ce9a-4480-8873-d8c9b6495cbc\") " pod="openstack/dnsmasq-dns-57d769cc4f-bz4dn" Feb 27 08:07:13 crc kubenswrapper[4612]: I0227 08:07:13.895559 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rh2tp\" (UniqueName: \"kubernetes.io/projected/72e3b749-ce9a-4480-8873-d8c9b6495cbc-kube-api-access-rh2tp\") pod \"dnsmasq-dns-57d769cc4f-bz4dn\" (UID: \"72e3b749-ce9a-4480-8873-d8c9b6495cbc\") " pod="openstack/dnsmasq-dns-57d769cc4f-bz4dn" Feb 27 08:07:13 crc kubenswrapper[4612]: I0227 08:07:13.895587 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/72e3b749-ce9a-4480-8873-d8c9b6495cbc-config\") pod \"dnsmasq-dns-57d769cc4f-bz4dn\" (UID: \"72e3b749-ce9a-4480-8873-d8c9b6495cbc\") " pod="openstack/dnsmasq-dns-57d769cc4f-bz4dn" Feb 27 08:07:13 crc kubenswrapper[4612]: I0227 08:07:13.899107 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-bz4dn"] Feb 27 08:07:13 crc kubenswrapper[4612]: I0227 08:07:13.996878 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/72e3b749-ce9a-4480-8873-d8c9b6495cbc-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-bz4dn\" (UID: \"72e3b749-ce9a-4480-8873-d8c9b6495cbc\") " pod="openstack/dnsmasq-dns-57d769cc4f-bz4dn" Feb 27 08:07:13 crc kubenswrapper[4612]: I0227 08:07:13.996926 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rh2tp\" (UniqueName: \"kubernetes.io/projected/72e3b749-ce9a-4480-8873-d8c9b6495cbc-kube-api-access-rh2tp\") pod \"dnsmasq-dns-57d769cc4f-bz4dn\" (UID: \"72e3b749-ce9a-4480-8873-d8c9b6495cbc\") " pod="openstack/dnsmasq-dns-57d769cc4f-bz4dn" Feb 27 08:07:13 crc kubenswrapper[4612]: I0227 08:07:13.996949 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/72e3b749-ce9a-4480-8873-d8c9b6495cbc-config\") pod \"dnsmasq-dns-57d769cc4f-bz4dn\" (UID: \"72e3b749-ce9a-4480-8873-d8c9b6495cbc\") " pod="openstack/dnsmasq-dns-57d769cc4f-bz4dn" Feb 27 08:07:13 crc kubenswrapper[4612]: I0227 08:07:13.998290 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/72e3b749-ce9a-4480-8873-d8c9b6495cbc-config\") pod \"dnsmasq-dns-57d769cc4f-bz4dn\" (UID: \"72e3b749-ce9a-4480-8873-d8c9b6495cbc\") " pod="openstack/dnsmasq-dns-57d769cc4f-bz4dn" Feb 27 08:07:13 crc kubenswrapper[4612]: I0227 08:07:13.998674 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/72e3b749-ce9a-4480-8873-d8c9b6495cbc-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-bz4dn\" (UID: \"72e3b749-ce9a-4480-8873-d8c9b6495cbc\") " pod="openstack/dnsmasq-dns-57d769cc4f-bz4dn" Feb 27 08:07:14 crc kubenswrapper[4612]: I0227 08:07:14.036717 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rh2tp\" (UniqueName: \"kubernetes.io/projected/72e3b749-ce9a-4480-8873-d8c9b6495cbc-kube-api-access-rh2tp\") pod \"dnsmasq-dns-57d769cc4f-bz4dn\" (UID: \"72e3b749-ce9a-4480-8873-d8c9b6495cbc\") " pod="openstack/dnsmasq-dns-57d769cc4f-bz4dn" Feb 27 08:07:14 crc kubenswrapper[4612]: I0227 08:07:14.266330 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-bz4dn" Feb 27 08:07:14 crc kubenswrapper[4612]: I0227 08:07:14.335103 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-jghh4"] Feb 27 08:07:14 crc kubenswrapper[4612]: I0227 08:07:14.547727 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Feb 27 08:07:14 crc kubenswrapper[4612]: I0227 08:07:14.549460 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Feb 27 08:07:14 crc kubenswrapper[4612]: I0227 08:07:14.552667 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Feb 27 08:07:14 crc kubenswrapper[4612]: I0227 08:07:14.553309 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Feb 27 08:07:14 crc kubenswrapper[4612]: I0227 08:07:14.555718 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-r2sww" Feb 27 08:07:14 crc kubenswrapper[4612]: I0227 08:07:14.556472 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Feb 27 08:07:14 crc kubenswrapper[4612]: I0227 08:07:14.556484 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Feb 27 08:07:14 crc kubenswrapper[4612]: I0227 08:07:14.556607 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Feb 27 08:07:14 crc kubenswrapper[4612]: I0227 08:07:14.556939 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Feb 27 08:07:14 crc kubenswrapper[4612]: I0227 08:07:14.588231 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Feb 27 08:07:14 crc kubenswrapper[4612]: I0227 08:07:14.714702 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/0da436b1-803d-41bb-94ab-9cfd6fa769eb-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"0da436b1-803d-41bb-94ab-9cfd6fa769eb\") " pod="openstack/rabbitmq-server-0" Feb 27 08:07:14 crc kubenswrapper[4612]: I0227 08:07:14.714748 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/0da436b1-803d-41bb-94ab-9cfd6fa769eb-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"0da436b1-803d-41bb-94ab-9cfd6fa769eb\") " pod="openstack/rabbitmq-server-0" Feb 27 08:07:14 crc kubenswrapper[4612]: I0227 08:07:14.714778 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vjjpx\" (UniqueName: \"kubernetes.io/projected/0da436b1-803d-41bb-94ab-9cfd6fa769eb-kube-api-access-vjjpx\") pod \"rabbitmq-server-0\" (UID: \"0da436b1-803d-41bb-94ab-9cfd6fa769eb\") " pod="openstack/rabbitmq-server-0" Feb 27 08:07:14 crc kubenswrapper[4612]: I0227 08:07:14.714801 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-server-0\" (UID: \"0da436b1-803d-41bb-94ab-9cfd6fa769eb\") " pod="openstack/rabbitmq-server-0" Feb 27 08:07:14 crc kubenswrapper[4612]: I0227 08:07:14.714818 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/0da436b1-803d-41bb-94ab-9cfd6fa769eb-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"0da436b1-803d-41bb-94ab-9cfd6fa769eb\") " pod="openstack/rabbitmq-server-0" Feb 27 08:07:14 crc kubenswrapper[4612]: I0227 08:07:14.714843 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/0da436b1-803d-41bb-94ab-9cfd6fa769eb-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"0da436b1-803d-41bb-94ab-9cfd6fa769eb\") " pod="openstack/rabbitmq-server-0" Feb 27 08:07:14 crc kubenswrapper[4612]: I0227 08:07:14.714870 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/0da436b1-803d-41bb-94ab-9cfd6fa769eb-pod-info\") pod \"rabbitmq-server-0\" (UID: \"0da436b1-803d-41bb-94ab-9cfd6fa769eb\") " pod="openstack/rabbitmq-server-0" Feb 27 08:07:14 crc kubenswrapper[4612]: I0227 08:07:14.714884 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/0da436b1-803d-41bb-94ab-9cfd6fa769eb-server-conf\") pod \"rabbitmq-server-0\" (UID: \"0da436b1-803d-41bb-94ab-9cfd6fa769eb\") " pod="openstack/rabbitmq-server-0" Feb 27 08:07:14 crc kubenswrapper[4612]: I0227 08:07:14.714905 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/0da436b1-803d-41bb-94ab-9cfd6fa769eb-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"0da436b1-803d-41bb-94ab-9cfd6fa769eb\") " pod="openstack/rabbitmq-server-0" Feb 27 08:07:14 crc kubenswrapper[4612]: I0227 08:07:14.714929 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/0da436b1-803d-41bb-94ab-9cfd6fa769eb-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"0da436b1-803d-41bb-94ab-9cfd6fa769eb\") " pod="openstack/rabbitmq-server-0" Feb 27 08:07:14 crc kubenswrapper[4612]: I0227 08:07:14.714947 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0da436b1-803d-41bb-94ab-9cfd6fa769eb-config-data\") pod \"rabbitmq-server-0\" (UID: \"0da436b1-803d-41bb-94ab-9cfd6fa769eb\") " pod="openstack/rabbitmq-server-0" Feb 27 08:07:14 crc kubenswrapper[4612]: I0227 08:07:14.815803 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/0da436b1-803d-41bb-94ab-9cfd6fa769eb-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"0da436b1-803d-41bb-94ab-9cfd6fa769eb\") " pod="openstack/rabbitmq-server-0" Feb 27 08:07:14 crc kubenswrapper[4612]: I0227 08:07:14.815862 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/0da436b1-803d-41bb-94ab-9cfd6fa769eb-pod-info\") pod \"rabbitmq-server-0\" (UID: \"0da436b1-803d-41bb-94ab-9cfd6fa769eb\") " pod="openstack/rabbitmq-server-0" Feb 27 08:07:14 crc kubenswrapper[4612]: I0227 08:07:14.815887 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/0da436b1-803d-41bb-94ab-9cfd6fa769eb-server-conf\") pod \"rabbitmq-server-0\" (UID: \"0da436b1-803d-41bb-94ab-9cfd6fa769eb\") " pod="openstack/rabbitmq-server-0" Feb 27 08:07:14 crc kubenswrapper[4612]: I0227 08:07:14.815913 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/0da436b1-803d-41bb-94ab-9cfd6fa769eb-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"0da436b1-803d-41bb-94ab-9cfd6fa769eb\") " pod="openstack/rabbitmq-server-0" Feb 27 08:07:14 crc kubenswrapper[4612]: I0227 08:07:14.815952 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/0da436b1-803d-41bb-94ab-9cfd6fa769eb-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"0da436b1-803d-41bb-94ab-9cfd6fa769eb\") " pod="openstack/rabbitmq-server-0" Feb 27 08:07:14 crc kubenswrapper[4612]: I0227 08:07:14.815976 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0da436b1-803d-41bb-94ab-9cfd6fa769eb-config-data\") pod \"rabbitmq-server-0\" (UID: \"0da436b1-803d-41bb-94ab-9cfd6fa769eb\") " pod="openstack/rabbitmq-server-0" Feb 27 08:07:14 crc kubenswrapper[4612]: I0227 08:07:14.816026 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/0da436b1-803d-41bb-94ab-9cfd6fa769eb-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"0da436b1-803d-41bb-94ab-9cfd6fa769eb\") " pod="openstack/rabbitmq-server-0" Feb 27 08:07:14 crc kubenswrapper[4612]: I0227 08:07:14.816049 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/0da436b1-803d-41bb-94ab-9cfd6fa769eb-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"0da436b1-803d-41bb-94ab-9cfd6fa769eb\") " pod="openstack/rabbitmq-server-0" Feb 27 08:07:14 crc kubenswrapper[4612]: I0227 08:07:14.816079 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vjjpx\" (UniqueName: \"kubernetes.io/projected/0da436b1-803d-41bb-94ab-9cfd6fa769eb-kube-api-access-vjjpx\") pod \"rabbitmq-server-0\" (UID: \"0da436b1-803d-41bb-94ab-9cfd6fa769eb\") " pod="openstack/rabbitmq-server-0" Feb 27 08:07:14 crc kubenswrapper[4612]: I0227 08:07:14.816105 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-server-0\" (UID: \"0da436b1-803d-41bb-94ab-9cfd6fa769eb\") " pod="openstack/rabbitmq-server-0" Feb 27 08:07:14 crc kubenswrapper[4612]: I0227 08:07:14.816127 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/0da436b1-803d-41bb-94ab-9cfd6fa769eb-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"0da436b1-803d-41bb-94ab-9cfd6fa769eb\") " pod="openstack/rabbitmq-server-0" Feb 27 08:07:14 crc kubenswrapper[4612]: I0227 08:07:14.816682 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/0da436b1-803d-41bb-94ab-9cfd6fa769eb-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"0da436b1-803d-41bb-94ab-9cfd6fa769eb\") " pod="openstack/rabbitmq-server-0" Feb 27 08:07:14 crc kubenswrapper[4612]: I0227 08:07:14.819351 4612 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-server-0\" (UID: \"0da436b1-803d-41bb-94ab-9cfd6fa769eb\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/rabbitmq-server-0" Feb 27 08:07:14 crc kubenswrapper[4612]: I0227 08:07:14.823875 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/0da436b1-803d-41bb-94ab-9cfd6fa769eb-server-conf\") pod \"rabbitmq-server-0\" (UID: \"0da436b1-803d-41bb-94ab-9cfd6fa769eb\") " pod="openstack/rabbitmq-server-0" Feb 27 08:07:14 crc kubenswrapper[4612]: I0227 08:07:14.846967 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-jghh4" event={"ID":"57036ec4-7696-4e30-b365-44176b0afdc2","Type":"ContainerStarted","Data":"08028231b4e92b3303f61c2ac8b27f9df6050f3da4933accfb44811d6e216955"} Feb 27 08:07:14 crc kubenswrapper[4612]: I0227 08:07:14.854869 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/0da436b1-803d-41bb-94ab-9cfd6fa769eb-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"0da436b1-803d-41bb-94ab-9cfd6fa769eb\") " pod="openstack/rabbitmq-server-0" Feb 27 08:07:14 crc kubenswrapper[4612]: I0227 08:07:14.863515 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0da436b1-803d-41bb-94ab-9cfd6fa769eb-config-data\") pod \"rabbitmq-server-0\" (UID: \"0da436b1-803d-41bb-94ab-9cfd6fa769eb\") " pod="openstack/rabbitmq-server-0" Feb 27 08:07:14 crc kubenswrapper[4612]: I0227 08:07:14.871910 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/0da436b1-803d-41bb-94ab-9cfd6fa769eb-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"0da436b1-803d-41bb-94ab-9cfd6fa769eb\") " pod="openstack/rabbitmq-server-0" Feb 27 08:07:14 crc kubenswrapper[4612]: I0227 08:07:14.872718 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/0da436b1-803d-41bb-94ab-9cfd6fa769eb-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"0da436b1-803d-41bb-94ab-9cfd6fa769eb\") " pod="openstack/rabbitmq-server-0" Feb 27 08:07:14 crc kubenswrapper[4612]: I0227 08:07:14.872846 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/0da436b1-803d-41bb-94ab-9cfd6fa769eb-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"0da436b1-803d-41bb-94ab-9cfd6fa769eb\") " pod="openstack/rabbitmq-server-0" Feb 27 08:07:14 crc kubenswrapper[4612]: I0227 08:07:14.873850 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/0da436b1-803d-41bb-94ab-9cfd6fa769eb-pod-info\") pod \"rabbitmq-server-0\" (UID: \"0da436b1-803d-41bb-94ab-9cfd6fa769eb\") " pod="openstack/rabbitmq-server-0" Feb 27 08:07:14 crc kubenswrapper[4612]: I0227 08:07:14.873895 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/0da436b1-803d-41bb-94ab-9cfd6fa769eb-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"0da436b1-803d-41bb-94ab-9cfd6fa769eb\") " pod="openstack/rabbitmq-server-0" Feb 27 08:07:14 crc kubenswrapper[4612]: I0227 08:07:14.877587 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vjjpx\" (UniqueName: \"kubernetes.io/projected/0da436b1-803d-41bb-94ab-9cfd6fa769eb-kube-api-access-vjjpx\") pod \"rabbitmq-server-0\" (UID: \"0da436b1-803d-41bb-94ab-9cfd6fa769eb\") " pod="openstack/rabbitmq-server-0" Feb 27 08:07:14 crc kubenswrapper[4612]: I0227 08:07:14.893377 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-server-0\" (UID: \"0da436b1-803d-41bb-94ab-9cfd6fa769eb\") " pod="openstack/rabbitmq-server-0" Feb 27 08:07:14 crc kubenswrapper[4612]: I0227 08:07:14.970239 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-bz4dn"] Feb 27 08:07:14 crc kubenswrapper[4612]: W0227 08:07:14.986107 4612 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod72e3b749_ce9a_4480_8873_d8c9b6495cbc.slice/crio-152840ba98fe6de0f95953025ad658d0f6483f322820a14332d945558142f743 WatchSource:0}: Error finding container 152840ba98fe6de0f95953025ad658d0f6483f322820a14332d945558142f743: Status 404 returned error can't find the container with id 152840ba98fe6de0f95953025ad658d0f6483f322820a14332d945558142f743 Feb 27 08:07:15 crc kubenswrapper[4612]: I0227 08:07:15.022197 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Feb 27 08:07:15 crc kubenswrapper[4612]: I0227 08:07:15.023386 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Feb 27 08:07:15 crc kubenswrapper[4612]: I0227 08:07:15.028031 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Feb 27 08:07:15 crc kubenswrapper[4612]: I0227 08:07:15.028259 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Feb 27 08:07:15 crc kubenswrapper[4612]: I0227 08:07:15.028408 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Feb 27 08:07:15 crc kubenswrapper[4612]: I0227 08:07:15.028586 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-lznsf" Feb 27 08:07:15 crc kubenswrapper[4612]: I0227 08:07:15.028733 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Feb 27 08:07:15 crc kubenswrapper[4612]: I0227 08:07:15.029562 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Feb 27 08:07:15 crc kubenswrapper[4612]: I0227 08:07:15.032005 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Feb 27 08:07:15 crc kubenswrapper[4612]: I0227 08:07:15.072150 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Feb 27 08:07:15 crc kubenswrapper[4612]: I0227 08:07:15.128089 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/6d1ba6db-1cae-461e-9716-5d5b36350218-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"6d1ba6db-1cae-461e-9716-5d5b36350218\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 08:07:15 crc kubenswrapper[4612]: I0227 08:07:15.128175 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-86pqk\" (UniqueName: \"kubernetes.io/projected/6d1ba6db-1cae-461e-9716-5d5b36350218-kube-api-access-86pqk\") pod \"rabbitmq-cell1-server-0\" (UID: \"6d1ba6db-1cae-461e-9716-5d5b36350218\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 08:07:15 crc kubenswrapper[4612]: I0227 08:07:15.128248 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/6d1ba6db-1cae-461e-9716-5d5b36350218-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"6d1ba6db-1cae-461e-9716-5d5b36350218\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 08:07:15 crc kubenswrapper[4612]: I0227 08:07:15.128301 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/6d1ba6db-1cae-461e-9716-5d5b36350218-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"6d1ba6db-1cae-461e-9716-5d5b36350218\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 08:07:15 crc kubenswrapper[4612]: I0227 08:07:15.128320 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/6d1ba6db-1cae-461e-9716-5d5b36350218-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"6d1ba6db-1cae-461e-9716-5d5b36350218\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 08:07:15 crc kubenswrapper[4612]: I0227 08:07:15.128337 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/6d1ba6db-1cae-461e-9716-5d5b36350218-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"6d1ba6db-1cae-461e-9716-5d5b36350218\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 08:07:15 crc kubenswrapper[4612]: I0227 08:07:15.128377 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/6d1ba6db-1cae-461e-9716-5d5b36350218-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"6d1ba6db-1cae-461e-9716-5d5b36350218\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 08:07:15 crc kubenswrapper[4612]: I0227 08:07:15.128396 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"6d1ba6db-1cae-461e-9716-5d5b36350218\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 08:07:15 crc kubenswrapper[4612]: I0227 08:07:15.128440 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6d1ba6db-1cae-461e-9716-5d5b36350218-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"6d1ba6db-1cae-461e-9716-5d5b36350218\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 08:07:15 crc kubenswrapper[4612]: I0227 08:07:15.128461 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/6d1ba6db-1cae-461e-9716-5d5b36350218-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"6d1ba6db-1cae-461e-9716-5d5b36350218\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 08:07:15 crc kubenswrapper[4612]: I0227 08:07:15.128483 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/6d1ba6db-1cae-461e-9716-5d5b36350218-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"6d1ba6db-1cae-461e-9716-5d5b36350218\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 08:07:15 crc kubenswrapper[4612]: I0227 08:07:15.186208 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Feb 27 08:07:15 crc kubenswrapper[4612]: I0227 08:07:15.230223 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/6d1ba6db-1cae-461e-9716-5d5b36350218-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"6d1ba6db-1cae-461e-9716-5d5b36350218\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 08:07:15 crc kubenswrapper[4612]: I0227 08:07:15.230267 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/6d1ba6db-1cae-461e-9716-5d5b36350218-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"6d1ba6db-1cae-461e-9716-5d5b36350218\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 08:07:15 crc kubenswrapper[4612]: I0227 08:07:15.230285 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/6d1ba6db-1cae-461e-9716-5d5b36350218-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"6d1ba6db-1cae-461e-9716-5d5b36350218\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 08:07:15 crc kubenswrapper[4612]: I0227 08:07:15.230306 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"6d1ba6db-1cae-461e-9716-5d5b36350218\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 08:07:15 crc kubenswrapper[4612]: I0227 08:07:15.230333 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6d1ba6db-1cae-461e-9716-5d5b36350218-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"6d1ba6db-1cae-461e-9716-5d5b36350218\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 08:07:15 crc kubenswrapper[4612]: I0227 08:07:15.230352 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/6d1ba6db-1cae-461e-9716-5d5b36350218-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"6d1ba6db-1cae-461e-9716-5d5b36350218\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 08:07:15 crc kubenswrapper[4612]: I0227 08:07:15.230372 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/6d1ba6db-1cae-461e-9716-5d5b36350218-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"6d1ba6db-1cae-461e-9716-5d5b36350218\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 08:07:15 crc kubenswrapper[4612]: I0227 08:07:15.230393 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/6d1ba6db-1cae-461e-9716-5d5b36350218-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"6d1ba6db-1cae-461e-9716-5d5b36350218\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 08:07:15 crc kubenswrapper[4612]: I0227 08:07:15.230431 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-86pqk\" (UniqueName: \"kubernetes.io/projected/6d1ba6db-1cae-461e-9716-5d5b36350218-kube-api-access-86pqk\") pod \"rabbitmq-cell1-server-0\" (UID: \"6d1ba6db-1cae-461e-9716-5d5b36350218\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 08:07:15 crc kubenswrapper[4612]: I0227 08:07:15.230470 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/6d1ba6db-1cae-461e-9716-5d5b36350218-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"6d1ba6db-1cae-461e-9716-5d5b36350218\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 08:07:15 crc kubenswrapper[4612]: I0227 08:07:15.230500 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/6d1ba6db-1cae-461e-9716-5d5b36350218-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"6d1ba6db-1cae-461e-9716-5d5b36350218\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 08:07:15 crc kubenswrapper[4612]: I0227 08:07:15.231940 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/6d1ba6db-1cae-461e-9716-5d5b36350218-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"6d1ba6db-1cae-461e-9716-5d5b36350218\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 08:07:15 crc kubenswrapper[4612]: I0227 08:07:15.232029 4612 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"6d1ba6db-1cae-461e-9716-5d5b36350218\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/rabbitmq-cell1-server-0" Feb 27 08:07:15 crc kubenswrapper[4612]: I0227 08:07:15.232554 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/6d1ba6db-1cae-461e-9716-5d5b36350218-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"6d1ba6db-1cae-461e-9716-5d5b36350218\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 08:07:15 crc kubenswrapper[4612]: I0227 08:07:15.233015 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/6d1ba6db-1cae-461e-9716-5d5b36350218-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"6d1ba6db-1cae-461e-9716-5d5b36350218\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 08:07:15 crc kubenswrapper[4612]: I0227 08:07:15.237136 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/6d1ba6db-1cae-461e-9716-5d5b36350218-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"6d1ba6db-1cae-461e-9716-5d5b36350218\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 08:07:15 crc kubenswrapper[4612]: I0227 08:07:15.238173 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/6d1ba6db-1cae-461e-9716-5d5b36350218-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"6d1ba6db-1cae-461e-9716-5d5b36350218\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 08:07:15 crc kubenswrapper[4612]: I0227 08:07:15.238195 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6d1ba6db-1cae-461e-9716-5d5b36350218-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"6d1ba6db-1cae-461e-9716-5d5b36350218\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 08:07:15 crc kubenswrapper[4612]: I0227 08:07:15.238621 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/6d1ba6db-1cae-461e-9716-5d5b36350218-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"6d1ba6db-1cae-461e-9716-5d5b36350218\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 08:07:15 crc kubenswrapper[4612]: I0227 08:07:15.239150 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/6d1ba6db-1cae-461e-9716-5d5b36350218-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"6d1ba6db-1cae-461e-9716-5d5b36350218\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 08:07:15 crc kubenswrapper[4612]: I0227 08:07:15.252976 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/6d1ba6db-1cae-461e-9716-5d5b36350218-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"6d1ba6db-1cae-461e-9716-5d5b36350218\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 08:07:15 crc kubenswrapper[4612]: I0227 08:07:15.256469 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-86pqk\" (UniqueName: \"kubernetes.io/projected/6d1ba6db-1cae-461e-9716-5d5b36350218-kube-api-access-86pqk\") pod \"rabbitmq-cell1-server-0\" (UID: \"6d1ba6db-1cae-461e-9716-5d5b36350218\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 08:07:15 crc kubenswrapper[4612]: I0227 08:07:15.274020 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"6d1ba6db-1cae-461e-9716-5d5b36350218\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 08:07:15 crc kubenswrapper[4612]: I0227 08:07:15.364494 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Feb 27 08:07:15 crc kubenswrapper[4612]: I0227 08:07:15.754211 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Feb 27 08:07:15 crc kubenswrapper[4612]: I0227 08:07:15.866182 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"0da436b1-803d-41bb-94ab-9cfd6fa769eb","Type":"ContainerStarted","Data":"b0d2efa8e535b610aa785074d361a1c941c0b5146f868873b1ed12043662d340"} Feb 27 08:07:15 crc kubenswrapper[4612]: I0227 08:07:15.867505 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-bz4dn" event={"ID":"72e3b749-ce9a-4480-8873-d8c9b6495cbc","Type":"ContainerStarted","Data":"152840ba98fe6de0f95953025ad658d0f6483f322820a14332d945558142f743"} Feb 27 08:07:15 crc kubenswrapper[4612]: I0227 08:07:15.899383 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Feb 27 08:07:15 crc kubenswrapper[4612]: I0227 08:07:15.956394 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Feb 27 08:07:15 crc kubenswrapper[4612]: I0227 08:07:15.958348 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Feb 27 08:07:15 crc kubenswrapper[4612]: I0227 08:07:15.968309 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Feb 27 08:07:15 crc kubenswrapper[4612]: I0227 08:07:15.968583 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Feb 27 08:07:15 crc kubenswrapper[4612]: I0227 08:07:15.968782 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-jwwpt" Feb 27 08:07:15 crc kubenswrapper[4612]: I0227 08:07:15.968897 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Feb 27 08:07:15 crc kubenswrapper[4612]: I0227 08:07:15.972394 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Feb 27 08:07:15 crc kubenswrapper[4612]: I0227 08:07:15.976558 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Feb 27 08:07:16 crc kubenswrapper[4612]: I0227 08:07:16.045838 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f189e394-0c97-48a4-9c55-4bd6710923f1-operator-scripts\") pod \"openstack-galera-0\" (UID: \"f189e394-0c97-48a4-9c55-4bd6710923f1\") " pod="openstack/openstack-galera-0" Feb 27 08:07:16 crc kubenswrapper[4612]: I0227 08:07:16.045933 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d4fl8\" (UniqueName: \"kubernetes.io/projected/f189e394-0c97-48a4-9c55-4bd6710923f1-kube-api-access-d4fl8\") pod \"openstack-galera-0\" (UID: \"f189e394-0c97-48a4-9c55-4bd6710923f1\") " pod="openstack/openstack-galera-0" Feb 27 08:07:16 crc kubenswrapper[4612]: I0227 08:07:16.045972 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/f189e394-0c97-48a4-9c55-4bd6710923f1-config-data-default\") pod \"openstack-galera-0\" (UID: \"f189e394-0c97-48a4-9c55-4bd6710923f1\") " pod="openstack/openstack-galera-0" Feb 27 08:07:16 crc kubenswrapper[4612]: I0227 08:07:16.046009 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f189e394-0c97-48a4-9c55-4bd6710923f1-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"f189e394-0c97-48a4-9c55-4bd6710923f1\") " pod="openstack/openstack-galera-0" Feb 27 08:07:16 crc kubenswrapper[4612]: I0227 08:07:16.046065 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/f189e394-0c97-48a4-9c55-4bd6710923f1-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"f189e394-0c97-48a4-9c55-4bd6710923f1\") " pod="openstack/openstack-galera-0" Feb 27 08:07:16 crc kubenswrapper[4612]: I0227 08:07:16.046100 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"openstack-galera-0\" (UID: \"f189e394-0c97-48a4-9c55-4bd6710923f1\") " pod="openstack/openstack-galera-0" Feb 27 08:07:16 crc kubenswrapper[4612]: I0227 08:07:16.046127 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/f189e394-0c97-48a4-9c55-4bd6710923f1-config-data-generated\") pod \"openstack-galera-0\" (UID: \"f189e394-0c97-48a4-9c55-4bd6710923f1\") " pod="openstack/openstack-galera-0" Feb 27 08:07:16 crc kubenswrapper[4612]: I0227 08:07:16.046152 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/f189e394-0c97-48a4-9c55-4bd6710923f1-kolla-config\") pod \"openstack-galera-0\" (UID: \"f189e394-0c97-48a4-9c55-4bd6710923f1\") " pod="openstack/openstack-galera-0" Feb 27 08:07:16 crc kubenswrapper[4612]: I0227 08:07:16.147113 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d4fl8\" (UniqueName: \"kubernetes.io/projected/f189e394-0c97-48a4-9c55-4bd6710923f1-kube-api-access-d4fl8\") pod \"openstack-galera-0\" (UID: \"f189e394-0c97-48a4-9c55-4bd6710923f1\") " pod="openstack/openstack-galera-0" Feb 27 08:07:16 crc kubenswrapper[4612]: I0227 08:07:16.147173 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/f189e394-0c97-48a4-9c55-4bd6710923f1-config-data-default\") pod \"openstack-galera-0\" (UID: \"f189e394-0c97-48a4-9c55-4bd6710923f1\") " pod="openstack/openstack-galera-0" Feb 27 08:07:16 crc kubenswrapper[4612]: I0227 08:07:16.147210 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f189e394-0c97-48a4-9c55-4bd6710923f1-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"f189e394-0c97-48a4-9c55-4bd6710923f1\") " pod="openstack/openstack-galera-0" Feb 27 08:07:16 crc kubenswrapper[4612]: I0227 08:07:16.147258 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/f189e394-0c97-48a4-9c55-4bd6710923f1-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"f189e394-0c97-48a4-9c55-4bd6710923f1\") " pod="openstack/openstack-galera-0" Feb 27 08:07:16 crc kubenswrapper[4612]: I0227 08:07:16.147285 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"openstack-galera-0\" (UID: \"f189e394-0c97-48a4-9c55-4bd6710923f1\") " pod="openstack/openstack-galera-0" Feb 27 08:07:16 crc kubenswrapper[4612]: I0227 08:07:16.147313 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/f189e394-0c97-48a4-9c55-4bd6710923f1-config-data-generated\") pod \"openstack-galera-0\" (UID: \"f189e394-0c97-48a4-9c55-4bd6710923f1\") " pod="openstack/openstack-galera-0" Feb 27 08:07:16 crc kubenswrapper[4612]: I0227 08:07:16.147336 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/f189e394-0c97-48a4-9c55-4bd6710923f1-kolla-config\") pod \"openstack-galera-0\" (UID: \"f189e394-0c97-48a4-9c55-4bd6710923f1\") " pod="openstack/openstack-galera-0" Feb 27 08:07:16 crc kubenswrapper[4612]: I0227 08:07:16.147388 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f189e394-0c97-48a4-9c55-4bd6710923f1-operator-scripts\") pod \"openstack-galera-0\" (UID: \"f189e394-0c97-48a4-9c55-4bd6710923f1\") " pod="openstack/openstack-galera-0" Feb 27 08:07:16 crc kubenswrapper[4612]: I0227 08:07:16.148589 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/f189e394-0c97-48a4-9c55-4bd6710923f1-config-data-default\") pod \"openstack-galera-0\" (UID: \"f189e394-0c97-48a4-9c55-4bd6710923f1\") " pod="openstack/openstack-galera-0" Feb 27 08:07:16 crc kubenswrapper[4612]: I0227 08:07:16.148982 4612 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"openstack-galera-0\" (UID: \"f189e394-0c97-48a4-9c55-4bd6710923f1\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/openstack-galera-0" Feb 27 08:07:16 crc kubenswrapper[4612]: I0227 08:07:16.149027 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/f189e394-0c97-48a4-9c55-4bd6710923f1-kolla-config\") pod \"openstack-galera-0\" (UID: \"f189e394-0c97-48a4-9c55-4bd6710923f1\") " pod="openstack/openstack-galera-0" Feb 27 08:07:16 crc kubenswrapper[4612]: I0227 08:07:16.149879 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f189e394-0c97-48a4-9c55-4bd6710923f1-operator-scripts\") pod \"openstack-galera-0\" (UID: \"f189e394-0c97-48a4-9c55-4bd6710923f1\") " pod="openstack/openstack-galera-0" Feb 27 08:07:16 crc kubenswrapper[4612]: I0227 08:07:16.168241 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/f189e394-0c97-48a4-9c55-4bd6710923f1-config-data-generated\") pod \"openstack-galera-0\" (UID: \"f189e394-0c97-48a4-9c55-4bd6710923f1\") " pod="openstack/openstack-galera-0" Feb 27 08:07:16 crc kubenswrapper[4612]: I0227 08:07:16.184516 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f189e394-0c97-48a4-9c55-4bd6710923f1-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"f189e394-0c97-48a4-9c55-4bd6710923f1\") " pod="openstack/openstack-galera-0" Feb 27 08:07:16 crc kubenswrapper[4612]: I0227 08:07:16.184548 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d4fl8\" (UniqueName: \"kubernetes.io/projected/f189e394-0c97-48a4-9c55-4bd6710923f1-kube-api-access-d4fl8\") pod \"openstack-galera-0\" (UID: \"f189e394-0c97-48a4-9c55-4bd6710923f1\") " pod="openstack/openstack-galera-0" Feb 27 08:07:16 crc kubenswrapper[4612]: I0227 08:07:16.187798 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"openstack-galera-0\" (UID: \"f189e394-0c97-48a4-9c55-4bd6710923f1\") " pod="openstack/openstack-galera-0" Feb 27 08:07:16 crc kubenswrapper[4612]: I0227 08:07:16.195885 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/f189e394-0c97-48a4-9c55-4bd6710923f1-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"f189e394-0c97-48a4-9c55-4bd6710923f1\") " pod="openstack/openstack-galera-0" Feb 27 08:07:16 crc kubenswrapper[4612]: I0227 08:07:16.288289 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Feb 27 08:07:16 crc kubenswrapper[4612]: I0227 08:07:16.876338 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Feb 27 08:07:16 crc kubenswrapper[4612]: I0227 08:07:16.911858 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"6d1ba6db-1cae-461e-9716-5d5b36350218","Type":"ContainerStarted","Data":"152b8e50f185ca631e828e8d4886ab57c0abda78df5a4124b28847d2a6cd7111"} Feb 27 08:07:17 crc kubenswrapper[4612]: I0227 08:07:17.358378 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Feb 27 08:07:17 crc kubenswrapper[4612]: I0227 08:07:17.365889 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Feb 27 08:07:17 crc kubenswrapper[4612]: I0227 08:07:17.366078 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Feb 27 08:07:17 crc kubenswrapper[4612]: I0227 08:07:17.369933 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Feb 27 08:07:17 crc kubenswrapper[4612]: I0227 08:07:17.370245 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Feb 27 08:07:17 crc kubenswrapper[4612]: I0227 08:07:17.370519 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-9dghk" Feb 27 08:07:17 crc kubenswrapper[4612]: I0227 08:07:17.371966 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Feb 27 08:07:17 crc kubenswrapper[4612]: I0227 08:07:17.472822 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-drmd6\" (UniqueName: \"kubernetes.io/projected/e02a40f2-3de1-4f01-84b0-2fe56acdd396-kube-api-access-drmd6\") pod \"openstack-cell1-galera-0\" (UID: \"e02a40f2-3de1-4f01-84b0-2fe56acdd396\") " pod="openstack/openstack-cell1-galera-0" Feb 27 08:07:17 crc kubenswrapper[4612]: I0227 08:07:17.472885 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e02a40f2-3de1-4f01-84b0-2fe56acdd396-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"e02a40f2-3de1-4f01-84b0-2fe56acdd396\") " pod="openstack/openstack-cell1-galera-0" Feb 27 08:07:17 crc kubenswrapper[4612]: I0227 08:07:17.472913 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/e02a40f2-3de1-4f01-84b0-2fe56acdd396-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"e02a40f2-3de1-4f01-84b0-2fe56acdd396\") " pod="openstack/openstack-cell1-galera-0" Feb 27 08:07:17 crc kubenswrapper[4612]: I0227 08:07:17.472971 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e02a40f2-3de1-4f01-84b0-2fe56acdd396-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"e02a40f2-3de1-4f01-84b0-2fe56acdd396\") " pod="openstack/openstack-cell1-galera-0" Feb 27 08:07:17 crc kubenswrapper[4612]: I0227 08:07:17.473134 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/e02a40f2-3de1-4f01-84b0-2fe56acdd396-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"e02a40f2-3de1-4f01-84b0-2fe56acdd396\") " pod="openstack/openstack-cell1-galera-0" Feb 27 08:07:17 crc kubenswrapper[4612]: I0227 08:07:17.473278 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/e02a40f2-3de1-4f01-84b0-2fe56acdd396-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"e02a40f2-3de1-4f01-84b0-2fe56acdd396\") " pod="openstack/openstack-cell1-galera-0" Feb 27 08:07:17 crc kubenswrapper[4612]: I0227 08:07:17.473307 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"openstack-cell1-galera-0\" (UID: \"e02a40f2-3de1-4f01-84b0-2fe56acdd396\") " pod="openstack/openstack-cell1-galera-0" Feb 27 08:07:17 crc kubenswrapper[4612]: I0227 08:07:17.473333 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/e02a40f2-3de1-4f01-84b0-2fe56acdd396-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"e02a40f2-3de1-4f01-84b0-2fe56acdd396\") " pod="openstack/openstack-cell1-galera-0" Feb 27 08:07:17 crc kubenswrapper[4612]: I0227 08:07:17.574362 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"openstack-cell1-galera-0\" (UID: \"e02a40f2-3de1-4f01-84b0-2fe56acdd396\") " pod="openstack/openstack-cell1-galera-0" Feb 27 08:07:17 crc kubenswrapper[4612]: I0227 08:07:17.574419 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/e02a40f2-3de1-4f01-84b0-2fe56acdd396-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"e02a40f2-3de1-4f01-84b0-2fe56acdd396\") " pod="openstack/openstack-cell1-galera-0" Feb 27 08:07:17 crc kubenswrapper[4612]: I0227 08:07:17.574456 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-drmd6\" (UniqueName: \"kubernetes.io/projected/e02a40f2-3de1-4f01-84b0-2fe56acdd396-kube-api-access-drmd6\") pod \"openstack-cell1-galera-0\" (UID: \"e02a40f2-3de1-4f01-84b0-2fe56acdd396\") " pod="openstack/openstack-cell1-galera-0" Feb 27 08:07:17 crc kubenswrapper[4612]: I0227 08:07:17.574482 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e02a40f2-3de1-4f01-84b0-2fe56acdd396-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"e02a40f2-3de1-4f01-84b0-2fe56acdd396\") " pod="openstack/openstack-cell1-galera-0" Feb 27 08:07:17 crc kubenswrapper[4612]: I0227 08:07:17.574505 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/e02a40f2-3de1-4f01-84b0-2fe56acdd396-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"e02a40f2-3de1-4f01-84b0-2fe56acdd396\") " pod="openstack/openstack-cell1-galera-0" Feb 27 08:07:17 crc kubenswrapper[4612]: I0227 08:07:17.574534 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e02a40f2-3de1-4f01-84b0-2fe56acdd396-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"e02a40f2-3de1-4f01-84b0-2fe56acdd396\") " pod="openstack/openstack-cell1-galera-0" Feb 27 08:07:17 crc kubenswrapper[4612]: I0227 08:07:17.574587 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/e02a40f2-3de1-4f01-84b0-2fe56acdd396-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"e02a40f2-3de1-4f01-84b0-2fe56acdd396\") " pod="openstack/openstack-cell1-galera-0" Feb 27 08:07:17 crc kubenswrapper[4612]: I0227 08:07:17.574652 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/e02a40f2-3de1-4f01-84b0-2fe56acdd396-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"e02a40f2-3de1-4f01-84b0-2fe56acdd396\") " pod="openstack/openstack-cell1-galera-0" Feb 27 08:07:17 crc kubenswrapper[4612]: I0227 08:07:17.575548 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/e02a40f2-3de1-4f01-84b0-2fe56acdd396-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"e02a40f2-3de1-4f01-84b0-2fe56acdd396\") " pod="openstack/openstack-cell1-galera-0" Feb 27 08:07:17 crc kubenswrapper[4612]: I0227 08:07:17.575847 4612 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"openstack-cell1-galera-0\" (UID: \"e02a40f2-3de1-4f01-84b0-2fe56acdd396\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/openstack-cell1-galera-0" Feb 27 08:07:17 crc kubenswrapper[4612]: I0227 08:07:17.575979 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/e02a40f2-3de1-4f01-84b0-2fe56acdd396-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"e02a40f2-3de1-4f01-84b0-2fe56acdd396\") " pod="openstack/openstack-cell1-galera-0" Feb 27 08:07:17 crc kubenswrapper[4612]: I0227 08:07:17.576128 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/e02a40f2-3de1-4f01-84b0-2fe56acdd396-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"e02a40f2-3de1-4f01-84b0-2fe56acdd396\") " pod="openstack/openstack-cell1-galera-0" Feb 27 08:07:17 crc kubenswrapper[4612]: I0227 08:07:17.578060 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e02a40f2-3de1-4f01-84b0-2fe56acdd396-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"e02a40f2-3de1-4f01-84b0-2fe56acdd396\") " pod="openstack/openstack-cell1-galera-0" Feb 27 08:07:17 crc kubenswrapper[4612]: I0227 08:07:17.584817 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/e02a40f2-3de1-4f01-84b0-2fe56acdd396-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"e02a40f2-3de1-4f01-84b0-2fe56acdd396\") " pod="openstack/openstack-cell1-galera-0" Feb 27 08:07:17 crc kubenswrapper[4612]: I0227 08:07:17.597296 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e02a40f2-3de1-4f01-84b0-2fe56acdd396-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"e02a40f2-3de1-4f01-84b0-2fe56acdd396\") " pod="openstack/openstack-cell1-galera-0" Feb 27 08:07:17 crc kubenswrapper[4612]: I0227 08:07:17.598176 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-drmd6\" (UniqueName: \"kubernetes.io/projected/e02a40f2-3de1-4f01-84b0-2fe56acdd396-kube-api-access-drmd6\") pod \"openstack-cell1-galera-0\" (UID: \"e02a40f2-3de1-4f01-84b0-2fe56acdd396\") " pod="openstack/openstack-cell1-galera-0" Feb 27 08:07:17 crc kubenswrapper[4612]: I0227 08:07:17.640209 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"openstack-cell1-galera-0\" (UID: \"e02a40f2-3de1-4f01-84b0-2fe56acdd396\") " pod="openstack/openstack-cell1-galera-0" Feb 27 08:07:17 crc kubenswrapper[4612]: I0227 08:07:17.662654 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Feb 27 08:07:17 crc kubenswrapper[4612]: I0227 08:07:17.670800 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Feb 27 08:07:17 crc kubenswrapper[4612]: I0227 08:07:17.679836 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Feb 27 08:07:17 crc kubenswrapper[4612]: I0227 08:07:17.680316 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Feb 27 08:07:17 crc kubenswrapper[4612]: I0227 08:07:17.680531 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-tpr57" Feb 27 08:07:17 crc kubenswrapper[4612]: I0227 08:07:17.702054 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Feb 27 08:07:17 crc kubenswrapper[4612]: I0227 08:07:17.709007 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Feb 27 08:07:17 crc kubenswrapper[4612]: I0227 08:07:17.779733 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/d19bfe90-c5ca-4089-b4f0-bf3eccff30c8-kolla-config\") pod \"memcached-0\" (UID: \"d19bfe90-c5ca-4089-b4f0-bf3eccff30c8\") " pod="openstack/memcached-0" Feb 27 08:07:17 crc kubenswrapper[4612]: I0227 08:07:17.780057 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d19bfe90-c5ca-4089-b4f0-bf3eccff30c8-config-data\") pod \"memcached-0\" (UID: \"d19bfe90-c5ca-4089-b4f0-bf3eccff30c8\") " pod="openstack/memcached-0" Feb 27 08:07:17 crc kubenswrapper[4612]: I0227 08:07:17.780160 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dxx5j\" (UniqueName: \"kubernetes.io/projected/d19bfe90-c5ca-4089-b4f0-bf3eccff30c8-kube-api-access-dxx5j\") pod \"memcached-0\" (UID: \"d19bfe90-c5ca-4089-b4f0-bf3eccff30c8\") " pod="openstack/memcached-0" Feb 27 08:07:17 crc kubenswrapper[4612]: I0227 08:07:17.780313 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d19bfe90-c5ca-4089-b4f0-bf3eccff30c8-combined-ca-bundle\") pod \"memcached-0\" (UID: \"d19bfe90-c5ca-4089-b4f0-bf3eccff30c8\") " pod="openstack/memcached-0" Feb 27 08:07:17 crc kubenswrapper[4612]: I0227 08:07:17.780755 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/d19bfe90-c5ca-4089-b4f0-bf3eccff30c8-memcached-tls-certs\") pod \"memcached-0\" (UID: \"d19bfe90-c5ca-4089-b4f0-bf3eccff30c8\") " pod="openstack/memcached-0" Feb 27 08:07:17 crc kubenswrapper[4612]: I0227 08:07:17.882066 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/d19bfe90-c5ca-4089-b4f0-bf3eccff30c8-kolla-config\") pod \"memcached-0\" (UID: \"d19bfe90-c5ca-4089-b4f0-bf3eccff30c8\") " pod="openstack/memcached-0" Feb 27 08:07:17 crc kubenswrapper[4612]: I0227 08:07:17.882121 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d19bfe90-c5ca-4089-b4f0-bf3eccff30c8-config-data\") pod \"memcached-0\" (UID: \"d19bfe90-c5ca-4089-b4f0-bf3eccff30c8\") " pod="openstack/memcached-0" Feb 27 08:07:17 crc kubenswrapper[4612]: I0227 08:07:17.882138 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dxx5j\" (UniqueName: \"kubernetes.io/projected/d19bfe90-c5ca-4089-b4f0-bf3eccff30c8-kube-api-access-dxx5j\") pod \"memcached-0\" (UID: \"d19bfe90-c5ca-4089-b4f0-bf3eccff30c8\") " pod="openstack/memcached-0" Feb 27 08:07:17 crc kubenswrapper[4612]: I0227 08:07:17.882167 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d19bfe90-c5ca-4089-b4f0-bf3eccff30c8-combined-ca-bundle\") pod \"memcached-0\" (UID: \"d19bfe90-c5ca-4089-b4f0-bf3eccff30c8\") " pod="openstack/memcached-0" Feb 27 08:07:17 crc kubenswrapper[4612]: I0227 08:07:17.882224 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/d19bfe90-c5ca-4089-b4f0-bf3eccff30c8-memcached-tls-certs\") pod \"memcached-0\" (UID: \"d19bfe90-c5ca-4089-b4f0-bf3eccff30c8\") " pod="openstack/memcached-0" Feb 27 08:07:17 crc kubenswrapper[4612]: I0227 08:07:17.883049 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/d19bfe90-c5ca-4089-b4f0-bf3eccff30c8-kolla-config\") pod \"memcached-0\" (UID: \"d19bfe90-c5ca-4089-b4f0-bf3eccff30c8\") " pod="openstack/memcached-0" Feb 27 08:07:17 crc kubenswrapper[4612]: I0227 08:07:17.885355 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d19bfe90-c5ca-4089-b4f0-bf3eccff30c8-config-data\") pod \"memcached-0\" (UID: \"d19bfe90-c5ca-4089-b4f0-bf3eccff30c8\") " pod="openstack/memcached-0" Feb 27 08:07:17 crc kubenswrapper[4612]: I0227 08:07:17.901773 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/d19bfe90-c5ca-4089-b4f0-bf3eccff30c8-memcached-tls-certs\") pod \"memcached-0\" (UID: \"d19bfe90-c5ca-4089-b4f0-bf3eccff30c8\") " pod="openstack/memcached-0" Feb 27 08:07:17 crc kubenswrapper[4612]: I0227 08:07:17.904648 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d19bfe90-c5ca-4089-b4f0-bf3eccff30c8-combined-ca-bundle\") pod \"memcached-0\" (UID: \"d19bfe90-c5ca-4089-b4f0-bf3eccff30c8\") " pod="openstack/memcached-0" Feb 27 08:07:17 crc kubenswrapper[4612]: I0227 08:07:17.909600 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dxx5j\" (UniqueName: \"kubernetes.io/projected/d19bfe90-c5ca-4089-b4f0-bf3eccff30c8-kube-api-access-dxx5j\") pod \"memcached-0\" (UID: \"d19bfe90-c5ca-4089-b4f0-bf3eccff30c8\") " pod="openstack/memcached-0" Feb 27 08:07:17 crc kubenswrapper[4612]: I0227 08:07:17.922686 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"f189e394-0c97-48a4-9c55-4bd6710923f1","Type":"ContainerStarted","Data":"cad7e4d2258f65bb2ec78d84e60b28f4d04bb8969c13b19ff753b3f147cd4270"} Feb 27 08:07:18 crc kubenswrapper[4612]: I0227 08:07:18.011361 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Feb 27 08:07:18 crc kubenswrapper[4612]: I0227 08:07:18.348599 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Feb 27 08:07:18 crc kubenswrapper[4612]: I0227 08:07:18.781022 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Feb 27 08:07:18 crc kubenswrapper[4612]: I0227 08:07:18.961109 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"d19bfe90-c5ca-4089-b4f0-bf3eccff30c8","Type":"ContainerStarted","Data":"6f0f3251e94ace1ab4e01e1ed13f865a01afea520b48b5aa5e1e17735bf6ac93"} Feb 27 08:07:18 crc kubenswrapper[4612]: I0227 08:07:18.966658 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"e02a40f2-3de1-4f01-84b0-2fe56acdd396","Type":"ContainerStarted","Data":"2771aacd82e00d63cf01e68025d16e22cff73edb79d65ac699370800ab437e36"} Feb 27 08:07:20 crc kubenswrapper[4612]: I0227 08:07:20.354911 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Feb 27 08:07:20 crc kubenswrapper[4612]: I0227 08:07:20.360221 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Feb 27 08:07:20 crc kubenswrapper[4612]: I0227 08:07:20.369543 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-x4vk8" Feb 27 08:07:20 crc kubenswrapper[4612]: I0227 08:07:20.376829 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Feb 27 08:07:20 crc kubenswrapper[4612]: I0227 08:07:20.453940 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b9h48\" (UniqueName: \"kubernetes.io/projected/7c51bfaf-2c6a-49aa-afe7-eb4789351857-kube-api-access-b9h48\") pod \"kube-state-metrics-0\" (UID: \"7c51bfaf-2c6a-49aa-afe7-eb4789351857\") " pod="openstack/kube-state-metrics-0" Feb 27 08:07:20 crc kubenswrapper[4612]: I0227 08:07:20.555704 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b9h48\" (UniqueName: \"kubernetes.io/projected/7c51bfaf-2c6a-49aa-afe7-eb4789351857-kube-api-access-b9h48\") pod \"kube-state-metrics-0\" (UID: \"7c51bfaf-2c6a-49aa-afe7-eb4789351857\") " pod="openstack/kube-state-metrics-0" Feb 27 08:07:20 crc kubenswrapper[4612]: I0227 08:07:20.598419 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b9h48\" (UniqueName: \"kubernetes.io/projected/7c51bfaf-2c6a-49aa-afe7-eb4789351857-kube-api-access-b9h48\") pod \"kube-state-metrics-0\" (UID: \"7c51bfaf-2c6a-49aa-afe7-eb4789351857\") " pod="openstack/kube-state-metrics-0" Feb 27 08:07:20 crc kubenswrapper[4612]: I0227 08:07:20.695162 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Feb 27 08:07:21 crc kubenswrapper[4612]: I0227 08:07:21.412341 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Feb 27 08:07:21 crc kubenswrapper[4612]: W0227 08:07:21.421684 4612 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7c51bfaf_2c6a_49aa_afe7_eb4789351857.slice/crio-a6c6141ac16efb9dd3d28c920fec9a6735719a3bc398e647edc3c62558c8250b WatchSource:0}: Error finding container a6c6141ac16efb9dd3d28c920fec9a6735719a3bc398e647edc3c62558c8250b: Status 404 returned error can't find the container with id a6c6141ac16efb9dd3d28c920fec9a6735719a3bc398e647edc3c62558c8250b Feb 27 08:07:22 crc kubenswrapper[4612]: I0227 08:07:22.022997 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"7c51bfaf-2c6a-49aa-afe7-eb4789351857","Type":"ContainerStarted","Data":"a6c6141ac16efb9dd3d28c920fec9a6735719a3bc398e647edc3c62558c8250b"} Feb 27 08:07:24 crc kubenswrapper[4612]: I0227 08:07:24.008869 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-z7f9r"] Feb 27 08:07:24 crc kubenswrapper[4612]: I0227 08:07:24.010132 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-z7f9r" Feb 27 08:07:24 crc kubenswrapper[4612]: I0227 08:07:24.018238 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Feb 27 08:07:24 crc kubenswrapper[4612]: I0227 08:07:24.018502 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Feb 27 08:07:24 crc kubenswrapper[4612]: I0227 08:07:24.018661 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-85npx" Feb 27 08:07:24 crc kubenswrapper[4612]: I0227 08:07:24.046792 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-x255x"] Feb 27 08:07:24 crc kubenswrapper[4612]: I0227 08:07:24.049217 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-x255x" Feb 27 08:07:24 crc kubenswrapper[4612]: I0227 08:07:24.057190 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-z7f9r"] Feb 27 08:07:24 crc kubenswrapper[4612]: I0227 08:07:24.073441 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-x255x"] Feb 27 08:07:24 crc kubenswrapper[4612]: I0227 08:07:24.148572 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1ef66a51-46e7-4b76-a70b-b475d47822fd-scripts\") pod \"ovn-controller-z7f9r\" (UID: \"1ef66a51-46e7-4b76-a70b-b475d47822fd\") " pod="openstack/ovn-controller-z7f9r" Feb 27 08:07:24 crc kubenswrapper[4612]: I0227 08:07:24.148632 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/1ef66a51-46e7-4b76-a70b-b475d47822fd-var-run\") pod \"ovn-controller-z7f9r\" (UID: \"1ef66a51-46e7-4b76-a70b-b475d47822fd\") " pod="openstack/ovn-controller-z7f9r" Feb 27 08:07:24 crc kubenswrapper[4612]: I0227 08:07:24.148829 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4kdqc\" (UniqueName: \"kubernetes.io/projected/1ef66a51-46e7-4b76-a70b-b475d47822fd-kube-api-access-4kdqc\") pod \"ovn-controller-z7f9r\" (UID: \"1ef66a51-46e7-4b76-a70b-b475d47822fd\") " pod="openstack/ovn-controller-z7f9r" Feb 27 08:07:24 crc kubenswrapper[4612]: I0227 08:07:24.148910 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/1ef66a51-46e7-4b76-a70b-b475d47822fd-var-run-ovn\") pod \"ovn-controller-z7f9r\" (UID: \"1ef66a51-46e7-4b76-a70b-b475d47822fd\") " pod="openstack/ovn-controller-z7f9r" Feb 27 08:07:24 crc kubenswrapper[4612]: I0227 08:07:24.148974 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1ef66a51-46e7-4b76-a70b-b475d47822fd-combined-ca-bundle\") pod \"ovn-controller-z7f9r\" (UID: \"1ef66a51-46e7-4b76-a70b-b475d47822fd\") " pod="openstack/ovn-controller-z7f9r" Feb 27 08:07:24 crc kubenswrapper[4612]: I0227 08:07:24.149016 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/1ef66a51-46e7-4b76-a70b-b475d47822fd-var-log-ovn\") pod \"ovn-controller-z7f9r\" (UID: \"1ef66a51-46e7-4b76-a70b-b475d47822fd\") " pod="openstack/ovn-controller-z7f9r" Feb 27 08:07:24 crc kubenswrapper[4612]: I0227 08:07:24.149060 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/1ef66a51-46e7-4b76-a70b-b475d47822fd-ovn-controller-tls-certs\") pod \"ovn-controller-z7f9r\" (UID: \"1ef66a51-46e7-4b76-a70b-b475d47822fd\") " pod="openstack/ovn-controller-z7f9r" Feb 27 08:07:24 crc kubenswrapper[4612]: I0227 08:07:24.250219 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-njlpm\" (UniqueName: \"kubernetes.io/projected/20c3f793-4286-4067-a372-37b2de08fa8e-kube-api-access-njlpm\") pod \"ovn-controller-ovs-x255x\" (UID: \"20c3f793-4286-4067-a372-37b2de08fa8e\") " pod="openstack/ovn-controller-ovs-x255x" Feb 27 08:07:24 crc kubenswrapper[4612]: I0227 08:07:24.250281 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1ef66a51-46e7-4b76-a70b-b475d47822fd-combined-ca-bundle\") pod \"ovn-controller-z7f9r\" (UID: \"1ef66a51-46e7-4b76-a70b-b475d47822fd\") " pod="openstack/ovn-controller-z7f9r" Feb 27 08:07:24 crc kubenswrapper[4612]: I0227 08:07:24.250312 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/20c3f793-4286-4067-a372-37b2de08fa8e-scripts\") pod \"ovn-controller-ovs-x255x\" (UID: \"20c3f793-4286-4067-a372-37b2de08fa8e\") " pod="openstack/ovn-controller-ovs-x255x" Feb 27 08:07:24 crc kubenswrapper[4612]: I0227 08:07:24.250332 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/1ef66a51-46e7-4b76-a70b-b475d47822fd-var-log-ovn\") pod \"ovn-controller-z7f9r\" (UID: \"1ef66a51-46e7-4b76-a70b-b475d47822fd\") " pod="openstack/ovn-controller-z7f9r" Feb 27 08:07:24 crc kubenswrapper[4612]: I0227 08:07:24.251104 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/20c3f793-4286-4067-a372-37b2de08fa8e-var-log\") pod \"ovn-controller-ovs-x255x\" (UID: \"20c3f793-4286-4067-a372-37b2de08fa8e\") " pod="openstack/ovn-controller-ovs-x255x" Feb 27 08:07:24 crc kubenswrapper[4612]: I0227 08:07:24.251180 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/20c3f793-4286-4067-a372-37b2de08fa8e-etc-ovs\") pod \"ovn-controller-ovs-x255x\" (UID: \"20c3f793-4286-4067-a372-37b2de08fa8e\") " pod="openstack/ovn-controller-ovs-x255x" Feb 27 08:07:24 crc kubenswrapper[4612]: I0227 08:07:24.251256 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/1ef66a51-46e7-4b76-a70b-b475d47822fd-ovn-controller-tls-certs\") pod \"ovn-controller-z7f9r\" (UID: \"1ef66a51-46e7-4b76-a70b-b475d47822fd\") " pod="openstack/ovn-controller-z7f9r" Feb 27 08:07:24 crc kubenswrapper[4612]: I0227 08:07:24.251277 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/20c3f793-4286-4067-a372-37b2de08fa8e-var-lib\") pod \"ovn-controller-ovs-x255x\" (UID: \"20c3f793-4286-4067-a372-37b2de08fa8e\") " pod="openstack/ovn-controller-ovs-x255x" Feb 27 08:07:24 crc kubenswrapper[4612]: I0227 08:07:24.251365 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1ef66a51-46e7-4b76-a70b-b475d47822fd-scripts\") pod \"ovn-controller-z7f9r\" (UID: \"1ef66a51-46e7-4b76-a70b-b475d47822fd\") " pod="openstack/ovn-controller-z7f9r" Feb 27 08:07:24 crc kubenswrapper[4612]: I0227 08:07:24.251398 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/1ef66a51-46e7-4b76-a70b-b475d47822fd-var-run\") pod \"ovn-controller-z7f9r\" (UID: \"1ef66a51-46e7-4b76-a70b-b475d47822fd\") " pod="openstack/ovn-controller-z7f9r" Feb 27 08:07:24 crc kubenswrapper[4612]: I0227 08:07:24.251449 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/20c3f793-4286-4067-a372-37b2de08fa8e-var-run\") pod \"ovn-controller-ovs-x255x\" (UID: \"20c3f793-4286-4067-a372-37b2de08fa8e\") " pod="openstack/ovn-controller-ovs-x255x" Feb 27 08:07:24 crc kubenswrapper[4612]: I0227 08:07:24.251497 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4kdqc\" (UniqueName: \"kubernetes.io/projected/1ef66a51-46e7-4b76-a70b-b475d47822fd-kube-api-access-4kdqc\") pod \"ovn-controller-z7f9r\" (UID: \"1ef66a51-46e7-4b76-a70b-b475d47822fd\") " pod="openstack/ovn-controller-z7f9r" Feb 27 08:07:24 crc kubenswrapper[4612]: I0227 08:07:24.251525 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/1ef66a51-46e7-4b76-a70b-b475d47822fd-var-log-ovn\") pod \"ovn-controller-z7f9r\" (UID: \"1ef66a51-46e7-4b76-a70b-b475d47822fd\") " pod="openstack/ovn-controller-z7f9r" Feb 27 08:07:24 crc kubenswrapper[4612]: I0227 08:07:24.251537 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/1ef66a51-46e7-4b76-a70b-b475d47822fd-var-run-ovn\") pod \"ovn-controller-z7f9r\" (UID: \"1ef66a51-46e7-4b76-a70b-b475d47822fd\") " pod="openstack/ovn-controller-z7f9r" Feb 27 08:07:24 crc kubenswrapper[4612]: I0227 08:07:24.251757 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/1ef66a51-46e7-4b76-a70b-b475d47822fd-var-run-ovn\") pod \"ovn-controller-z7f9r\" (UID: \"1ef66a51-46e7-4b76-a70b-b475d47822fd\") " pod="openstack/ovn-controller-z7f9r" Feb 27 08:07:24 crc kubenswrapper[4612]: I0227 08:07:24.251964 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/1ef66a51-46e7-4b76-a70b-b475d47822fd-var-run\") pod \"ovn-controller-z7f9r\" (UID: \"1ef66a51-46e7-4b76-a70b-b475d47822fd\") " pod="openstack/ovn-controller-z7f9r" Feb 27 08:07:24 crc kubenswrapper[4612]: I0227 08:07:24.253812 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1ef66a51-46e7-4b76-a70b-b475d47822fd-scripts\") pod \"ovn-controller-z7f9r\" (UID: \"1ef66a51-46e7-4b76-a70b-b475d47822fd\") " pod="openstack/ovn-controller-z7f9r" Feb 27 08:07:24 crc kubenswrapper[4612]: I0227 08:07:24.259334 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1ef66a51-46e7-4b76-a70b-b475d47822fd-combined-ca-bundle\") pod \"ovn-controller-z7f9r\" (UID: \"1ef66a51-46e7-4b76-a70b-b475d47822fd\") " pod="openstack/ovn-controller-z7f9r" Feb 27 08:07:24 crc kubenswrapper[4612]: I0227 08:07:24.259791 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/1ef66a51-46e7-4b76-a70b-b475d47822fd-ovn-controller-tls-certs\") pod \"ovn-controller-z7f9r\" (UID: \"1ef66a51-46e7-4b76-a70b-b475d47822fd\") " pod="openstack/ovn-controller-z7f9r" Feb 27 08:07:24 crc kubenswrapper[4612]: I0227 08:07:24.273758 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4kdqc\" (UniqueName: \"kubernetes.io/projected/1ef66a51-46e7-4b76-a70b-b475d47822fd-kube-api-access-4kdqc\") pod \"ovn-controller-z7f9r\" (UID: \"1ef66a51-46e7-4b76-a70b-b475d47822fd\") " pod="openstack/ovn-controller-z7f9r" Feb 27 08:07:24 crc kubenswrapper[4612]: I0227 08:07:24.351391 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-z7f9r" Feb 27 08:07:24 crc kubenswrapper[4612]: I0227 08:07:24.353022 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/20c3f793-4286-4067-a372-37b2de08fa8e-var-run\") pod \"ovn-controller-ovs-x255x\" (UID: \"20c3f793-4286-4067-a372-37b2de08fa8e\") " pod="openstack/ovn-controller-ovs-x255x" Feb 27 08:07:24 crc kubenswrapper[4612]: I0227 08:07:24.353110 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-njlpm\" (UniqueName: \"kubernetes.io/projected/20c3f793-4286-4067-a372-37b2de08fa8e-kube-api-access-njlpm\") pod \"ovn-controller-ovs-x255x\" (UID: \"20c3f793-4286-4067-a372-37b2de08fa8e\") " pod="openstack/ovn-controller-ovs-x255x" Feb 27 08:07:24 crc kubenswrapper[4612]: I0227 08:07:24.353174 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/20c3f793-4286-4067-a372-37b2de08fa8e-scripts\") pod \"ovn-controller-ovs-x255x\" (UID: \"20c3f793-4286-4067-a372-37b2de08fa8e\") " pod="openstack/ovn-controller-ovs-x255x" Feb 27 08:07:24 crc kubenswrapper[4612]: I0227 08:07:24.353196 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/20c3f793-4286-4067-a372-37b2de08fa8e-var-log\") pod \"ovn-controller-ovs-x255x\" (UID: \"20c3f793-4286-4067-a372-37b2de08fa8e\") " pod="openstack/ovn-controller-ovs-x255x" Feb 27 08:07:24 crc kubenswrapper[4612]: I0227 08:07:24.353238 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/20c3f793-4286-4067-a372-37b2de08fa8e-etc-ovs\") pod \"ovn-controller-ovs-x255x\" (UID: \"20c3f793-4286-4067-a372-37b2de08fa8e\") " pod="openstack/ovn-controller-ovs-x255x" Feb 27 08:07:24 crc kubenswrapper[4612]: I0227 08:07:24.353280 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/20c3f793-4286-4067-a372-37b2de08fa8e-var-lib\") pod \"ovn-controller-ovs-x255x\" (UID: \"20c3f793-4286-4067-a372-37b2de08fa8e\") " pod="openstack/ovn-controller-ovs-x255x" Feb 27 08:07:24 crc kubenswrapper[4612]: I0227 08:07:24.353965 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/20c3f793-4286-4067-a372-37b2de08fa8e-var-run\") pod \"ovn-controller-ovs-x255x\" (UID: \"20c3f793-4286-4067-a372-37b2de08fa8e\") " pod="openstack/ovn-controller-ovs-x255x" Feb 27 08:07:24 crc kubenswrapper[4612]: I0227 08:07:24.354570 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/20c3f793-4286-4067-a372-37b2de08fa8e-etc-ovs\") pod \"ovn-controller-ovs-x255x\" (UID: \"20c3f793-4286-4067-a372-37b2de08fa8e\") " pod="openstack/ovn-controller-ovs-x255x" Feb 27 08:07:24 crc kubenswrapper[4612]: I0227 08:07:24.353984 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/20c3f793-4286-4067-a372-37b2de08fa8e-var-lib\") pod \"ovn-controller-ovs-x255x\" (UID: \"20c3f793-4286-4067-a372-37b2de08fa8e\") " pod="openstack/ovn-controller-ovs-x255x" Feb 27 08:07:24 crc kubenswrapper[4612]: I0227 08:07:24.354826 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/20c3f793-4286-4067-a372-37b2de08fa8e-var-log\") pod \"ovn-controller-ovs-x255x\" (UID: \"20c3f793-4286-4067-a372-37b2de08fa8e\") " pod="openstack/ovn-controller-ovs-x255x" Feb 27 08:07:24 crc kubenswrapper[4612]: I0227 08:07:24.355741 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/20c3f793-4286-4067-a372-37b2de08fa8e-scripts\") pod \"ovn-controller-ovs-x255x\" (UID: \"20c3f793-4286-4067-a372-37b2de08fa8e\") " pod="openstack/ovn-controller-ovs-x255x" Feb 27 08:07:24 crc kubenswrapper[4612]: I0227 08:07:24.395412 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-njlpm\" (UniqueName: \"kubernetes.io/projected/20c3f793-4286-4067-a372-37b2de08fa8e-kube-api-access-njlpm\") pod \"ovn-controller-ovs-x255x\" (UID: \"20c3f793-4286-4067-a372-37b2de08fa8e\") " pod="openstack/ovn-controller-ovs-x255x" Feb 27 08:07:24 crc kubenswrapper[4612]: I0227 08:07:24.687930 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-x255x" Feb 27 08:07:24 crc kubenswrapper[4612]: I0227 08:07:24.955942 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Feb 27 08:07:24 crc kubenswrapper[4612]: I0227 08:07:24.957421 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Feb 27 08:07:24 crc kubenswrapper[4612]: I0227 08:07:24.963114 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-6tnv9" Feb 27 08:07:24 crc kubenswrapper[4612]: I0227 08:07:24.963256 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Feb 27 08:07:24 crc kubenswrapper[4612]: I0227 08:07:24.963257 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Feb 27 08:07:24 crc kubenswrapper[4612]: I0227 08:07:24.963365 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Feb 27 08:07:24 crc kubenswrapper[4612]: I0227 08:07:24.963483 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Feb 27 08:07:24 crc kubenswrapper[4612]: I0227 08:07:24.971606 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Feb 27 08:07:25 crc kubenswrapper[4612]: I0227 08:07:25.076522 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s59t8\" (UniqueName: \"kubernetes.io/projected/1336636f-bc7a-45d4-a2ee-886a6f64ecd3-kube-api-access-s59t8\") pod \"ovsdbserver-nb-0\" (UID: \"1336636f-bc7a-45d4-a2ee-886a6f64ecd3\") " pod="openstack/ovsdbserver-nb-0" Feb 27 08:07:25 crc kubenswrapper[4612]: I0227 08:07:25.076577 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"ovsdbserver-nb-0\" (UID: \"1336636f-bc7a-45d4-a2ee-886a6f64ecd3\") " pod="openstack/ovsdbserver-nb-0" Feb 27 08:07:25 crc kubenswrapper[4612]: I0227 08:07:25.076616 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/1336636f-bc7a-45d4-a2ee-886a6f64ecd3-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"1336636f-bc7a-45d4-a2ee-886a6f64ecd3\") " pod="openstack/ovsdbserver-nb-0" Feb 27 08:07:25 crc kubenswrapper[4612]: I0227 08:07:25.076655 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/1336636f-bc7a-45d4-a2ee-886a6f64ecd3-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"1336636f-bc7a-45d4-a2ee-886a6f64ecd3\") " pod="openstack/ovsdbserver-nb-0" Feb 27 08:07:25 crc kubenswrapper[4612]: I0227 08:07:25.076713 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1336636f-bc7a-45d4-a2ee-886a6f64ecd3-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"1336636f-bc7a-45d4-a2ee-886a6f64ecd3\") " pod="openstack/ovsdbserver-nb-0" Feb 27 08:07:25 crc kubenswrapper[4612]: I0227 08:07:25.076752 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/1336636f-bc7a-45d4-a2ee-886a6f64ecd3-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"1336636f-bc7a-45d4-a2ee-886a6f64ecd3\") " pod="openstack/ovsdbserver-nb-0" Feb 27 08:07:25 crc kubenswrapper[4612]: I0227 08:07:25.076808 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1336636f-bc7a-45d4-a2ee-886a6f64ecd3-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"1336636f-bc7a-45d4-a2ee-886a6f64ecd3\") " pod="openstack/ovsdbserver-nb-0" Feb 27 08:07:25 crc kubenswrapper[4612]: I0227 08:07:25.076838 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1336636f-bc7a-45d4-a2ee-886a6f64ecd3-config\") pod \"ovsdbserver-nb-0\" (UID: \"1336636f-bc7a-45d4-a2ee-886a6f64ecd3\") " pod="openstack/ovsdbserver-nb-0" Feb 27 08:07:25 crc kubenswrapper[4612]: I0227 08:07:25.178611 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/1336636f-bc7a-45d4-a2ee-886a6f64ecd3-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"1336636f-bc7a-45d4-a2ee-886a6f64ecd3\") " pod="openstack/ovsdbserver-nb-0" Feb 27 08:07:25 crc kubenswrapper[4612]: I0227 08:07:25.178760 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1336636f-bc7a-45d4-a2ee-886a6f64ecd3-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"1336636f-bc7a-45d4-a2ee-886a6f64ecd3\") " pod="openstack/ovsdbserver-nb-0" Feb 27 08:07:25 crc kubenswrapper[4612]: I0227 08:07:25.179025 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1336636f-bc7a-45d4-a2ee-886a6f64ecd3-config\") pod \"ovsdbserver-nb-0\" (UID: \"1336636f-bc7a-45d4-a2ee-886a6f64ecd3\") " pod="openstack/ovsdbserver-nb-0" Feb 27 08:07:25 crc kubenswrapper[4612]: I0227 08:07:25.179088 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s59t8\" (UniqueName: \"kubernetes.io/projected/1336636f-bc7a-45d4-a2ee-886a6f64ecd3-kube-api-access-s59t8\") pod \"ovsdbserver-nb-0\" (UID: \"1336636f-bc7a-45d4-a2ee-886a6f64ecd3\") " pod="openstack/ovsdbserver-nb-0" Feb 27 08:07:25 crc kubenswrapper[4612]: I0227 08:07:25.179113 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"ovsdbserver-nb-0\" (UID: \"1336636f-bc7a-45d4-a2ee-886a6f64ecd3\") " pod="openstack/ovsdbserver-nb-0" Feb 27 08:07:25 crc kubenswrapper[4612]: I0227 08:07:25.179154 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/1336636f-bc7a-45d4-a2ee-886a6f64ecd3-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"1336636f-bc7a-45d4-a2ee-886a6f64ecd3\") " pod="openstack/ovsdbserver-nb-0" Feb 27 08:07:25 crc kubenswrapper[4612]: I0227 08:07:25.179195 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/1336636f-bc7a-45d4-a2ee-886a6f64ecd3-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"1336636f-bc7a-45d4-a2ee-886a6f64ecd3\") " pod="openstack/ovsdbserver-nb-0" Feb 27 08:07:25 crc kubenswrapper[4612]: I0227 08:07:25.179632 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1336636f-bc7a-45d4-a2ee-886a6f64ecd3-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"1336636f-bc7a-45d4-a2ee-886a6f64ecd3\") " pod="openstack/ovsdbserver-nb-0" Feb 27 08:07:25 crc kubenswrapper[4612]: I0227 08:07:25.179821 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/1336636f-bc7a-45d4-a2ee-886a6f64ecd3-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"1336636f-bc7a-45d4-a2ee-886a6f64ecd3\") " pod="openstack/ovsdbserver-nb-0" Feb 27 08:07:25 crc kubenswrapper[4612]: I0227 08:07:25.180061 4612 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"ovsdbserver-nb-0\" (UID: \"1336636f-bc7a-45d4-a2ee-886a6f64ecd3\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/ovsdbserver-nb-0" Feb 27 08:07:25 crc kubenswrapper[4612]: I0227 08:07:25.180325 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1336636f-bc7a-45d4-a2ee-886a6f64ecd3-config\") pod \"ovsdbserver-nb-0\" (UID: \"1336636f-bc7a-45d4-a2ee-886a6f64ecd3\") " pod="openstack/ovsdbserver-nb-0" Feb 27 08:07:25 crc kubenswrapper[4612]: I0227 08:07:25.180536 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1336636f-bc7a-45d4-a2ee-886a6f64ecd3-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"1336636f-bc7a-45d4-a2ee-886a6f64ecd3\") " pod="openstack/ovsdbserver-nb-0" Feb 27 08:07:25 crc kubenswrapper[4612]: I0227 08:07:25.197157 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"ovsdbserver-nb-0\" (UID: \"1336636f-bc7a-45d4-a2ee-886a6f64ecd3\") " pod="openstack/ovsdbserver-nb-0" Feb 27 08:07:25 crc kubenswrapper[4612]: I0227 08:07:25.217536 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s59t8\" (UniqueName: \"kubernetes.io/projected/1336636f-bc7a-45d4-a2ee-886a6f64ecd3-kube-api-access-s59t8\") pod \"ovsdbserver-nb-0\" (UID: \"1336636f-bc7a-45d4-a2ee-886a6f64ecd3\") " pod="openstack/ovsdbserver-nb-0" Feb 27 08:07:25 crc kubenswrapper[4612]: I0227 08:07:25.225093 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/1336636f-bc7a-45d4-a2ee-886a6f64ecd3-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"1336636f-bc7a-45d4-a2ee-886a6f64ecd3\") " pod="openstack/ovsdbserver-nb-0" Feb 27 08:07:25 crc kubenswrapper[4612]: I0227 08:07:25.231832 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/1336636f-bc7a-45d4-a2ee-886a6f64ecd3-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"1336636f-bc7a-45d4-a2ee-886a6f64ecd3\") " pod="openstack/ovsdbserver-nb-0" Feb 27 08:07:25 crc kubenswrapper[4612]: I0227 08:07:25.239020 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1336636f-bc7a-45d4-a2ee-886a6f64ecd3-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"1336636f-bc7a-45d4-a2ee-886a6f64ecd3\") " pod="openstack/ovsdbserver-nb-0" Feb 27 08:07:25 crc kubenswrapper[4612]: I0227 08:07:25.292736 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Feb 27 08:07:26 crc kubenswrapper[4612]: I0227 08:07:26.684172 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Feb 27 08:07:26 crc kubenswrapper[4612]: I0227 08:07:26.689164 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Feb 27 08:07:26 crc kubenswrapper[4612]: I0227 08:07:26.691358 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Feb 27 08:07:26 crc kubenswrapper[4612]: I0227 08:07:26.696393 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Feb 27 08:07:26 crc kubenswrapper[4612]: I0227 08:07:26.700297 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Feb 27 08:07:26 crc kubenswrapper[4612]: I0227 08:07:26.700443 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-zmw8g" Feb 27 08:07:26 crc kubenswrapper[4612]: I0227 08:07:26.700483 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Feb 27 08:07:26 crc kubenswrapper[4612]: I0227 08:07:26.866801 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ddac95c9-8d81-420b-836e-d1902ce8beb5-config\") pod \"ovsdbserver-sb-0\" (UID: \"ddac95c9-8d81-420b-836e-d1902ce8beb5\") " pod="openstack/ovsdbserver-sb-0" Feb 27 08:07:26 crc kubenswrapper[4612]: I0227 08:07:26.866919 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/ddac95c9-8d81-420b-836e-d1902ce8beb5-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"ddac95c9-8d81-420b-836e-d1902ce8beb5\") " pod="openstack/ovsdbserver-sb-0" Feb 27 08:07:26 crc kubenswrapper[4612]: I0227 08:07:26.867002 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"ovsdbserver-sb-0\" (UID: \"ddac95c9-8d81-420b-836e-d1902ce8beb5\") " pod="openstack/ovsdbserver-sb-0" Feb 27 08:07:26 crc kubenswrapper[4612]: I0227 08:07:26.867051 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wdbzd\" (UniqueName: \"kubernetes.io/projected/ddac95c9-8d81-420b-836e-d1902ce8beb5-kube-api-access-wdbzd\") pod \"ovsdbserver-sb-0\" (UID: \"ddac95c9-8d81-420b-836e-d1902ce8beb5\") " pod="openstack/ovsdbserver-sb-0" Feb 27 08:07:26 crc kubenswrapper[4612]: I0227 08:07:26.867189 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/ddac95c9-8d81-420b-836e-d1902ce8beb5-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"ddac95c9-8d81-420b-836e-d1902ce8beb5\") " pod="openstack/ovsdbserver-sb-0" Feb 27 08:07:26 crc kubenswrapper[4612]: I0227 08:07:26.867292 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ddac95c9-8d81-420b-836e-d1902ce8beb5-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"ddac95c9-8d81-420b-836e-d1902ce8beb5\") " pod="openstack/ovsdbserver-sb-0" Feb 27 08:07:26 crc kubenswrapper[4612]: I0227 08:07:26.867318 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ddac95c9-8d81-420b-836e-d1902ce8beb5-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"ddac95c9-8d81-420b-836e-d1902ce8beb5\") " pod="openstack/ovsdbserver-sb-0" Feb 27 08:07:26 crc kubenswrapper[4612]: I0227 08:07:26.867375 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/ddac95c9-8d81-420b-836e-d1902ce8beb5-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"ddac95c9-8d81-420b-836e-d1902ce8beb5\") " pod="openstack/ovsdbserver-sb-0" Feb 27 08:07:26 crc kubenswrapper[4612]: I0227 08:07:26.969369 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/ddac95c9-8d81-420b-836e-d1902ce8beb5-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"ddac95c9-8d81-420b-836e-d1902ce8beb5\") " pod="openstack/ovsdbserver-sb-0" Feb 27 08:07:26 crc kubenswrapper[4612]: I0227 08:07:26.969801 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"ovsdbserver-sb-0\" (UID: \"ddac95c9-8d81-420b-836e-d1902ce8beb5\") " pod="openstack/ovsdbserver-sb-0" Feb 27 08:07:26 crc kubenswrapper[4612]: I0227 08:07:26.969934 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wdbzd\" (UniqueName: \"kubernetes.io/projected/ddac95c9-8d81-420b-836e-d1902ce8beb5-kube-api-access-wdbzd\") pod \"ovsdbserver-sb-0\" (UID: \"ddac95c9-8d81-420b-836e-d1902ce8beb5\") " pod="openstack/ovsdbserver-sb-0" Feb 27 08:07:26 crc kubenswrapper[4612]: I0227 08:07:26.970000 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/ddac95c9-8d81-420b-836e-d1902ce8beb5-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"ddac95c9-8d81-420b-836e-d1902ce8beb5\") " pod="openstack/ovsdbserver-sb-0" Feb 27 08:07:26 crc kubenswrapper[4612]: I0227 08:07:26.970011 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/ddac95c9-8d81-420b-836e-d1902ce8beb5-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"ddac95c9-8d81-420b-836e-d1902ce8beb5\") " pod="openstack/ovsdbserver-sb-0" Feb 27 08:07:26 crc kubenswrapper[4612]: I0227 08:07:26.970233 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ddac95c9-8d81-420b-836e-d1902ce8beb5-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"ddac95c9-8d81-420b-836e-d1902ce8beb5\") " pod="openstack/ovsdbserver-sb-0" Feb 27 08:07:26 crc kubenswrapper[4612]: I0227 08:07:26.970386 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ddac95c9-8d81-420b-836e-d1902ce8beb5-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"ddac95c9-8d81-420b-836e-d1902ce8beb5\") " pod="openstack/ovsdbserver-sb-0" Feb 27 08:07:26 crc kubenswrapper[4612]: I0227 08:07:26.970508 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/ddac95c9-8d81-420b-836e-d1902ce8beb5-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"ddac95c9-8d81-420b-836e-d1902ce8beb5\") " pod="openstack/ovsdbserver-sb-0" Feb 27 08:07:26 crc kubenswrapper[4612]: I0227 08:07:26.970662 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ddac95c9-8d81-420b-836e-d1902ce8beb5-config\") pod \"ovsdbserver-sb-0\" (UID: \"ddac95c9-8d81-420b-836e-d1902ce8beb5\") " pod="openstack/ovsdbserver-sb-0" Feb 27 08:07:26 crc kubenswrapper[4612]: I0227 08:07:26.970781 4612 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"ovsdbserver-sb-0\" (UID: \"ddac95c9-8d81-420b-836e-d1902ce8beb5\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/ovsdbserver-sb-0" Feb 27 08:07:26 crc kubenswrapper[4612]: I0227 08:07:26.972923 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ddac95c9-8d81-420b-836e-d1902ce8beb5-config\") pod \"ovsdbserver-sb-0\" (UID: \"ddac95c9-8d81-420b-836e-d1902ce8beb5\") " pod="openstack/ovsdbserver-sb-0" Feb 27 08:07:26 crc kubenswrapper[4612]: I0227 08:07:26.973930 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ddac95c9-8d81-420b-836e-d1902ce8beb5-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"ddac95c9-8d81-420b-836e-d1902ce8beb5\") " pod="openstack/ovsdbserver-sb-0" Feb 27 08:07:26 crc kubenswrapper[4612]: I0227 08:07:26.977419 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/ddac95c9-8d81-420b-836e-d1902ce8beb5-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"ddac95c9-8d81-420b-836e-d1902ce8beb5\") " pod="openstack/ovsdbserver-sb-0" Feb 27 08:07:26 crc kubenswrapper[4612]: I0227 08:07:26.980034 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ddac95c9-8d81-420b-836e-d1902ce8beb5-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"ddac95c9-8d81-420b-836e-d1902ce8beb5\") " pod="openstack/ovsdbserver-sb-0" Feb 27 08:07:26 crc kubenswrapper[4612]: I0227 08:07:26.983487 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/ddac95c9-8d81-420b-836e-d1902ce8beb5-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"ddac95c9-8d81-420b-836e-d1902ce8beb5\") " pod="openstack/ovsdbserver-sb-0" Feb 27 08:07:26 crc kubenswrapper[4612]: I0227 08:07:26.988597 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wdbzd\" (UniqueName: \"kubernetes.io/projected/ddac95c9-8d81-420b-836e-d1902ce8beb5-kube-api-access-wdbzd\") pod \"ovsdbserver-sb-0\" (UID: \"ddac95c9-8d81-420b-836e-d1902ce8beb5\") " pod="openstack/ovsdbserver-sb-0" Feb 27 08:07:27 crc kubenswrapper[4612]: I0227 08:07:27.006956 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"ovsdbserver-sb-0\" (UID: \"ddac95c9-8d81-420b-836e-d1902ce8beb5\") " pod="openstack/ovsdbserver-sb-0" Feb 27 08:07:27 crc kubenswrapper[4612]: I0227 08:07:27.083173 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Feb 27 08:07:40 crc kubenswrapper[4612]: E0227 08:07:40.771483 4612 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified" Feb 27 08:07:40 crc kubenswrapper[4612]: E0227 08:07:40.773081 4612 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:setup-container,Image:quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified,Command:[sh -c cp /tmp/erlang-cookie-secret/.erlang.cookie /var/lib/rabbitmq/.erlang.cookie && chmod 600 /var/lib/rabbitmq/.erlang.cookie ; cp /tmp/rabbitmq-plugins/enabled_plugins /operator/enabled_plugins ; echo '[default]' > /var/lib/rabbitmq/.rabbitmqadmin.conf && sed -e 's/default_user/username/' -e 's/default_pass/password/' /tmp/default_user.conf >> /var/lib/rabbitmq/.rabbitmqadmin.conf && chmod 600 /var/lib/rabbitmq/.rabbitmqadmin.conf ; sleep 30],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Requests:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:plugins-conf,ReadOnly:false,MountPath:/tmp/rabbitmq-plugins/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-erlang-cookie,ReadOnly:false,MountPath:/var/lib/rabbitmq/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:erlang-cookie-secret,ReadOnly:false,MountPath:/tmp/erlang-cookie-secret/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-plugins,ReadOnly:false,MountPath:/operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:persistence,ReadOnly:false,MountPath:/var/lib/rabbitmq/mnesia/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-confd,ReadOnly:false,MountPath:/tmp/default_user.conf,SubPath:default_user.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-86pqk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cell1-server-0_openstack(6d1ba6db-1cae-461e-9716-5d5b36350218): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 27 08:07:40 crc kubenswrapper[4612]: E0227 08:07:40.774972 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/rabbitmq-cell1-server-0" podUID="6d1ba6db-1cae-461e-9716-5d5b36350218" Feb 27 08:07:41 crc kubenswrapper[4612]: E0227 08:07:41.205713 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified\\\"\"" pod="openstack/rabbitmq-cell1-server-0" podUID="6d1ba6db-1cae-461e-9716-5d5b36350218" Feb 27 08:07:41 crc kubenswrapper[4612]: E0227 08:07:41.533809 4612 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-memcached:current-podified" Feb 27 08:07:41 crc kubenswrapper[4612]: E0227 08:07:41.534023 4612 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:memcached,Image:quay.io/podified-antelope-centos9/openstack-memcached:current-podified,Command:[/usr/bin/dumb-init -- /usr/local/bin/kolla_start],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:memcached,HostPort:0,ContainerPort:11211,Protocol:TCP,HostIP:,},ContainerPort{Name:memcached-tls,HostPort:0,ContainerPort:11212,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:POD_IPS,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIPs,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:CONFIG_HASH,Value:n86hbhf5h5f8hfdh646h579hc5h659h5f7h54ch6h656h557h6bh66bh67chb9h58chb4hfbh677h599h9fh565h56h7fh5c4h649h56hfch6bq,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/src,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kolla-config,ReadOnly:true,MountPath:/var/lib/kolla/config_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:memcached-tls-certs,ReadOnly:true,MountPath:/var/lib/config-data/tls/certs/memcached.crt,SubPath:tls.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:memcached-tls-certs,ReadOnly:true,MountPath:/var/lib/config-data/tls/private/memcached.key,SubPath:tls.key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-dxx5j,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 11211 },Host:,},GRPC:nil,},InitialDelaySeconds:3,TimeoutSeconds:5,PeriodSeconds:3,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 11211 },Host:,},GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42457,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42457,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod memcached-0_openstack(d19bfe90-c5ca-4089-b4f0-bf3eccff30c8): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 27 08:07:41 crc kubenswrapper[4612]: E0227 08:07:41.535734 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"memcached\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/memcached-0" podUID="d19bfe90-c5ca-4089-b4f0-bf3eccff30c8" Feb 27 08:07:41 crc kubenswrapper[4612]: E0227 08:07:41.552622 4612 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified" Feb 27 08:07:41 crc kubenswrapper[4612]: E0227 08:07:41.552873 4612 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:setup-container,Image:quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified,Command:[sh -c cp /tmp/erlang-cookie-secret/.erlang.cookie /var/lib/rabbitmq/.erlang.cookie && chmod 600 /var/lib/rabbitmq/.erlang.cookie ; cp /tmp/rabbitmq-plugins/enabled_plugins /operator/enabled_plugins ; echo '[default]' > /var/lib/rabbitmq/.rabbitmqadmin.conf && sed -e 's/default_user/username/' -e 's/default_pass/password/' /tmp/default_user.conf >> /var/lib/rabbitmq/.rabbitmqadmin.conf && chmod 600 /var/lib/rabbitmq/.rabbitmqadmin.conf ; sleep 30],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Requests:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:plugins-conf,ReadOnly:false,MountPath:/tmp/rabbitmq-plugins/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-erlang-cookie,ReadOnly:false,MountPath:/var/lib/rabbitmq/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:erlang-cookie-secret,ReadOnly:false,MountPath:/tmp/erlang-cookie-secret/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-plugins,ReadOnly:false,MountPath:/operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:persistence,ReadOnly:false,MountPath:/var/lib/rabbitmq/mnesia/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-confd,ReadOnly:false,MountPath:/tmp/default_user.conf,SubPath:default_user.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-vjjpx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-server-0_openstack(0da436b1-803d-41bb-94ab-9cfd6fa769eb): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 27 08:07:41 crc kubenswrapper[4612]: E0227 08:07:41.554142 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/rabbitmq-server-0" podUID="0da436b1-803d-41bb-94ab-9cfd6fa769eb" Feb 27 08:07:41 crc kubenswrapper[4612]: E0227 08:07:41.555244 4612 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-mariadb:current-podified" Feb 27 08:07:41 crc kubenswrapper[4612]: E0227 08:07:41.555433 4612 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:mysql-bootstrap,Image:quay.io/podified-antelope-centos9/openstack-mariadb:current-podified,Command:[bash /var/lib/operator-scripts/mysql_bootstrap.sh],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:True,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:mysql-db,ReadOnly:false,MountPath:/var/lib/mysql,SubPath:mysql,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data-default,ReadOnly:true,MountPath:/var/lib/config-data/default,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data-generated,ReadOnly:false,MountPath:/var/lib/config-data/generated,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:operator-scripts,ReadOnly:true,MountPath:/var/lib/operator-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kolla-config,ReadOnly:true,MountPath:/var/lib/kolla/config_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-drmd6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod openstack-cell1-galera-0_openstack(e02a40f2-3de1-4f01-84b0-2fe56acdd396): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 27 08:07:41 crc kubenswrapper[4612]: E0227 08:07:41.556604 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mysql-bootstrap\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/openstack-cell1-galera-0" podUID="e02a40f2-3de1-4f01-84b0-2fe56acdd396" Feb 27 08:07:42 crc kubenswrapper[4612]: E0227 08:07:42.224075 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mysql-bootstrap\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-mariadb:current-podified\\\"\"" pod="openstack/openstack-cell1-galera-0" podUID="e02a40f2-3de1-4f01-84b0-2fe56acdd396" Feb 27 08:07:42 crc kubenswrapper[4612]: E0227 08:07:42.224094 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"memcached\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-memcached:current-podified\\\"\"" pod="openstack/memcached-0" podUID="d19bfe90-c5ca-4089-b4f0-bf3eccff30c8" Feb 27 08:07:42 crc kubenswrapper[4612]: E0227 08:07:42.224143 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified\\\"\"" pod="openstack/rabbitmq-server-0" podUID="0da436b1-803d-41bb-94ab-9cfd6fa769eb" Feb 27 08:07:42 crc kubenswrapper[4612]: E0227 08:07:42.277149 4612 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Feb 27 08:07:42 crc kubenswrapper[4612]: E0227 08:07:42.277298 4612 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nffh5bdhf4h5f8h79h55h77h58fh56dh7bh6fh578hbch55dh68h56bhd9h65dh57ch658hc9h566h666h688h58h65dh684h5d7h6ch575h5d6h88q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-pzl2m,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-675f4bcbfc-7smgg_openstack(e2653a32-b431-40b0-b9de-6a431d7f0644): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 27 08:07:42 crc kubenswrapper[4612]: E0227 08:07:42.279045 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-675f4bcbfc-7smgg" podUID="e2653a32-b431-40b0-b9de-6a431d7f0644" Feb 27 08:07:42 crc kubenswrapper[4612]: E0227 08:07:42.403357 4612 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Feb 27 08:07:42 crc kubenswrapper[4612]: E0227 08:07:42.403577 4612 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n68chd6h679hbfh55fhc6h5ffh5d8h94h56ch589hb4hc5h57bh677hcdh655h8dh667h675h654h66ch567h8fh659h5b4h675h566h55bh54h67dh6dq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-jssxq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-666b6646f7-jghh4_openstack(57036ec4-7696-4e30-b365-44176b0afdc2): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 27 08:07:42 crc kubenswrapper[4612]: E0227 08:07:42.403980 4612 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Feb 27 08:07:42 crc kubenswrapper[4612]: E0227 08:07:42.404073 4612 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n659h4h664hbh658h587h67ch89h587h8fh679hc6hf9h55fh644h5d5h698h68dh5cdh5ffh669h54ch9h689hb8hd4h5bfhd8h5d7h5fh665h574q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rh2tp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-57d769cc4f-bz4dn_openstack(72e3b749-ce9a-4480-8873-d8c9b6495cbc): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 27 08:07:42 crc kubenswrapper[4612]: E0227 08:07:42.405236 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-57d769cc4f-bz4dn" podUID="72e3b749-ce9a-4480-8873-d8c9b6495cbc" Feb 27 08:07:42 crc kubenswrapper[4612]: E0227 08:07:42.405291 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-666b6646f7-jghh4" podUID="57036ec4-7696-4e30-b365-44176b0afdc2" Feb 27 08:07:42 crc kubenswrapper[4612]: E0227 08:07:42.406650 4612 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Feb 27 08:07:42 crc kubenswrapper[4612]: E0227 08:07:42.406836 4612 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndfhb5h667h568h584h5f9h58dh565h664h587h597h577h64bh5c4h66fh647hbdh68ch5c5h68dh686h5f7h64hd7hc6h55fh57bh98h57fh87h5fh57fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-2hwtr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-78dd6ddcc-579pm_openstack(93fcf290-888e-43d5-b46b-0a235776bdbc): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 27 08:07:42 crc kubenswrapper[4612]: E0227 08:07:42.407970 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-78dd6ddcc-579pm" podUID="93fcf290-888e-43d5-b46b-0a235776bdbc" Feb 27 08:07:43 crc kubenswrapper[4612]: E0227 08:07:43.229047 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified\\\"\"" pod="openstack/dnsmasq-dns-57d769cc4f-bz4dn" podUID="72e3b749-ce9a-4480-8873-d8c9b6495cbc" Feb 27 08:07:43 crc kubenswrapper[4612]: E0227 08:07:43.236439 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified\\\"\"" pod="openstack/dnsmasq-dns-666b6646f7-jghh4" podUID="57036ec4-7696-4e30-b365-44176b0afdc2" Feb 27 08:07:43 crc kubenswrapper[4612]: I0227 08:07:43.363424 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-z7f9r"] Feb 27 08:07:43 crc kubenswrapper[4612]: I0227 08:07:43.609148 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Feb 27 08:07:43 crc kubenswrapper[4612]: W0227 08:07:43.897944 4612 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podddac95c9_8d81_420b_836e_d1902ce8beb5.slice/crio-8e5d67a7201867cba0a0ad2a7099489b3fcd3217231f151b73986ab8cd301790 WatchSource:0}: Error finding container 8e5d67a7201867cba0a0ad2a7099489b3fcd3217231f151b73986ab8cd301790: Status 404 returned error can't find the container with id 8e5d67a7201867cba0a0ad2a7099489b3fcd3217231f151b73986ab8cd301790 Feb 27 08:07:43 crc kubenswrapper[4612]: W0227 08:07:43.903908 4612 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1ef66a51_46e7_4b76_a70b_b475d47822fd.slice/crio-9287158ce48823c02d401389985847065364a6987290c655370f3066b8d796de WatchSource:0}: Error finding container 9287158ce48823c02d401389985847065364a6987290c655370f3066b8d796de: Status 404 returned error can't find the container with id 9287158ce48823c02d401389985847065364a6987290c655370f3066b8d796de Feb 27 08:07:44 crc kubenswrapper[4612]: I0227 08:07:44.001004 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-579pm" Feb 27 08:07:44 crc kubenswrapper[4612]: I0227 08:07:44.044267 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-7smgg" Feb 27 08:07:44 crc kubenswrapper[4612]: I0227 08:07:44.168988 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/93fcf290-888e-43d5-b46b-0a235776bdbc-config\") pod \"93fcf290-888e-43d5-b46b-0a235776bdbc\" (UID: \"93fcf290-888e-43d5-b46b-0a235776bdbc\") " Feb 27 08:07:44 crc kubenswrapper[4612]: I0227 08:07:44.169332 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2hwtr\" (UniqueName: \"kubernetes.io/projected/93fcf290-888e-43d5-b46b-0a235776bdbc-kube-api-access-2hwtr\") pod \"93fcf290-888e-43d5-b46b-0a235776bdbc\" (UID: \"93fcf290-888e-43d5-b46b-0a235776bdbc\") " Feb 27 08:07:44 crc kubenswrapper[4612]: I0227 08:07:44.169420 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pzl2m\" (UniqueName: \"kubernetes.io/projected/e2653a32-b431-40b0-b9de-6a431d7f0644-kube-api-access-pzl2m\") pod \"e2653a32-b431-40b0-b9de-6a431d7f0644\" (UID: \"e2653a32-b431-40b0-b9de-6a431d7f0644\") " Feb 27 08:07:44 crc kubenswrapper[4612]: I0227 08:07:44.169437 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/93fcf290-888e-43d5-b46b-0a235776bdbc-dns-svc\") pod \"93fcf290-888e-43d5-b46b-0a235776bdbc\" (UID: \"93fcf290-888e-43d5-b46b-0a235776bdbc\") " Feb 27 08:07:44 crc kubenswrapper[4612]: I0227 08:07:44.169458 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e2653a32-b431-40b0-b9de-6a431d7f0644-config\") pod \"e2653a32-b431-40b0-b9de-6a431d7f0644\" (UID: \"e2653a32-b431-40b0-b9de-6a431d7f0644\") " Feb 27 08:07:44 crc kubenswrapper[4612]: I0227 08:07:44.169746 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/93fcf290-888e-43d5-b46b-0a235776bdbc-config" (OuterVolumeSpecName: "config") pod "93fcf290-888e-43d5-b46b-0a235776bdbc" (UID: "93fcf290-888e-43d5-b46b-0a235776bdbc"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 08:07:44 crc kubenswrapper[4612]: I0227 08:07:44.170163 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/93fcf290-888e-43d5-b46b-0a235776bdbc-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "93fcf290-888e-43d5-b46b-0a235776bdbc" (UID: "93fcf290-888e-43d5-b46b-0a235776bdbc"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 08:07:44 crc kubenswrapper[4612]: I0227 08:07:44.170433 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e2653a32-b431-40b0-b9de-6a431d7f0644-config" (OuterVolumeSpecName: "config") pod "e2653a32-b431-40b0-b9de-6a431d7f0644" (UID: "e2653a32-b431-40b0-b9de-6a431d7f0644"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 08:07:44 crc kubenswrapper[4612]: I0227 08:07:44.176852 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/93fcf290-888e-43d5-b46b-0a235776bdbc-kube-api-access-2hwtr" (OuterVolumeSpecName: "kube-api-access-2hwtr") pod "93fcf290-888e-43d5-b46b-0a235776bdbc" (UID: "93fcf290-888e-43d5-b46b-0a235776bdbc"). InnerVolumeSpecName "kube-api-access-2hwtr". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 08:07:44 crc kubenswrapper[4612]: I0227 08:07:44.180829 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e2653a32-b431-40b0-b9de-6a431d7f0644-kube-api-access-pzl2m" (OuterVolumeSpecName: "kube-api-access-pzl2m") pod "e2653a32-b431-40b0-b9de-6a431d7f0644" (UID: "e2653a32-b431-40b0-b9de-6a431d7f0644"). InnerVolumeSpecName "kube-api-access-pzl2m". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 08:07:44 crc kubenswrapper[4612]: I0227 08:07:44.233138 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"ddac95c9-8d81-420b-836e-d1902ce8beb5","Type":"ContainerStarted","Data":"8e5d67a7201867cba0a0ad2a7099489b3fcd3217231f151b73986ab8cd301790"} Feb 27 08:07:44 crc kubenswrapper[4612]: I0227 08:07:44.234726 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"f189e394-0c97-48a4-9c55-4bd6710923f1","Type":"ContainerStarted","Data":"a05a735130fbad58220807c9fcbc842cec9ed97bd45b1d0fd4b4dc1f3f5aa50f"} Feb 27 08:07:44 crc kubenswrapper[4612]: I0227 08:07:44.237102 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"7c51bfaf-2c6a-49aa-afe7-eb4789351857","Type":"ContainerStarted","Data":"6631aea4a97b9613eab069c3da8c94e2ca649e395147956fd3276596b9588520"} Feb 27 08:07:44 crc kubenswrapper[4612]: I0227 08:07:44.237234 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Feb 27 08:07:44 crc kubenswrapper[4612]: I0227 08:07:44.238249 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-7smgg" event={"ID":"e2653a32-b431-40b0-b9de-6a431d7f0644","Type":"ContainerDied","Data":"63f5baf13892f51a4eba21cfaa2db5067796cd93aea442bf49b7482c73cf2162"} Feb 27 08:07:44 crc kubenswrapper[4612]: I0227 08:07:44.238257 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-7smgg" Feb 27 08:07:44 crc kubenswrapper[4612]: I0227 08:07:44.239171 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-z7f9r" event={"ID":"1ef66a51-46e7-4b76-a70b-b475d47822fd","Type":"ContainerStarted","Data":"9287158ce48823c02d401389985847065364a6987290c655370f3066b8d796de"} Feb 27 08:07:44 crc kubenswrapper[4612]: I0227 08:07:44.240249 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-579pm" event={"ID":"93fcf290-888e-43d5-b46b-0a235776bdbc","Type":"ContainerDied","Data":"bef2679160619fc74b69b6fb9240cff0be2dd3e92d67f536c29de144de780ec3"} Feb 27 08:07:44 crc kubenswrapper[4612]: I0227 08:07:44.240296 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-579pm" Feb 27 08:07:44 crc kubenswrapper[4612]: I0227 08:07:44.271609 4612 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/93fcf290-888e-43d5-b46b-0a235776bdbc-config\") on node \"crc\" DevicePath \"\"" Feb 27 08:07:44 crc kubenswrapper[4612]: I0227 08:07:44.271641 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2hwtr\" (UniqueName: \"kubernetes.io/projected/93fcf290-888e-43d5-b46b-0a235776bdbc-kube-api-access-2hwtr\") on node \"crc\" DevicePath \"\"" Feb 27 08:07:44 crc kubenswrapper[4612]: I0227 08:07:44.271652 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pzl2m\" (UniqueName: \"kubernetes.io/projected/e2653a32-b431-40b0-b9de-6a431d7f0644-kube-api-access-pzl2m\") on node \"crc\" DevicePath \"\"" Feb 27 08:07:44 crc kubenswrapper[4612]: I0227 08:07:44.271661 4612 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/93fcf290-888e-43d5-b46b-0a235776bdbc-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 27 08:07:44 crc kubenswrapper[4612]: I0227 08:07:44.271669 4612 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e2653a32-b431-40b0-b9de-6a431d7f0644-config\") on node \"crc\" DevicePath \"\"" Feb 27 08:07:44 crc kubenswrapper[4612]: I0227 08:07:44.279422 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=1.7391483 podStartE2EDuration="24.279403128s" podCreationTimestamp="2026-02-27 08:07:20 +0000 UTC" firstStartedPulling="2026-02-27 08:07:21.428269112 +0000 UTC m=+1099.282199110" lastFinishedPulling="2026-02-27 08:07:43.96852394 +0000 UTC m=+1121.822453938" observedRunningTime="2026-02-27 08:07:44.273267212 +0000 UTC m=+1122.127197210" watchObservedRunningTime="2026-02-27 08:07:44.279403128 +0000 UTC m=+1122.133333126" Feb 27 08:07:44 crc kubenswrapper[4612]: I0227 08:07:44.299440 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Feb 27 08:07:44 crc kubenswrapper[4612]: W0227 08:07:44.307452 4612 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1336636f_bc7a_45d4_a2ee_886a6f64ecd3.slice/crio-320b3e8c0eface7724a8b81cfff96ecd2d62ee08f0496163f545eabbf4af789c WatchSource:0}: Error finding container 320b3e8c0eface7724a8b81cfff96ecd2d62ee08f0496163f545eabbf4af789c: Status 404 returned error can't find the container with id 320b3e8c0eface7724a8b81cfff96ecd2d62ee08f0496163f545eabbf4af789c Feb 27 08:07:44 crc kubenswrapper[4612]: I0227 08:07:44.332920 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-7smgg"] Feb 27 08:07:44 crc kubenswrapper[4612]: I0227 08:07:44.340174 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-7smgg"] Feb 27 08:07:44 crc kubenswrapper[4612]: I0227 08:07:44.377776 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-579pm"] Feb 27 08:07:44 crc kubenswrapper[4612]: I0227 08:07:44.384070 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-579pm"] Feb 27 08:07:44 crc kubenswrapper[4612]: I0227 08:07:44.571215 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-x255x"] Feb 27 08:07:44 crc kubenswrapper[4612]: W0227 08:07:44.581568 4612 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod20c3f793_4286_4067_a372_37b2de08fa8e.slice/crio-8fe6520d1b34fd52b077f7a2e19bc45fd37d9ac2dbb79676b126294e0170ddc5 WatchSource:0}: Error finding container 8fe6520d1b34fd52b077f7a2e19bc45fd37d9ac2dbb79676b126294e0170ddc5: Status 404 returned error can't find the container with id 8fe6520d1b34fd52b077f7a2e19bc45fd37d9ac2dbb79676b126294e0170ddc5 Feb 27 08:07:44 crc kubenswrapper[4612]: I0227 08:07:44.864249 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="93fcf290-888e-43d5-b46b-0a235776bdbc" path="/var/lib/kubelet/pods/93fcf290-888e-43d5-b46b-0a235776bdbc/volumes" Feb 27 08:07:44 crc kubenswrapper[4612]: I0227 08:07:44.865170 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e2653a32-b431-40b0-b9de-6a431d7f0644" path="/var/lib/kubelet/pods/e2653a32-b431-40b0-b9de-6a431d7f0644/volumes" Feb 27 08:07:45 crc kubenswrapper[4612]: I0227 08:07:45.250751 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-x255x" event={"ID":"20c3f793-4286-4067-a372-37b2de08fa8e","Type":"ContainerStarted","Data":"8fe6520d1b34fd52b077f7a2e19bc45fd37d9ac2dbb79676b126294e0170ddc5"} Feb 27 08:07:45 crc kubenswrapper[4612]: I0227 08:07:45.252215 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"1336636f-bc7a-45d4-a2ee-886a6f64ecd3","Type":"ContainerStarted","Data":"320b3e8c0eface7724a8b81cfff96ecd2d62ee08f0496163f545eabbf4af789c"} Feb 27 08:07:46 crc kubenswrapper[4612]: I0227 08:07:46.027013 4612 patch_prober.go:28] interesting pod/machine-config-daemon-924vb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 27 08:07:46 crc kubenswrapper[4612]: I0227 08:07:46.027098 4612 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 27 08:07:50 crc kubenswrapper[4612]: I0227 08:07:50.701077 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Feb 27 08:07:51 crc kubenswrapper[4612]: I0227 08:07:51.302898 4612 generic.go:334] "Generic (PLEG): container finished" podID="f189e394-0c97-48a4-9c55-4bd6710923f1" containerID="a05a735130fbad58220807c9fcbc842cec9ed97bd45b1d0fd4b4dc1f3f5aa50f" exitCode=0 Feb 27 08:07:51 crc kubenswrapper[4612]: I0227 08:07:51.302931 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"f189e394-0c97-48a4-9c55-4bd6710923f1","Type":"ContainerDied","Data":"a05a735130fbad58220807c9fcbc842cec9ed97bd45b1d0fd4b4dc1f3f5aa50f"} Feb 27 08:07:53 crc kubenswrapper[4612]: I0227 08:07:53.317645 4612 generic.go:334] "Generic (PLEG): container finished" podID="20c3f793-4286-4067-a372-37b2de08fa8e" containerID="8d83584881cecd2c70e46cf3b38246e180449f484e407f79a151dfcca133f53c" exitCode=0 Feb 27 08:07:53 crc kubenswrapper[4612]: I0227 08:07:53.317788 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-x255x" event={"ID":"20c3f793-4286-4067-a372-37b2de08fa8e","Type":"ContainerDied","Data":"8d83584881cecd2c70e46cf3b38246e180449f484e407f79a151dfcca133f53c"} Feb 27 08:07:53 crc kubenswrapper[4612]: I0227 08:07:53.323620 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-z7f9r" event={"ID":"1ef66a51-46e7-4b76-a70b-b475d47822fd","Type":"ContainerStarted","Data":"73ba5bd40ae9480c29e0910c319b4f5276de02cd309a0ca3f81ba26bb88669eb"} Feb 27 08:07:53 crc kubenswrapper[4612]: I0227 08:07:53.323813 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-z7f9r" Feb 27 08:07:53 crc kubenswrapper[4612]: I0227 08:07:53.329065 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"ddac95c9-8d81-420b-836e-d1902ce8beb5","Type":"ContainerStarted","Data":"d2733378f8761e7f53bf0f576ed820ab568f551d64099c55074f71f505aefe86"} Feb 27 08:07:53 crc kubenswrapper[4612]: I0227 08:07:53.337275 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"f189e394-0c97-48a4-9c55-4bd6710923f1","Type":"ContainerStarted","Data":"b2cde4c738407d67958e2c488f33845e48ef4e6956455f07caad04ff3ac48145"} Feb 27 08:07:53 crc kubenswrapper[4612]: I0227 08:07:53.347195 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"1336636f-bc7a-45d4-a2ee-886a6f64ecd3","Type":"ContainerStarted","Data":"6084ea6830b6c13bca8ead2b33fdd20fc6d4b05378f30687db14d8f900d30629"} Feb 27 08:07:53 crc kubenswrapper[4612]: I0227 08:07:53.380678 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=13.618158847 podStartE2EDuration="39.380662365s" podCreationTimestamp="2026-02-27 08:07:14 +0000 UTC" firstStartedPulling="2026-02-27 08:07:16.922592126 +0000 UTC m=+1094.776522114" lastFinishedPulling="2026-02-27 08:07:42.685095634 +0000 UTC m=+1120.539025632" observedRunningTime="2026-02-27 08:07:53.375320662 +0000 UTC m=+1131.229250660" watchObservedRunningTime="2026-02-27 08:07:53.380662365 +0000 UTC m=+1131.234592363" Feb 27 08:07:53 crc kubenswrapper[4612]: I0227 08:07:53.398018 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-z7f9r" podStartSLOduration=21.794423325 podStartE2EDuration="30.397999421s" podCreationTimestamp="2026-02-27 08:07:23 +0000 UTC" firstStartedPulling="2026-02-27 08:07:43.909271322 +0000 UTC m=+1121.763201320" lastFinishedPulling="2026-02-27 08:07:52.512847408 +0000 UTC m=+1130.366777416" observedRunningTime="2026-02-27 08:07:53.397351923 +0000 UTC m=+1131.251281921" watchObservedRunningTime="2026-02-27 08:07:53.397999421 +0000 UTC m=+1131.251929419" Feb 27 08:07:54 crc kubenswrapper[4612]: I0227 08:07:54.360764 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-x255x" event={"ID":"20c3f793-4286-4067-a372-37b2de08fa8e","Type":"ContainerStarted","Data":"504a779fbc3d440ccd00edb90508cadac48fe42a3d1a1008ff814215f5621dae"} Feb 27 08:07:54 crc kubenswrapper[4612]: I0227 08:07:54.361654 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-x255x" event={"ID":"20c3f793-4286-4067-a372-37b2de08fa8e","Type":"ContainerStarted","Data":"5310a79855f87ef168f0349f34ef2ba1ce85c423efe6bd6a9057a04a57e6df87"} Feb 27 08:07:54 crc kubenswrapper[4612]: I0227 08:07:54.362876 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-x255x" Feb 27 08:07:54 crc kubenswrapper[4612]: I0227 08:07:54.362924 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-x255x" Feb 27 08:07:54 crc kubenswrapper[4612]: I0227 08:07:54.873479 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-x255x" podStartSLOduration=23.970854739 podStartE2EDuration="31.87346269s" podCreationTimestamp="2026-02-27 08:07:23 +0000 UTC" firstStartedPulling="2026-02-27 08:07:44.590827011 +0000 UTC m=+1122.444757009" lastFinishedPulling="2026-02-27 08:07:52.493434962 +0000 UTC m=+1130.347364960" observedRunningTime="2026-02-27 08:07:54.387139284 +0000 UTC m=+1132.241069282" watchObservedRunningTime="2026-02-27 08:07:54.87346269 +0000 UTC m=+1132.727392688" Feb 27 08:07:55 crc kubenswrapper[4612]: I0227 08:07:55.370424 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"1336636f-bc7a-45d4-a2ee-886a6f64ecd3","Type":"ContainerStarted","Data":"d424049131770203825a1326265e8ccf9af60370ad9e26eb0a90f3088d156fd6"} Feb 27 08:07:55 crc kubenswrapper[4612]: I0227 08:07:55.374411 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"ddac95c9-8d81-420b-836e-d1902ce8beb5","Type":"ContainerStarted","Data":"ffc3499b1cb9246e073b90aee322ec39ead1eb7d1873402a59cd068293df9b95"} Feb 27 08:07:55 crc kubenswrapper[4612]: I0227 08:07:55.409372 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=22.002747525 podStartE2EDuration="32.409352505s" podCreationTimestamp="2026-02-27 08:07:23 +0000 UTC" firstStartedPulling="2026-02-27 08:07:44.320614839 +0000 UTC m=+1122.174544837" lastFinishedPulling="2026-02-27 08:07:54.727219809 +0000 UTC m=+1132.581149817" observedRunningTime="2026-02-27 08:07:55.405141945 +0000 UTC m=+1133.259071953" watchObservedRunningTime="2026-02-27 08:07:55.409352505 +0000 UTC m=+1133.263282503" Feb 27 08:07:55 crc kubenswrapper[4612]: I0227 08:07:55.441596 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=19.621195512 podStartE2EDuration="30.441573339s" podCreationTimestamp="2026-02-27 08:07:25 +0000 UTC" firstStartedPulling="2026-02-27 08:07:43.900506261 +0000 UTC m=+1121.754436259" lastFinishedPulling="2026-02-27 08:07:54.720884088 +0000 UTC m=+1132.574814086" observedRunningTime="2026-02-27 08:07:55.440723684 +0000 UTC m=+1133.294653702" watchObservedRunningTime="2026-02-27 08:07:55.441573339 +0000 UTC m=+1133.295503337" Feb 27 08:07:56 crc kubenswrapper[4612]: I0227 08:07:56.289951 4612 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Feb 27 08:07:56 crc kubenswrapper[4612]: I0227 08:07:56.290269 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Feb 27 08:07:56 crc kubenswrapper[4612]: I0227 08:07:56.383505 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"d19bfe90-c5ca-4089-b4f0-bf3eccff30c8","Type":"ContainerStarted","Data":"e7efab0dcc8d7c30fa22e0e71f87da7815d8fd555a9cd154064544d4ad537c0b"} Feb 27 08:07:56 crc kubenswrapper[4612]: I0227 08:07:56.384321 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Feb 27 08:07:56 crc kubenswrapper[4612]: I0227 08:07:56.887605 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=3.461234572 podStartE2EDuration="39.887583562s" podCreationTimestamp="2026-02-27 08:07:17 +0000 UTC" firstStartedPulling="2026-02-27 08:07:18.83658023 +0000 UTC m=+1096.690510228" lastFinishedPulling="2026-02-27 08:07:55.26292922 +0000 UTC m=+1133.116859218" observedRunningTime="2026-02-27 08:07:56.407072303 +0000 UTC m=+1134.261002301" watchObservedRunningTime="2026-02-27 08:07:56.887583562 +0000 UTC m=+1134.741513560" Feb 27 08:07:57 crc kubenswrapper[4612]: I0227 08:07:57.083917 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Feb 27 08:07:57 crc kubenswrapper[4612]: I0227 08:07:57.083975 4612 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Feb 27 08:07:57 crc kubenswrapper[4612]: I0227 08:07:57.122721 4612 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Feb 27 08:07:57 crc kubenswrapper[4612]: I0227 08:07:57.393940 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"e02a40f2-3de1-4f01-84b0-2fe56acdd396","Type":"ContainerStarted","Data":"2c2dd24bb0984e2f0667b21aefd75dd2cdbda5fdb59e4abd8219a1cf61d6b59c"} Feb 27 08:07:57 crc kubenswrapper[4612]: I0227 08:07:57.398615 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"0da436b1-803d-41bb-94ab-9cfd6fa769eb","Type":"ContainerStarted","Data":"839c0ba56d6f1ff42f16427d18510198097a735f0b3fdf5b535bc49979e9110b"} Feb 27 08:07:57 crc kubenswrapper[4612]: I0227 08:07:57.402903 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"6d1ba6db-1cae-461e-9716-5d5b36350218","Type":"ContainerStarted","Data":"eb1bb5464e5938959380380e0adb9e0444116a9729409966f1310218a5de0a2a"} Feb 27 08:07:57 crc kubenswrapper[4612]: I0227 08:07:57.499357 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Feb 27 08:07:57 crc kubenswrapper[4612]: I0227 08:07:57.812190 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-bz4dn"] Feb 27 08:07:57 crc kubenswrapper[4612]: I0227 08:07:57.877736 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-m6w89"] Feb 27 08:07:57 crc kubenswrapper[4612]: I0227 08:07:57.878582 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-m6w89" Feb 27 08:07:57 crc kubenswrapper[4612]: I0227 08:07:57.917542 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Feb 27 08:07:57 crc kubenswrapper[4612]: I0227 08:07:57.935735 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/99cc2aa0-2437-49fe-a7e0-1df45cecd195-config\") pod \"ovn-controller-metrics-m6w89\" (UID: \"99cc2aa0-2437-49fe-a7e0-1df45cecd195\") " pod="openstack/ovn-controller-metrics-m6w89" Feb 27 08:07:57 crc kubenswrapper[4612]: I0227 08:07:57.936165 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/99cc2aa0-2437-49fe-a7e0-1df45cecd195-ovs-rundir\") pod \"ovn-controller-metrics-m6w89\" (UID: \"99cc2aa0-2437-49fe-a7e0-1df45cecd195\") " pod="openstack/ovn-controller-metrics-m6w89" Feb 27 08:07:57 crc kubenswrapper[4612]: I0227 08:07:57.936278 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/99cc2aa0-2437-49fe-a7e0-1df45cecd195-ovn-rundir\") pod \"ovn-controller-metrics-m6w89\" (UID: \"99cc2aa0-2437-49fe-a7e0-1df45cecd195\") " pod="openstack/ovn-controller-metrics-m6w89" Feb 27 08:07:57 crc kubenswrapper[4612]: I0227 08:07:57.936500 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/99cc2aa0-2437-49fe-a7e0-1df45cecd195-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-m6w89\" (UID: \"99cc2aa0-2437-49fe-a7e0-1df45cecd195\") " pod="openstack/ovn-controller-metrics-m6w89" Feb 27 08:07:57 crc kubenswrapper[4612]: I0227 08:07:57.936616 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99cc2aa0-2437-49fe-a7e0-1df45cecd195-combined-ca-bundle\") pod \"ovn-controller-metrics-m6w89\" (UID: \"99cc2aa0-2437-49fe-a7e0-1df45cecd195\") " pod="openstack/ovn-controller-metrics-m6w89" Feb 27 08:07:57 crc kubenswrapper[4612]: I0227 08:07:57.936674 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xm76g\" (UniqueName: \"kubernetes.io/projected/99cc2aa0-2437-49fe-a7e0-1df45cecd195-kube-api-access-xm76g\") pod \"ovn-controller-metrics-m6w89\" (UID: \"99cc2aa0-2437-49fe-a7e0-1df45cecd195\") " pod="openstack/ovn-controller-metrics-m6w89" Feb 27 08:07:57 crc kubenswrapper[4612]: I0227 08:07:57.936989 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6bc7876d45-k2v9x"] Feb 27 08:07:57 crc kubenswrapper[4612]: I0227 08:07:57.939392 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bc7876d45-k2v9x" Feb 27 08:07:57 crc kubenswrapper[4612]: I0227 08:07:57.973094 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Feb 27 08:07:57 crc kubenswrapper[4612]: I0227 08:07:57.997533 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-m6w89"] Feb 27 08:07:58 crc kubenswrapper[4612]: I0227 08:07:58.011369 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6bc7876d45-k2v9x"] Feb 27 08:07:58 crc kubenswrapper[4612]: I0227 08:07:58.038207 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/99cc2aa0-2437-49fe-a7e0-1df45cecd195-config\") pod \"ovn-controller-metrics-m6w89\" (UID: \"99cc2aa0-2437-49fe-a7e0-1df45cecd195\") " pod="openstack/ovn-controller-metrics-m6w89" Feb 27 08:07:58 crc kubenswrapper[4612]: I0227 08:07:58.038269 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/99cc2aa0-2437-49fe-a7e0-1df45cecd195-ovs-rundir\") pod \"ovn-controller-metrics-m6w89\" (UID: \"99cc2aa0-2437-49fe-a7e0-1df45cecd195\") " pod="openstack/ovn-controller-metrics-m6w89" Feb 27 08:07:58 crc kubenswrapper[4612]: I0227 08:07:58.038303 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/15aa4339-acbf-4d66-b8d8-ca4ef14bad44-ovsdbserver-sb\") pod \"dnsmasq-dns-6bc7876d45-k2v9x\" (UID: \"15aa4339-acbf-4d66-b8d8-ca4ef14bad44\") " pod="openstack/dnsmasq-dns-6bc7876d45-k2v9x" Feb 27 08:07:58 crc kubenswrapper[4612]: I0227 08:07:58.038331 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c7tbx\" (UniqueName: \"kubernetes.io/projected/15aa4339-acbf-4d66-b8d8-ca4ef14bad44-kube-api-access-c7tbx\") pod \"dnsmasq-dns-6bc7876d45-k2v9x\" (UID: \"15aa4339-acbf-4d66-b8d8-ca4ef14bad44\") " pod="openstack/dnsmasq-dns-6bc7876d45-k2v9x" Feb 27 08:07:58 crc kubenswrapper[4612]: I0227 08:07:58.038360 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/99cc2aa0-2437-49fe-a7e0-1df45cecd195-ovn-rundir\") pod \"ovn-controller-metrics-m6w89\" (UID: \"99cc2aa0-2437-49fe-a7e0-1df45cecd195\") " pod="openstack/ovn-controller-metrics-m6w89" Feb 27 08:07:58 crc kubenswrapper[4612]: I0227 08:07:58.038388 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/15aa4339-acbf-4d66-b8d8-ca4ef14bad44-dns-svc\") pod \"dnsmasq-dns-6bc7876d45-k2v9x\" (UID: \"15aa4339-acbf-4d66-b8d8-ca4ef14bad44\") " pod="openstack/dnsmasq-dns-6bc7876d45-k2v9x" Feb 27 08:07:58 crc kubenswrapper[4612]: I0227 08:07:58.038419 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/15aa4339-acbf-4d66-b8d8-ca4ef14bad44-config\") pod \"dnsmasq-dns-6bc7876d45-k2v9x\" (UID: \"15aa4339-acbf-4d66-b8d8-ca4ef14bad44\") " pod="openstack/dnsmasq-dns-6bc7876d45-k2v9x" Feb 27 08:07:58 crc kubenswrapper[4612]: I0227 08:07:58.038458 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/99cc2aa0-2437-49fe-a7e0-1df45cecd195-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-m6w89\" (UID: \"99cc2aa0-2437-49fe-a7e0-1df45cecd195\") " pod="openstack/ovn-controller-metrics-m6w89" Feb 27 08:07:58 crc kubenswrapper[4612]: I0227 08:07:58.038484 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99cc2aa0-2437-49fe-a7e0-1df45cecd195-combined-ca-bundle\") pod \"ovn-controller-metrics-m6w89\" (UID: \"99cc2aa0-2437-49fe-a7e0-1df45cecd195\") " pod="openstack/ovn-controller-metrics-m6w89" Feb 27 08:07:58 crc kubenswrapper[4612]: I0227 08:07:58.038500 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xm76g\" (UniqueName: \"kubernetes.io/projected/99cc2aa0-2437-49fe-a7e0-1df45cecd195-kube-api-access-xm76g\") pod \"ovn-controller-metrics-m6w89\" (UID: \"99cc2aa0-2437-49fe-a7e0-1df45cecd195\") " pod="openstack/ovn-controller-metrics-m6w89" Feb 27 08:07:58 crc kubenswrapper[4612]: I0227 08:07:58.038968 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/99cc2aa0-2437-49fe-a7e0-1df45cecd195-ovn-rundir\") pod \"ovn-controller-metrics-m6w89\" (UID: \"99cc2aa0-2437-49fe-a7e0-1df45cecd195\") " pod="openstack/ovn-controller-metrics-m6w89" Feb 27 08:07:58 crc kubenswrapper[4612]: I0227 08:07:58.039257 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/99cc2aa0-2437-49fe-a7e0-1df45cecd195-ovs-rundir\") pod \"ovn-controller-metrics-m6w89\" (UID: \"99cc2aa0-2437-49fe-a7e0-1df45cecd195\") " pod="openstack/ovn-controller-metrics-m6w89" Feb 27 08:07:58 crc kubenswrapper[4612]: I0227 08:07:58.039562 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/99cc2aa0-2437-49fe-a7e0-1df45cecd195-config\") pod \"ovn-controller-metrics-m6w89\" (UID: \"99cc2aa0-2437-49fe-a7e0-1df45cecd195\") " pod="openstack/ovn-controller-metrics-m6w89" Feb 27 08:07:58 crc kubenswrapper[4612]: I0227 08:07:58.061272 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99cc2aa0-2437-49fe-a7e0-1df45cecd195-combined-ca-bundle\") pod \"ovn-controller-metrics-m6w89\" (UID: \"99cc2aa0-2437-49fe-a7e0-1df45cecd195\") " pod="openstack/ovn-controller-metrics-m6w89" Feb 27 08:07:58 crc kubenswrapper[4612]: I0227 08:07:58.069222 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/99cc2aa0-2437-49fe-a7e0-1df45cecd195-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-m6w89\" (UID: \"99cc2aa0-2437-49fe-a7e0-1df45cecd195\") " pod="openstack/ovn-controller-metrics-m6w89" Feb 27 08:07:58 crc kubenswrapper[4612]: I0227 08:07:58.079324 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xm76g\" (UniqueName: \"kubernetes.io/projected/99cc2aa0-2437-49fe-a7e0-1df45cecd195-kube-api-access-xm76g\") pod \"ovn-controller-metrics-m6w89\" (UID: \"99cc2aa0-2437-49fe-a7e0-1df45cecd195\") " pod="openstack/ovn-controller-metrics-m6w89" Feb 27 08:07:58 crc kubenswrapper[4612]: I0227 08:07:58.140325 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/15aa4339-acbf-4d66-b8d8-ca4ef14bad44-ovsdbserver-sb\") pod \"dnsmasq-dns-6bc7876d45-k2v9x\" (UID: \"15aa4339-acbf-4d66-b8d8-ca4ef14bad44\") " pod="openstack/dnsmasq-dns-6bc7876d45-k2v9x" Feb 27 08:07:58 crc kubenswrapper[4612]: I0227 08:07:58.140374 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c7tbx\" (UniqueName: \"kubernetes.io/projected/15aa4339-acbf-4d66-b8d8-ca4ef14bad44-kube-api-access-c7tbx\") pod \"dnsmasq-dns-6bc7876d45-k2v9x\" (UID: \"15aa4339-acbf-4d66-b8d8-ca4ef14bad44\") " pod="openstack/dnsmasq-dns-6bc7876d45-k2v9x" Feb 27 08:07:58 crc kubenswrapper[4612]: I0227 08:07:58.140412 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/15aa4339-acbf-4d66-b8d8-ca4ef14bad44-dns-svc\") pod \"dnsmasq-dns-6bc7876d45-k2v9x\" (UID: \"15aa4339-acbf-4d66-b8d8-ca4ef14bad44\") " pod="openstack/dnsmasq-dns-6bc7876d45-k2v9x" Feb 27 08:07:58 crc kubenswrapper[4612]: I0227 08:07:58.140437 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/15aa4339-acbf-4d66-b8d8-ca4ef14bad44-config\") pod \"dnsmasq-dns-6bc7876d45-k2v9x\" (UID: \"15aa4339-acbf-4d66-b8d8-ca4ef14bad44\") " pod="openstack/dnsmasq-dns-6bc7876d45-k2v9x" Feb 27 08:07:58 crc kubenswrapper[4612]: I0227 08:07:58.141242 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/15aa4339-acbf-4d66-b8d8-ca4ef14bad44-config\") pod \"dnsmasq-dns-6bc7876d45-k2v9x\" (UID: \"15aa4339-acbf-4d66-b8d8-ca4ef14bad44\") " pod="openstack/dnsmasq-dns-6bc7876d45-k2v9x" Feb 27 08:07:58 crc kubenswrapper[4612]: I0227 08:07:58.141536 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/15aa4339-acbf-4d66-b8d8-ca4ef14bad44-ovsdbserver-sb\") pod \"dnsmasq-dns-6bc7876d45-k2v9x\" (UID: \"15aa4339-acbf-4d66-b8d8-ca4ef14bad44\") " pod="openstack/dnsmasq-dns-6bc7876d45-k2v9x" Feb 27 08:07:58 crc kubenswrapper[4612]: I0227 08:07:58.147426 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/15aa4339-acbf-4d66-b8d8-ca4ef14bad44-dns-svc\") pod \"dnsmasq-dns-6bc7876d45-k2v9x\" (UID: \"15aa4339-acbf-4d66-b8d8-ca4ef14bad44\") " pod="openstack/dnsmasq-dns-6bc7876d45-k2v9x" Feb 27 08:07:58 crc kubenswrapper[4612]: I0227 08:07:58.185025 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c7tbx\" (UniqueName: \"kubernetes.io/projected/15aa4339-acbf-4d66-b8d8-ca4ef14bad44-kube-api-access-c7tbx\") pod \"dnsmasq-dns-6bc7876d45-k2v9x\" (UID: \"15aa4339-acbf-4d66-b8d8-ca4ef14bad44\") " pod="openstack/dnsmasq-dns-6bc7876d45-k2v9x" Feb 27 08:07:58 crc kubenswrapper[4612]: I0227 08:07:58.217237 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-jghh4"] Feb 27 08:07:58 crc kubenswrapper[4612]: I0227 08:07:58.248394 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-m6w89" Feb 27 08:07:58 crc kubenswrapper[4612]: I0227 08:07:58.271577 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bc7876d45-k2v9x" Feb 27 08:07:58 crc kubenswrapper[4612]: I0227 08:07:58.279760 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-8554648995-md72g"] Feb 27 08:07:58 crc kubenswrapper[4612]: I0227 08:07:58.281082 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8554648995-md72g" Feb 27 08:07:58 crc kubenswrapper[4612]: I0227 08:07:58.282311 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-bz4dn" Feb 27 08:07:58 crc kubenswrapper[4612]: I0227 08:07:58.283886 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Feb 27 08:07:58 crc kubenswrapper[4612]: I0227 08:07:58.294010 4612 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Feb 27 08:07:58 crc kubenswrapper[4612]: I0227 08:07:58.294843 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8554648995-md72g"] Feb 27 08:07:58 crc kubenswrapper[4612]: I0227 08:07:58.344677 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/72e3b749-ce9a-4480-8873-d8c9b6495cbc-dns-svc\") pod \"72e3b749-ce9a-4480-8873-d8c9b6495cbc\" (UID: \"72e3b749-ce9a-4480-8873-d8c9b6495cbc\") " Feb 27 08:07:58 crc kubenswrapper[4612]: I0227 08:07:58.344750 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/72e3b749-ce9a-4480-8873-d8c9b6495cbc-config\") pod \"72e3b749-ce9a-4480-8873-d8c9b6495cbc\" (UID: \"72e3b749-ce9a-4480-8873-d8c9b6495cbc\") " Feb 27 08:07:58 crc kubenswrapper[4612]: I0227 08:07:58.344790 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rh2tp\" (UniqueName: \"kubernetes.io/projected/72e3b749-ce9a-4480-8873-d8c9b6495cbc-kube-api-access-rh2tp\") pod \"72e3b749-ce9a-4480-8873-d8c9b6495cbc\" (UID: \"72e3b749-ce9a-4480-8873-d8c9b6495cbc\") " Feb 27 08:07:58 crc kubenswrapper[4612]: I0227 08:07:58.344924 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p7s4d\" (UniqueName: \"kubernetes.io/projected/8fc4951a-451c-4194-8776-5e18a3a82479-kube-api-access-p7s4d\") pod \"dnsmasq-dns-8554648995-md72g\" (UID: \"8fc4951a-451c-4194-8776-5e18a3a82479\") " pod="openstack/dnsmasq-dns-8554648995-md72g" Feb 27 08:07:58 crc kubenswrapper[4612]: I0227 08:07:58.344949 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8fc4951a-451c-4194-8776-5e18a3a82479-config\") pod \"dnsmasq-dns-8554648995-md72g\" (UID: \"8fc4951a-451c-4194-8776-5e18a3a82479\") " pod="openstack/dnsmasq-dns-8554648995-md72g" Feb 27 08:07:58 crc kubenswrapper[4612]: I0227 08:07:58.344979 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8fc4951a-451c-4194-8776-5e18a3a82479-dns-svc\") pod \"dnsmasq-dns-8554648995-md72g\" (UID: \"8fc4951a-451c-4194-8776-5e18a3a82479\") " pod="openstack/dnsmasq-dns-8554648995-md72g" Feb 27 08:07:58 crc kubenswrapper[4612]: I0227 08:07:58.345051 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8fc4951a-451c-4194-8776-5e18a3a82479-ovsdbserver-sb\") pod \"dnsmasq-dns-8554648995-md72g\" (UID: \"8fc4951a-451c-4194-8776-5e18a3a82479\") " pod="openstack/dnsmasq-dns-8554648995-md72g" Feb 27 08:07:58 crc kubenswrapper[4612]: I0227 08:07:58.345113 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8fc4951a-451c-4194-8776-5e18a3a82479-ovsdbserver-nb\") pod \"dnsmasq-dns-8554648995-md72g\" (UID: \"8fc4951a-451c-4194-8776-5e18a3a82479\") " pod="openstack/dnsmasq-dns-8554648995-md72g" Feb 27 08:07:58 crc kubenswrapper[4612]: I0227 08:07:58.346167 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/72e3b749-ce9a-4480-8873-d8c9b6495cbc-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "72e3b749-ce9a-4480-8873-d8c9b6495cbc" (UID: "72e3b749-ce9a-4480-8873-d8c9b6495cbc"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 08:07:58 crc kubenswrapper[4612]: I0227 08:07:58.346487 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/72e3b749-ce9a-4480-8873-d8c9b6495cbc-config" (OuterVolumeSpecName: "config") pod "72e3b749-ce9a-4480-8873-d8c9b6495cbc" (UID: "72e3b749-ce9a-4480-8873-d8c9b6495cbc"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 08:07:58 crc kubenswrapper[4612]: I0227 08:07:58.352815 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/72e3b749-ce9a-4480-8873-d8c9b6495cbc-kube-api-access-rh2tp" (OuterVolumeSpecName: "kube-api-access-rh2tp") pod "72e3b749-ce9a-4480-8873-d8c9b6495cbc" (UID: "72e3b749-ce9a-4480-8873-d8c9b6495cbc"). InnerVolumeSpecName "kube-api-access-rh2tp". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 08:07:58 crc kubenswrapper[4612]: I0227 08:07:58.370872 4612 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Feb 27 08:07:58 crc kubenswrapper[4612]: I0227 08:07:58.419187 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-bz4dn" event={"ID":"72e3b749-ce9a-4480-8873-d8c9b6495cbc","Type":"ContainerDied","Data":"152840ba98fe6de0f95953025ad658d0f6483f322820a14332d945558142f743"} Feb 27 08:07:58 crc kubenswrapper[4612]: I0227 08:07:58.419267 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-bz4dn" Feb 27 08:07:58 crc kubenswrapper[4612]: I0227 08:07:58.422953 4612 generic.go:334] "Generic (PLEG): container finished" podID="57036ec4-7696-4e30-b365-44176b0afdc2" containerID="0ace83ec8a3697a7f7d99a2f20428c464aec78a1b8231bb664cf12519160932d" exitCode=0 Feb 27 08:07:58 crc kubenswrapper[4612]: I0227 08:07:58.423457 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-jghh4" event={"ID":"57036ec4-7696-4e30-b365-44176b0afdc2","Type":"ContainerDied","Data":"0ace83ec8a3697a7f7d99a2f20428c464aec78a1b8231bb664cf12519160932d"} Feb 27 08:07:58 crc kubenswrapper[4612]: I0227 08:07:58.423780 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Feb 27 08:07:58 crc kubenswrapper[4612]: I0227 08:07:58.446235 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8fc4951a-451c-4194-8776-5e18a3a82479-ovsdbserver-nb\") pod \"dnsmasq-dns-8554648995-md72g\" (UID: \"8fc4951a-451c-4194-8776-5e18a3a82479\") " pod="openstack/dnsmasq-dns-8554648995-md72g" Feb 27 08:07:58 crc kubenswrapper[4612]: I0227 08:07:58.446295 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p7s4d\" (UniqueName: \"kubernetes.io/projected/8fc4951a-451c-4194-8776-5e18a3a82479-kube-api-access-p7s4d\") pod \"dnsmasq-dns-8554648995-md72g\" (UID: \"8fc4951a-451c-4194-8776-5e18a3a82479\") " pod="openstack/dnsmasq-dns-8554648995-md72g" Feb 27 08:07:58 crc kubenswrapper[4612]: I0227 08:07:58.446322 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8fc4951a-451c-4194-8776-5e18a3a82479-config\") pod \"dnsmasq-dns-8554648995-md72g\" (UID: \"8fc4951a-451c-4194-8776-5e18a3a82479\") " pod="openstack/dnsmasq-dns-8554648995-md72g" Feb 27 08:07:58 crc kubenswrapper[4612]: I0227 08:07:58.446352 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8fc4951a-451c-4194-8776-5e18a3a82479-dns-svc\") pod \"dnsmasq-dns-8554648995-md72g\" (UID: \"8fc4951a-451c-4194-8776-5e18a3a82479\") " pod="openstack/dnsmasq-dns-8554648995-md72g" Feb 27 08:07:58 crc kubenswrapper[4612]: I0227 08:07:58.446401 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8fc4951a-451c-4194-8776-5e18a3a82479-ovsdbserver-sb\") pod \"dnsmasq-dns-8554648995-md72g\" (UID: \"8fc4951a-451c-4194-8776-5e18a3a82479\") " pod="openstack/dnsmasq-dns-8554648995-md72g" Feb 27 08:07:58 crc kubenswrapper[4612]: I0227 08:07:58.446460 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rh2tp\" (UniqueName: \"kubernetes.io/projected/72e3b749-ce9a-4480-8873-d8c9b6495cbc-kube-api-access-rh2tp\") on node \"crc\" DevicePath \"\"" Feb 27 08:07:58 crc kubenswrapper[4612]: I0227 08:07:58.446472 4612 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/72e3b749-ce9a-4480-8873-d8c9b6495cbc-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 27 08:07:58 crc kubenswrapper[4612]: I0227 08:07:58.446482 4612 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/72e3b749-ce9a-4480-8873-d8c9b6495cbc-config\") on node \"crc\" DevicePath \"\"" Feb 27 08:07:58 crc kubenswrapper[4612]: I0227 08:07:58.447344 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8fc4951a-451c-4194-8776-5e18a3a82479-ovsdbserver-sb\") pod \"dnsmasq-dns-8554648995-md72g\" (UID: \"8fc4951a-451c-4194-8776-5e18a3a82479\") " pod="openstack/dnsmasq-dns-8554648995-md72g" Feb 27 08:07:58 crc kubenswrapper[4612]: I0227 08:07:58.447885 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8fc4951a-451c-4194-8776-5e18a3a82479-dns-svc\") pod \"dnsmasq-dns-8554648995-md72g\" (UID: \"8fc4951a-451c-4194-8776-5e18a3a82479\") " pod="openstack/dnsmasq-dns-8554648995-md72g" Feb 27 08:07:58 crc kubenswrapper[4612]: I0227 08:07:58.450541 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8fc4951a-451c-4194-8776-5e18a3a82479-config\") pod \"dnsmasq-dns-8554648995-md72g\" (UID: \"8fc4951a-451c-4194-8776-5e18a3a82479\") " pod="openstack/dnsmasq-dns-8554648995-md72g" Feb 27 08:07:58 crc kubenswrapper[4612]: I0227 08:07:58.451061 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8fc4951a-451c-4194-8776-5e18a3a82479-ovsdbserver-nb\") pod \"dnsmasq-dns-8554648995-md72g\" (UID: \"8fc4951a-451c-4194-8776-5e18a3a82479\") " pod="openstack/dnsmasq-dns-8554648995-md72g" Feb 27 08:07:58 crc kubenswrapper[4612]: I0227 08:07:58.477873 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p7s4d\" (UniqueName: \"kubernetes.io/projected/8fc4951a-451c-4194-8776-5e18a3a82479-kube-api-access-p7s4d\") pod \"dnsmasq-dns-8554648995-md72g\" (UID: \"8fc4951a-451c-4194-8776-5e18a3a82479\") " pod="openstack/dnsmasq-dns-8554648995-md72g" Feb 27 08:07:58 crc kubenswrapper[4612]: I0227 08:07:58.513303 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-bz4dn"] Feb 27 08:07:58 crc kubenswrapper[4612]: I0227 08:07:58.517843 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Feb 27 08:07:58 crc kubenswrapper[4612]: I0227 08:07:58.524195 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-bz4dn"] Feb 27 08:07:58 crc kubenswrapper[4612]: I0227 08:07:58.627067 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8554648995-md72g" Feb 27 08:07:58 crc kubenswrapper[4612]: I0227 08:07:58.761920 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-m6w89"] Feb 27 08:07:59 crc kubenswrapper[4612]: I0227 08:07:59.010525 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="72e3b749-ce9a-4480-8873-d8c9b6495cbc" path="/var/lib/kubelet/pods/72e3b749-ce9a-4480-8873-d8c9b6495cbc/volumes" Feb 27 08:07:59 crc kubenswrapper[4612]: I0227 08:07:59.011173 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Feb 27 08:07:59 crc kubenswrapper[4612]: I0227 08:07:59.012435 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Feb 27 08:07:59 crc kubenswrapper[4612]: I0227 08:07:59.012512 4612 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Feb 27 08:07:59 crc kubenswrapper[4612]: I0227 08:07:59.012637 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Feb 27 08:07:59 crc kubenswrapper[4612]: I0227 08:07:59.013462 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6bc7876d45-k2v9x"] Feb 27 08:07:59 crc kubenswrapper[4612]: I0227 08:07:59.019481 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Feb 27 08:07:59 crc kubenswrapper[4612]: I0227 08:07:59.019784 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Feb 27 08:07:59 crc kubenswrapper[4612]: I0227 08:07:59.019915 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Feb 27 08:07:59 crc kubenswrapper[4612]: I0227 08:07:59.020021 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-jvl8x" Feb 27 08:07:59 crc kubenswrapper[4612]: I0227 08:07:59.145191 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Feb 27 08:07:59 crc kubenswrapper[4612]: I0227 08:07:59.175884 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/44e7bca8-b600-4ae0-8205-28f110f82dae-scripts\") pod \"ovn-northd-0\" (UID: \"44e7bca8-b600-4ae0-8205-28f110f82dae\") " pod="openstack/ovn-northd-0" Feb 27 08:07:59 crc kubenswrapper[4612]: I0227 08:07:59.175954 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/44e7bca8-b600-4ae0-8205-28f110f82dae-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"44e7bca8-b600-4ae0-8205-28f110f82dae\") " pod="openstack/ovn-northd-0" Feb 27 08:07:59 crc kubenswrapper[4612]: I0227 08:07:59.175975 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/44e7bca8-b600-4ae0-8205-28f110f82dae-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"44e7bca8-b600-4ae0-8205-28f110f82dae\") " pod="openstack/ovn-northd-0" Feb 27 08:07:59 crc kubenswrapper[4612]: I0227 08:07:59.176009 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/44e7bca8-b600-4ae0-8205-28f110f82dae-config\") pod \"ovn-northd-0\" (UID: \"44e7bca8-b600-4ae0-8205-28f110f82dae\") " pod="openstack/ovn-northd-0" Feb 27 08:07:59 crc kubenswrapper[4612]: I0227 08:07:59.176024 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h42zt\" (UniqueName: \"kubernetes.io/projected/44e7bca8-b600-4ae0-8205-28f110f82dae-kube-api-access-h42zt\") pod \"ovn-northd-0\" (UID: \"44e7bca8-b600-4ae0-8205-28f110f82dae\") " pod="openstack/ovn-northd-0" Feb 27 08:07:59 crc kubenswrapper[4612]: I0227 08:07:59.176071 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/44e7bca8-b600-4ae0-8205-28f110f82dae-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"44e7bca8-b600-4ae0-8205-28f110f82dae\") " pod="openstack/ovn-northd-0" Feb 27 08:07:59 crc kubenswrapper[4612]: I0227 08:07:59.176109 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/44e7bca8-b600-4ae0-8205-28f110f82dae-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"44e7bca8-b600-4ae0-8205-28f110f82dae\") " pod="openstack/ovn-northd-0" Feb 27 08:07:59 crc kubenswrapper[4612]: I0227 08:07:59.252274 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-jghh4" Feb 27 08:07:59 crc kubenswrapper[4612]: I0227 08:07:59.280027 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/44e7bca8-b600-4ae0-8205-28f110f82dae-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"44e7bca8-b600-4ae0-8205-28f110f82dae\") " pod="openstack/ovn-northd-0" Feb 27 08:07:59 crc kubenswrapper[4612]: I0227 08:07:59.280892 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/44e7bca8-b600-4ae0-8205-28f110f82dae-scripts\") pod \"ovn-northd-0\" (UID: \"44e7bca8-b600-4ae0-8205-28f110f82dae\") " pod="openstack/ovn-northd-0" Feb 27 08:07:59 crc kubenswrapper[4612]: I0227 08:07:59.280984 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/44e7bca8-b600-4ae0-8205-28f110f82dae-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"44e7bca8-b600-4ae0-8205-28f110f82dae\") " pod="openstack/ovn-northd-0" Feb 27 08:07:59 crc kubenswrapper[4612]: I0227 08:07:59.281022 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/44e7bca8-b600-4ae0-8205-28f110f82dae-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"44e7bca8-b600-4ae0-8205-28f110f82dae\") " pod="openstack/ovn-northd-0" Feb 27 08:07:59 crc kubenswrapper[4612]: I0227 08:07:59.281082 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/44e7bca8-b600-4ae0-8205-28f110f82dae-config\") pod \"ovn-northd-0\" (UID: \"44e7bca8-b600-4ae0-8205-28f110f82dae\") " pod="openstack/ovn-northd-0" Feb 27 08:07:59 crc kubenswrapper[4612]: I0227 08:07:59.281111 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h42zt\" (UniqueName: \"kubernetes.io/projected/44e7bca8-b600-4ae0-8205-28f110f82dae-kube-api-access-h42zt\") pod \"ovn-northd-0\" (UID: \"44e7bca8-b600-4ae0-8205-28f110f82dae\") " pod="openstack/ovn-northd-0" Feb 27 08:07:59 crc kubenswrapper[4612]: I0227 08:07:59.281169 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/44e7bca8-b600-4ae0-8205-28f110f82dae-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"44e7bca8-b600-4ae0-8205-28f110f82dae\") " pod="openstack/ovn-northd-0" Feb 27 08:07:59 crc kubenswrapper[4612]: I0227 08:07:59.282010 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/44e7bca8-b600-4ae0-8205-28f110f82dae-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"44e7bca8-b600-4ae0-8205-28f110f82dae\") " pod="openstack/ovn-northd-0" Feb 27 08:07:59 crc kubenswrapper[4612]: I0227 08:07:59.283814 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/44e7bca8-b600-4ae0-8205-28f110f82dae-config\") pod \"ovn-northd-0\" (UID: \"44e7bca8-b600-4ae0-8205-28f110f82dae\") " pod="openstack/ovn-northd-0" Feb 27 08:07:59 crc kubenswrapper[4612]: I0227 08:07:59.291129 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/44e7bca8-b600-4ae0-8205-28f110f82dae-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"44e7bca8-b600-4ae0-8205-28f110f82dae\") " pod="openstack/ovn-northd-0" Feb 27 08:07:59 crc kubenswrapper[4612]: I0227 08:07:59.291786 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/44e7bca8-b600-4ae0-8205-28f110f82dae-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"44e7bca8-b600-4ae0-8205-28f110f82dae\") " pod="openstack/ovn-northd-0" Feb 27 08:07:59 crc kubenswrapper[4612]: I0227 08:07:59.292942 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/44e7bca8-b600-4ae0-8205-28f110f82dae-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"44e7bca8-b600-4ae0-8205-28f110f82dae\") " pod="openstack/ovn-northd-0" Feb 27 08:07:59 crc kubenswrapper[4612]: I0227 08:07:59.298720 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/44e7bca8-b600-4ae0-8205-28f110f82dae-scripts\") pod \"ovn-northd-0\" (UID: \"44e7bca8-b600-4ae0-8205-28f110f82dae\") " pod="openstack/ovn-northd-0" Feb 27 08:07:59 crc kubenswrapper[4612]: I0227 08:07:59.313150 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h42zt\" (UniqueName: \"kubernetes.io/projected/44e7bca8-b600-4ae0-8205-28f110f82dae-kube-api-access-h42zt\") pod \"ovn-northd-0\" (UID: \"44e7bca8-b600-4ae0-8205-28f110f82dae\") " pod="openstack/ovn-northd-0" Feb 27 08:07:59 crc kubenswrapper[4612]: I0227 08:07:59.347620 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8554648995-md72g"] Feb 27 08:07:59 crc kubenswrapper[4612]: I0227 08:07:59.382139 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jssxq\" (UniqueName: \"kubernetes.io/projected/57036ec4-7696-4e30-b365-44176b0afdc2-kube-api-access-jssxq\") pod \"57036ec4-7696-4e30-b365-44176b0afdc2\" (UID: \"57036ec4-7696-4e30-b365-44176b0afdc2\") " Feb 27 08:07:59 crc kubenswrapper[4612]: I0227 08:07:59.382182 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/57036ec4-7696-4e30-b365-44176b0afdc2-config\") pod \"57036ec4-7696-4e30-b365-44176b0afdc2\" (UID: \"57036ec4-7696-4e30-b365-44176b0afdc2\") " Feb 27 08:07:59 crc kubenswrapper[4612]: I0227 08:07:59.382202 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/57036ec4-7696-4e30-b365-44176b0afdc2-dns-svc\") pod \"57036ec4-7696-4e30-b365-44176b0afdc2\" (UID: \"57036ec4-7696-4e30-b365-44176b0afdc2\") " Feb 27 08:07:59 crc kubenswrapper[4612]: I0227 08:07:59.387350 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57036ec4-7696-4e30-b365-44176b0afdc2-kube-api-access-jssxq" (OuterVolumeSpecName: "kube-api-access-jssxq") pod "57036ec4-7696-4e30-b365-44176b0afdc2" (UID: "57036ec4-7696-4e30-b365-44176b0afdc2"). InnerVolumeSpecName "kube-api-access-jssxq". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 08:07:59 crc kubenswrapper[4612]: I0227 08:07:59.411342 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/57036ec4-7696-4e30-b365-44176b0afdc2-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "57036ec4-7696-4e30-b365-44176b0afdc2" (UID: "57036ec4-7696-4e30-b365-44176b0afdc2"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 08:07:59 crc kubenswrapper[4612]: I0227 08:07:59.412084 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/57036ec4-7696-4e30-b365-44176b0afdc2-config" (OuterVolumeSpecName: "config") pod "57036ec4-7696-4e30-b365-44176b0afdc2" (UID: "57036ec4-7696-4e30-b365-44176b0afdc2"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 08:07:59 crc kubenswrapper[4612]: I0227 08:07:59.430173 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-jghh4" event={"ID":"57036ec4-7696-4e30-b365-44176b0afdc2","Type":"ContainerDied","Data":"08028231b4e92b3303f61c2ac8b27f9df6050f3da4933accfb44811d6e216955"} Feb 27 08:07:59 crc kubenswrapper[4612]: I0227 08:07:59.430200 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-jghh4" Feb 27 08:07:59 crc kubenswrapper[4612]: I0227 08:07:59.430216 4612 scope.go:117] "RemoveContainer" containerID="0ace83ec8a3697a7f7d99a2f20428c464aec78a1b8231bb664cf12519160932d" Feb 27 08:07:59 crc kubenswrapper[4612]: I0227 08:07:59.432299 4612 generic.go:334] "Generic (PLEG): container finished" podID="15aa4339-acbf-4d66-b8d8-ca4ef14bad44" containerID="7cd099f6ddd6f47437ac0ab4a9df73420ed2890fe4fa713c2ee35e8a52b44d47" exitCode=0 Feb 27 08:07:59 crc kubenswrapper[4612]: I0227 08:07:59.432362 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bc7876d45-k2v9x" event={"ID":"15aa4339-acbf-4d66-b8d8-ca4ef14bad44","Type":"ContainerDied","Data":"7cd099f6ddd6f47437ac0ab4a9df73420ed2890fe4fa713c2ee35e8a52b44d47"} Feb 27 08:07:59 crc kubenswrapper[4612]: I0227 08:07:59.432377 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bc7876d45-k2v9x" event={"ID":"15aa4339-acbf-4d66-b8d8-ca4ef14bad44","Type":"ContainerStarted","Data":"d56d40a9f9e3aac6a8261f759cc6ecd7837bf6cbc9d1bd97f7fee3146dec8b72"} Feb 27 08:07:59 crc kubenswrapper[4612]: I0227 08:07:59.439384 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-m6w89" event={"ID":"99cc2aa0-2437-49fe-a7e0-1df45cecd195","Type":"ContainerStarted","Data":"a37ea832d81fb162868c54cc76647cc7e977b0412c40ea7b2f34475dd1ee06c2"} Feb 27 08:07:59 crc kubenswrapper[4612]: I0227 08:07:59.439432 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-m6w89" event={"ID":"99cc2aa0-2437-49fe-a7e0-1df45cecd195","Type":"ContainerStarted","Data":"be8416f4e44ec41dc313cc975fd12e4214287a24bf76d900631e6e3828788f91"} Feb 27 08:07:59 crc kubenswrapper[4612]: I0227 08:07:59.444809 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-md72g" event={"ID":"8fc4951a-451c-4194-8776-5e18a3a82479","Type":"ContainerStarted","Data":"68d164eaa6c2c77917d82649888e5d6e88e9fdad783de0455d9e4c61961c8e35"} Feb 27 08:07:59 crc kubenswrapper[4612]: I0227 08:07:59.484384 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jssxq\" (UniqueName: \"kubernetes.io/projected/57036ec4-7696-4e30-b365-44176b0afdc2-kube-api-access-jssxq\") on node \"crc\" DevicePath \"\"" Feb 27 08:07:59 crc kubenswrapper[4612]: I0227 08:07:59.484423 4612 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/57036ec4-7696-4e30-b365-44176b0afdc2-config\") on node \"crc\" DevicePath \"\"" Feb 27 08:07:59 crc kubenswrapper[4612]: I0227 08:07:59.484436 4612 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/57036ec4-7696-4e30-b365-44176b0afdc2-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 27 08:07:59 crc kubenswrapper[4612]: I0227 08:07:59.498546 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-m6w89" podStartSLOduration=2.498524105 podStartE2EDuration="2.498524105s" podCreationTimestamp="2026-02-27 08:07:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 08:07:59.488756355 +0000 UTC m=+1137.342686353" watchObservedRunningTime="2026-02-27 08:07:59.498524105 +0000 UTC m=+1137.352454103" Feb 27 08:07:59 crc kubenswrapper[4612]: I0227 08:07:59.527475 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Feb 27 08:07:59 crc kubenswrapper[4612]: I0227 08:07:59.548938 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-jghh4"] Feb 27 08:07:59 crc kubenswrapper[4612]: I0227 08:07:59.575879 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-jghh4"] Feb 27 08:08:00 crc kubenswrapper[4612]: I0227 08:08:00.045084 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Feb 27 08:08:00 crc kubenswrapper[4612]: I0227 08:08:00.133973 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29536328-gbk7h"] Feb 27 08:08:00 crc kubenswrapper[4612]: E0227 08:08:00.134644 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="57036ec4-7696-4e30-b365-44176b0afdc2" containerName="init" Feb 27 08:08:00 crc kubenswrapper[4612]: I0227 08:08:00.134664 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="57036ec4-7696-4e30-b365-44176b0afdc2" containerName="init" Feb 27 08:08:00 crc kubenswrapper[4612]: I0227 08:08:00.134887 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="57036ec4-7696-4e30-b365-44176b0afdc2" containerName="init" Feb 27 08:08:00 crc kubenswrapper[4612]: I0227 08:08:00.135574 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536328-gbk7h" Feb 27 08:08:00 crc kubenswrapper[4612]: I0227 08:08:00.141561 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 27 08:08:00 crc kubenswrapper[4612]: I0227 08:08:00.141767 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-5zwfx" Feb 27 08:08:00 crc kubenswrapper[4612]: I0227 08:08:00.142290 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 27 08:08:00 crc kubenswrapper[4612]: I0227 08:08:00.170672 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536328-gbk7h"] Feb 27 08:08:00 crc kubenswrapper[4612]: I0227 08:08:00.196066 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qvndq\" (UniqueName: \"kubernetes.io/projected/ff8b1f18-8c91-41d1-b157-ea9c37c2a4ff-kube-api-access-qvndq\") pod \"auto-csr-approver-29536328-gbk7h\" (UID: \"ff8b1f18-8c91-41d1-b157-ea9c37c2a4ff\") " pod="openshift-infra/auto-csr-approver-29536328-gbk7h" Feb 27 08:08:00 crc kubenswrapper[4612]: I0227 08:08:00.298065 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qvndq\" (UniqueName: \"kubernetes.io/projected/ff8b1f18-8c91-41d1-b157-ea9c37c2a4ff-kube-api-access-qvndq\") pod \"auto-csr-approver-29536328-gbk7h\" (UID: \"ff8b1f18-8c91-41d1-b157-ea9c37c2a4ff\") " pod="openshift-infra/auto-csr-approver-29536328-gbk7h" Feb 27 08:08:00 crc kubenswrapper[4612]: I0227 08:08:00.319638 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qvndq\" (UniqueName: \"kubernetes.io/projected/ff8b1f18-8c91-41d1-b157-ea9c37c2a4ff-kube-api-access-qvndq\") pod \"auto-csr-approver-29536328-gbk7h\" (UID: \"ff8b1f18-8c91-41d1-b157-ea9c37c2a4ff\") " pod="openshift-infra/auto-csr-approver-29536328-gbk7h" Feb 27 08:08:00 crc kubenswrapper[4612]: I0227 08:08:00.451854 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536328-gbk7h" Feb 27 08:08:00 crc kubenswrapper[4612]: I0227 08:08:00.452798 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"44e7bca8-b600-4ae0-8205-28f110f82dae","Type":"ContainerStarted","Data":"950d299a84b509cd295291871f95575af8afba3fdceb904c5b95e1c76193bae6"} Feb 27 08:08:00 crc kubenswrapper[4612]: I0227 08:08:00.454887 4612 generic.go:334] "Generic (PLEG): container finished" podID="8fc4951a-451c-4194-8776-5e18a3a82479" containerID="6520b230948a858cee9959aa97849aff4bac6864f6ea3c48969b03829394403c" exitCode=0 Feb 27 08:08:00 crc kubenswrapper[4612]: I0227 08:08:00.454919 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-md72g" event={"ID":"8fc4951a-451c-4194-8776-5e18a3a82479","Type":"ContainerDied","Data":"6520b230948a858cee9959aa97849aff4bac6864f6ea3c48969b03829394403c"} Feb 27 08:08:00 crc kubenswrapper[4612]: I0227 08:08:00.458300 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bc7876d45-k2v9x" event={"ID":"15aa4339-acbf-4d66-b8d8-ca4ef14bad44","Type":"ContainerStarted","Data":"06474ec45acf96a16a692d4e6bd2ac7bac092af130f80483e085703beac8dec7"} Feb 27 08:08:00 crc kubenswrapper[4612]: I0227 08:08:00.459597 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6bc7876d45-k2v9x" Feb 27 08:08:00 crc kubenswrapper[4612]: I0227 08:08:00.522220 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6bc7876d45-k2v9x" podStartSLOduration=3.522197958 podStartE2EDuration="3.522197958s" podCreationTimestamp="2026-02-27 08:07:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 08:08:00.511335967 +0000 UTC m=+1138.365265975" watchObservedRunningTime="2026-02-27 08:08:00.522197958 +0000 UTC m=+1138.376127956" Feb 27 08:08:00 crc kubenswrapper[4612]: I0227 08:08:00.880571 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57036ec4-7696-4e30-b365-44176b0afdc2" path="/var/lib/kubelet/pods/57036ec4-7696-4e30-b365-44176b0afdc2/volumes" Feb 27 08:08:01 crc kubenswrapper[4612]: I0227 08:08:01.024899 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536328-gbk7h"] Feb 27 08:08:01 crc kubenswrapper[4612]: I0227 08:08:01.520968 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536328-gbk7h" event={"ID":"ff8b1f18-8c91-41d1-b157-ea9c37c2a4ff","Type":"ContainerStarted","Data":"0603bc53ffd80e7489233536c6103eabd4f7b8085adeee872d9fd0e61997f8d6"} Feb 27 08:08:01 crc kubenswrapper[4612]: I0227 08:08:01.550316 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"44e7bca8-b600-4ae0-8205-28f110f82dae","Type":"ContainerStarted","Data":"71557db6ecab44d2d8f6a057b0b5d7911e5bde98243e23e910f60b50d047b497"} Feb 27 08:08:01 crc kubenswrapper[4612]: I0227 08:08:01.569073 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-md72g" event={"ID":"8fc4951a-451c-4194-8776-5e18a3a82479","Type":"ContainerStarted","Data":"026e77480afb4661e5bdd221bfb4cb9ff8c2fd02e88b87d3293ec2748322f273"} Feb 27 08:08:01 crc kubenswrapper[4612]: I0227 08:08:01.569479 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-8554648995-md72g" Feb 27 08:08:01 crc kubenswrapper[4612]: I0227 08:08:01.600644 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-8554648995-md72g" podStartSLOduration=3.600625629 podStartE2EDuration="3.600625629s" podCreationTimestamp="2026-02-27 08:07:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 08:08:01.595591805 +0000 UTC m=+1139.449521803" watchObservedRunningTime="2026-02-27 08:08:01.600625629 +0000 UTC m=+1139.454555627" Feb 27 08:08:02 crc kubenswrapper[4612]: I0227 08:08:02.576368 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"44e7bca8-b600-4ae0-8205-28f110f82dae","Type":"ContainerStarted","Data":"a6456c92b0a8856d1f5a7a12c7086ce6922e84d9dcb833cfd3a2ed9673409cda"} Feb 27 08:08:02 crc kubenswrapper[4612]: I0227 08:08:02.576716 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Feb 27 08:08:02 crc kubenswrapper[4612]: I0227 08:08:02.578030 4612 generic.go:334] "Generic (PLEG): container finished" podID="e02a40f2-3de1-4f01-84b0-2fe56acdd396" containerID="2c2dd24bb0984e2f0667b21aefd75dd2cdbda5fdb59e4abd8219a1cf61d6b59c" exitCode=0 Feb 27 08:08:02 crc kubenswrapper[4612]: I0227 08:08:02.578164 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"e02a40f2-3de1-4f01-84b0-2fe56acdd396","Type":"ContainerDied","Data":"2c2dd24bb0984e2f0667b21aefd75dd2cdbda5fdb59e4abd8219a1cf61d6b59c"} Feb 27 08:08:02 crc kubenswrapper[4612]: I0227 08:08:02.581208 4612 generic.go:334] "Generic (PLEG): container finished" podID="ff8b1f18-8c91-41d1-b157-ea9c37c2a4ff" containerID="0992a9fec848e35a9a3164334231e441c0a8eb1353824fed5c37ccddfc2eed38" exitCode=0 Feb 27 08:08:02 crc kubenswrapper[4612]: I0227 08:08:02.581254 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536328-gbk7h" event={"ID":"ff8b1f18-8c91-41d1-b157-ea9c37c2a4ff","Type":"ContainerDied","Data":"0992a9fec848e35a9a3164334231e441c0a8eb1353824fed5c37ccddfc2eed38"} Feb 27 08:08:02 crc kubenswrapper[4612]: I0227 08:08:02.608185 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=3.4739845 podStartE2EDuration="4.608164709s" podCreationTimestamp="2026-02-27 08:07:58 +0000 UTC" firstStartedPulling="2026-02-27 08:08:00.049239856 +0000 UTC m=+1137.903169854" lastFinishedPulling="2026-02-27 08:08:01.183420065 +0000 UTC m=+1139.037350063" observedRunningTime="2026-02-27 08:08:02.596874456 +0000 UTC m=+1140.450804454" watchObservedRunningTime="2026-02-27 08:08:02.608164709 +0000 UTC m=+1140.462094717" Feb 27 08:08:03 crc kubenswrapper[4612]: I0227 08:08:03.012819 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Feb 27 08:08:03 crc kubenswrapper[4612]: I0227 08:08:03.307559 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-a904-account-create-update-hsnsq"] Feb 27 08:08:03 crc kubenswrapper[4612]: I0227 08:08:03.308800 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-a904-account-create-update-hsnsq" Feb 27 08:08:03 crc kubenswrapper[4612]: I0227 08:08:03.316252 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Feb 27 08:08:03 crc kubenswrapper[4612]: I0227 08:08:03.323556 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-a904-account-create-update-hsnsq"] Feb 27 08:08:03 crc kubenswrapper[4612]: I0227 08:08:03.342161 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-t9l25"] Feb 27 08:08:03 crc kubenswrapper[4612]: I0227 08:08:03.343237 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-t9l25" Feb 27 08:08:03 crc kubenswrapper[4612]: I0227 08:08:03.362950 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vph9f\" (UniqueName: \"kubernetes.io/projected/395f27e6-3047-48b3-a63f-7931bd177ccb-kube-api-access-vph9f\") pod \"glance-a904-account-create-update-hsnsq\" (UID: \"395f27e6-3047-48b3-a63f-7931bd177ccb\") " pod="openstack/glance-a904-account-create-update-hsnsq" Feb 27 08:08:03 crc kubenswrapper[4612]: I0227 08:08:03.363022 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/395f27e6-3047-48b3-a63f-7931bd177ccb-operator-scripts\") pod \"glance-a904-account-create-update-hsnsq\" (UID: \"395f27e6-3047-48b3-a63f-7931bd177ccb\") " pod="openstack/glance-a904-account-create-update-hsnsq" Feb 27 08:08:03 crc kubenswrapper[4612]: I0227 08:08:03.365214 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-t9l25"] Feb 27 08:08:03 crc kubenswrapper[4612]: I0227 08:08:03.475130 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7754x\" (UniqueName: \"kubernetes.io/projected/48835d1f-acd8-41db-a37f-7a82a08561bc-kube-api-access-7754x\") pod \"glance-db-create-t9l25\" (UID: \"48835d1f-acd8-41db-a37f-7a82a08561bc\") " pod="openstack/glance-db-create-t9l25" Feb 27 08:08:03 crc kubenswrapper[4612]: I0227 08:08:03.475193 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/395f27e6-3047-48b3-a63f-7931bd177ccb-operator-scripts\") pod \"glance-a904-account-create-update-hsnsq\" (UID: \"395f27e6-3047-48b3-a63f-7931bd177ccb\") " pod="openstack/glance-a904-account-create-update-hsnsq" Feb 27 08:08:03 crc kubenswrapper[4612]: I0227 08:08:03.475417 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/48835d1f-acd8-41db-a37f-7a82a08561bc-operator-scripts\") pod \"glance-db-create-t9l25\" (UID: \"48835d1f-acd8-41db-a37f-7a82a08561bc\") " pod="openstack/glance-db-create-t9l25" Feb 27 08:08:03 crc kubenswrapper[4612]: I0227 08:08:03.475758 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vph9f\" (UniqueName: \"kubernetes.io/projected/395f27e6-3047-48b3-a63f-7931bd177ccb-kube-api-access-vph9f\") pod \"glance-a904-account-create-update-hsnsq\" (UID: \"395f27e6-3047-48b3-a63f-7931bd177ccb\") " pod="openstack/glance-a904-account-create-update-hsnsq" Feb 27 08:08:03 crc kubenswrapper[4612]: I0227 08:08:03.476012 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/395f27e6-3047-48b3-a63f-7931bd177ccb-operator-scripts\") pod \"glance-a904-account-create-update-hsnsq\" (UID: \"395f27e6-3047-48b3-a63f-7931bd177ccb\") " pod="openstack/glance-a904-account-create-update-hsnsq" Feb 27 08:08:03 crc kubenswrapper[4612]: I0227 08:08:03.509251 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vph9f\" (UniqueName: \"kubernetes.io/projected/395f27e6-3047-48b3-a63f-7931bd177ccb-kube-api-access-vph9f\") pod \"glance-a904-account-create-update-hsnsq\" (UID: \"395f27e6-3047-48b3-a63f-7931bd177ccb\") " pod="openstack/glance-a904-account-create-update-hsnsq" Feb 27 08:08:03 crc kubenswrapper[4612]: I0227 08:08:03.580961 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/48835d1f-acd8-41db-a37f-7a82a08561bc-operator-scripts\") pod \"glance-db-create-t9l25\" (UID: \"48835d1f-acd8-41db-a37f-7a82a08561bc\") " pod="openstack/glance-db-create-t9l25" Feb 27 08:08:03 crc kubenswrapper[4612]: I0227 08:08:03.581198 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7754x\" (UniqueName: \"kubernetes.io/projected/48835d1f-acd8-41db-a37f-7a82a08561bc-kube-api-access-7754x\") pod \"glance-db-create-t9l25\" (UID: \"48835d1f-acd8-41db-a37f-7a82a08561bc\") " pod="openstack/glance-db-create-t9l25" Feb 27 08:08:03 crc kubenswrapper[4612]: I0227 08:08:03.582450 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/48835d1f-acd8-41db-a37f-7a82a08561bc-operator-scripts\") pod \"glance-db-create-t9l25\" (UID: \"48835d1f-acd8-41db-a37f-7a82a08561bc\") " pod="openstack/glance-db-create-t9l25" Feb 27 08:08:03 crc kubenswrapper[4612]: I0227 08:08:03.608502 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7754x\" (UniqueName: \"kubernetes.io/projected/48835d1f-acd8-41db-a37f-7a82a08561bc-kube-api-access-7754x\") pod \"glance-db-create-t9l25\" (UID: \"48835d1f-acd8-41db-a37f-7a82a08561bc\") " pod="openstack/glance-db-create-t9l25" Feb 27 08:08:03 crc kubenswrapper[4612]: I0227 08:08:03.651373 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-a904-account-create-update-hsnsq" Feb 27 08:08:03 crc kubenswrapper[4612]: I0227 08:08:03.672155 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-t9l25" Feb 27 08:08:04 crc kubenswrapper[4612]: I0227 08:08:04.006285 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536328-gbk7h" Feb 27 08:08:04 crc kubenswrapper[4612]: I0227 08:08:04.093156 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qvndq\" (UniqueName: \"kubernetes.io/projected/ff8b1f18-8c91-41d1-b157-ea9c37c2a4ff-kube-api-access-qvndq\") pod \"ff8b1f18-8c91-41d1-b157-ea9c37c2a4ff\" (UID: \"ff8b1f18-8c91-41d1-b157-ea9c37c2a4ff\") " Feb 27 08:08:04 crc kubenswrapper[4612]: I0227 08:08:04.099373 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ff8b1f18-8c91-41d1-b157-ea9c37c2a4ff-kube-api-access-qvndq" (OuterVolumeSpecName: "kube-api-access-qvndq") pod "ff8b1f18-8c91-41d1-b157-ea9c37c2a4ff" (UID: "ff8b1f18-8c91-41d1-b157-ea9c37c2a4ff"). InnerVolumeSpecName "kube-api-access-qvndq". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 08:08:04 crc kubenswrapper[4612]: I0227 08:08:04.192895 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-a904-account-create-update-hsnsq"] Feb 27 08:08:04 crc kubenswrapper[4612]: I0227 08:08:04.195079 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qvndq\" (UniqueName: \"kubernetes.io/projected/ff8b1f18-8c91-41d1-b157-ea9c37c2a4ff-kube-api-access-qvndq\") on node \"crc\" DevicePath \"\"" Feb 27 08:08:04 crc kubenswrapper[4612]: I0227 08:08:04.199147 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-t9l25"] Feb 27 08:08:04 crc kubenswrapper[4612]: I0227 08:08:04.613551 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-a904-account-create-update-hsnsq" event={"ID":"395f27e6-3047-48b3-a63f-7931bd177ccb","Type":"ContainerStarted","Data":"35c80ed1f8f3c82bd3aa0b4ec5ee1313a6684ee2f21aa736d4d88e5cb69f16bd"} Feb 27 08:08:04 crc kubenswrapper[4612]: I0227 08:08:04.614515 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-t9l25" event={"ID":"48835d1f-acd8-41db-a37f-7a82a08561bc","Type":"ContainerStarted","Data":"f019225a20ece805697f3cee50c4f4d12330bd6494e6e98b527a6f6484fecc22"} Feb 27 08:08:04 crc kubenswrapper[4612]: I0227 08:08:04.616236 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536328-gbk7h" event={"ID":"ff8b1f18-8c91-41d1-b157-ea9c37c2a4ff","Type":"ContainerDied","Data":"0603bc53ffd80e7489233536c6103eabd4f7b8085adeee872d9fd0e61997f8d6"} Feb 27 08:08:04 crc kubenswrapper[4612]: I0227 08:08:04.616374 4612 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0603bc53ffd80e7489233536c6103eabd4f7b8085adeee872d9fd0e61997f8d6" Feb 27 08:08:04 crc kubenswrapper[4612]: I0227 08:08:04.616272 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536328-gbk7h" Feb 27 08:08:04 crc kubenswrapper[4612]: I0227 08:08:04.931558 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/root-account-create-update-6rx64"] Feb 27 08:08:04 crc kubenswrapper[4612]: E0227 08:08:04.932136 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff8b1f18-8c91-41d1-b157-ea9c37c2a4ff" containerName="oc" Feb 27 08:08:04 crc kubenswrapper[4612]: I0227 08:08:04.932344 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff8b1f18-8c91-41d1-b157-ea9c37c2a4ff" containerName="oc" Feb 27 08:08:04 crc kubenswrapper[4612]: I0227 08:08:04.932503 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="ff8b1f18-8c91-41d1-b157-ea9c37c2a4ff" containerName="oc" Feb 27 08:08:04 crc kubenswrapper[4612]: I0227 08:08:04.932999 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-6rx64" Feb 27 08:08:04 crc kubenswrapper[4612]: I0227 08:08:04.937303 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-mariadb-root-db-secret" Feb 27 08:08:04 crc kubenswrapper[4612]: I0227 08:08:04.947082 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-6rx64"] Feb 27 08:08:05 crc kubenswrapper[4612]: I0227 08:08:05.006420 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a21d121a-8f63-428c-81b9-cc4d1a52be44-operator-scripts\") pod \"root-account-create-update-6rx64\" (UID: \"a21d121a-8f63-428c-81b9-cc4d1a52be44\") " pod="openstack/root-account-create-update-6rx64" Feb 27 08:08:05 crc kubenswrapper[4612]: I0227 08:08:05.006653 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9xhk4\" (UniqueName: \"kubernetes.io/projected/a21d121a-8f63-428c-81b9-cc4d1a52be44-kube-api-access-9xhk4\") pod \"root-account-create-update-6rx64\" (UID: \"a21d121a-8f63-428c-81b9-cc4d1a52be44\") " pod="openstack/root-account-create-update-6rx64" Feb 27 08:08:05 crc kubenswrapper[4612]: I0227 08:08:05.076835 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29536322-w8jkp"] Feb 27 08:08:05 crc kubenswrapper[4612]: I0227 08:08:05.083012 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29536322-w8jkp"] Feb 27 08:08:05 crc kubenswrapper[4612]: I0227 08:08:05.108669 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9xhk4\" (UniqueName: \"kubernetes.io/projected/a21d121a-8f63-428c-81b9-cc4d1a52be44-kube-api-access-9xhk4\") pod \"root-account-create-update-6rx64\" (UID: \"a21d121a-8f63-428c-81b9-cc4d1a52be44\") " pod="openstack/root-account-create-update-6rx64" Feb 27 08:08:05 crc kubenswrapper[4612]: I0227 08:08:05.108905 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a21d121a-8f63-428c-81b9-cc4d1a52be44-operator-scripts\") pod \"root-account-create-update-6rx64\" (UID: \"a21d121a-8f63-428c-81b9-cc4d1a52be44\") " pod="openstack/root-account-create-update-6rx64" Feb 27 08:08:05 crc kubenswrapper[4612]: I0227 08:08:05.109798 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a21d121a-8f63-428c-81b9-cc4d1a52be44-operator-scripts\") pod \"root-account-create-update-6rx64\" (UID: \"a21d121a-8f63-428c-81b9-cc4d1a52be44\") " pod="openstack/root-account-create-update-6rx64" Feb 27 08:08:05 crc kubenswrapper[4612]: I0227 08:08:05.127813 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9xhk4\" (UniqueName: \"kubernetes.io/projected/a21d121a-8f63-428c-81b9-cc4d1a52be44-kube-api-access-9xhk4\") pod \"root-account-create-update-6rx64\" (UID: \"a21d121a-8f63-428c-81b9-cc4d1a52be44\") " pod="openstack/root-account-create-update-6rx64" Feb 27 08:08:05 crc kubenswrapper[4612]: I0227 08:08:05.263513 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-6rx64" Feb 27 08:08:05 crc kubenswrapper[4612]: I0227 08:08:05.624444 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"e02a40f2-3de1-4f01-84b0-2fe56acdd396","Type":"ContainerStarted","Data":"6fa2efe149c8c9d568ebec2cc81f33943d2d9ebd5f8a042943573cf18a77dc8a"} Feb 27 08:08:05 crc kubenswrapper[4612]: I0227 08:08:05.626967 4612 generic.go:334] "Generic (PLEG): container finished" podID="395f27e6-3047-48b3-a63f-7931bd177ccb" containerID="37544462aa650601b257f91f9bb8064b10303d0824e3aecc659444fad5ec504e" exitCode=0 Feb 27 08:08:05 crc kubenswrapper[4612]: I0227 08:08:05.627039 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-a904-account-create-update-hsnsq" event={"ID":"395f27e6-3047-48b3-a63f-7931bd177ccb","Type":"ContainerDied","Data":"37544462aa650601b257f91f9bb8064b10303d0824e3aecc659444fad5ec504e"} Feb 27 08:08:05 crc kubenswrapper[4612]: I0227 08:08:05.635393 4612 generic.go:334] "Generic (PLEG): container finished" podID="48835d1f-acd8-41db-a37f-7a82a08561bc" containerID="1355afc2e630e6ea544c146adeee1caff5bf48d29a747aab4df9750c639e1d92" exitCode=0 Feb 27 08:08:05 crc kubenswrapper[4612]: I0227 08:08:05.635459 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-t9l25" event={"ID":"48835d1f-acd8-41db-a37f-7a82a08561bc","Type":"ContainerDied","Data":"1355afc2e630e6ea544c146adeee1caff5bf48d29a747aab4df9750c639e1d92"} Feb 27 08:08:05 crc kubenswrapper[4612]: I0227 08:08:05.643643 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=-9223371987.211151 podStartE2EDuration="49.643624977s" podCreationTimestamp="2026-02-27 08:07:16 +0000 UTC" firstStartedPulling="2026-02-27 08:07:18.357814802 +0000 UTC m=+1096.211744800" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 08:08:05.642207806 +0000 UTC m=+1143.496137814" watchObservedRunningTime="2026-02-27 08:08:05.643624977 +0000 UTC m=+1143.497554995" Feb 27 08:08:05 crc kubenswrapper[4612]: I0227 08:08:05.741794 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-6rx64"] Feb 27 08:08:05 crc kubenswrapper[4612]: W0227 08:08:05.744241 4612 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda21d121a_8f63_428c_81b9_cc4d1a52be44.slice/crio-a9805d94413bcc5e2053ebec98234a9baab37507e8ae75c7a51128ce8435a19b WatchSource:0}: Error finding container a9805d94413bcc5e2053ebec98234a9baab37507e8ae75c7a51128ce8435a19b: Status 404 returned error can't find the container with id a9805d94413bcc5e2053ebec98234a9baab37507e8ae75c7a51128ce8435a19b Feb 27 08:08:06 crc kubenswrapper[4612]: I0227 08:08:06.641741 4612 generic.go:334] "Generic (PLEG): container finished" podID="a21d121a-8f63-428c-81b9-cc4d1a52be44" containerID="d673d33a8032b5a980f2b7df29867154100118156686660391827c5296f434f8" exitCode=0 Feb 27 08:08:06 crc kubenswrapper[4612]: I0227 08:08:06.642374 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-6rx64" event={"ID":"a21d121a-8f63-428c-81b9-cc4d1a52be44","Type":"ContainerDied","Data":"d673d33a8032b5a980f2b7df29867154100118156686660391827c5296f434f8"} Feb 27 08:08:06 crc kubenswrapper[4612]: I0227 08:08:06.642399 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-6rx64" event={"ID":"a21d121a-8f63-428c-81b9-cc4d1a52be44","Type":"ContainerStarted","Data":"a9805d94413bcc5e2053ebec98234a9baab37507e8ae75c7a51128ce8435a19b"} Feb 27 08:08:06 crc kubenswrapper[4612]: I0227 08:08:06.864444 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2114ee13-c2af-4646-a086-8491370cb05d" path="/var/lib/kubelet/pods/2114ee13-c2af-4646-a086-8491370cb05d/volumes" Feb 27 08:08:07 crc kubenswrapper[4612]: I0227 08:08:07.067490 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-t9l25" Feb 27 08:08:07 crc kubenswrapper[4612]: I0227 08:08:07.072879 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-a904-account-create-update-hsnsq" Feb 27 08:08:07 crc kubenswrapper[4612]: I0227 08:08:07.142042 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/395f27e6-3047-48b3-a63f-7931bd177ccb-operator-scripts\") pod \"395f27e6-3047-48b3-a63f-7931bd177ccb\" (UID: \"395f27e6-3047-48b3-a63f-7931bd177ccb\") " Feb 27 08:08:07 crc kubenswrapper[4612]: I0227 08:08:07.142139 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7754x\" (UniqueName: \"kubernetes.io/projected/48835d1f-acd8-41db-a37f-7a82a08561bc-kube-api-access-7754x\") pod \"48835d1f-acd8-41db-a37f-7a82a08561bc\" (UID: \"48835d1f-acd8-41db-a37f-7a82a08561bc\") " Feb 27 08:08:07 crc kubenswrapper[4612]: I0227 08:08:07.142157 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/48835d1f-acd8-41db-a37f-7a82a08561bc-operator-scripts\") pod \"48835d1f-acd8-41db-a37f-7a82a08561bc\" (UID: \"48835d1f-acd8-41db-a37f-7a82a08561bc\") " Feb 27 08:08:07 crc kubenswrapper[4612]: I0227 08:08:07.142226 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vph9f\" (UniqueName: \"kubernetes.io/projected/395f27e6-3047-48b3-a63f-7931bd177ccb-kube-api-access-vph9f\") pod \"395f27e6-3047-48b3-a63f-7931bd177ccb\" (UID: \"395f27e6-3047-48b3-a63f-7931bd177ccb\") " Feb 27 08:08:07 crc kubenswrapper[4612]: I0227 08:08:07.143338 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/395f27e6-3047-48b3-a63f-7931bd177ccb-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "395f27e6-3047-48b3-a63f-7931bd177ccb" (UID: "395f27e6-3047-48b3-a63f-7931bd177ccb"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 08:08:07 crc kubenswrapper[4612]: I0227 08:08:07.143884 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/48835d1f-acd8-41db-a37f-7a82a08561bc-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "48835d1f-acd8-41db-a37f-7a82a08561bc" (UID: "48835d1f-acd8-41db-a37f-7a82a08561bc"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 08:08:07 crc kubenswrapper[4612]: I0227 08:08:07.149320 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/48835d1f-acd8-41db-a37f-7a82a08561bc-kube-api-access-7754x" (OuterVolumeSpecName: "kube-api-access-7754x") pod "48835d1f-acd8-41db-a37f-7a82a08561bc" (UID: "48835d1f-acd8-41db-a37f-7a82a08561bc"). InnerVolumeSpecName "kube-api-access-7754x". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 08:08:07 crc kubenswrapper[4612]: I0227 08:08:07.149380 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/395f27e6-3047-48b3-a63f-7931bd177ccb-kube-api-access-vph9f" (OuterVolumeSpecName: "kube-api-access-vph9f") pod "395f27e6-3047-48b3-a63f-7931bd177ccb" (UID: "395f27e6-3047-48b3-a63f-7931bd177ccb"). InnerVolumeSpecName "kube-api-access-vph9f". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 08:08:07 crc kubenswrapper[4612]: I0227 08:08:07.244445 4612 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/395f27e6-3047-48b3-a63f-7931bd177ccb-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 27 08:08:07 crc kubenswrapper[4612]: I0227 08:08:07.244782 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7754x\" (UniqueName: \"kubernetes.io/projected/48835d1f-acd8-41db-a37f-7a82a08561bc-kube-api-access-7754x\") on node \"crc\" DevicePath \"\"" Feb 27 08:08:07 crc kubenswrapper[4612]: I0227 08:08:07.244795 4612 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/48835d1f-acd8-41db-a37f-7a82a08561bc-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 27 08:08:07 crc kubenswrapper[4612]: I0227 08:08:07.244803 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vph9f\" (UniqueName: \"kubernetes.io/projected/395f27e6-3047-48b3-a63f-7931bd177ccb-kube-api-access-vph9f\") on node \"crc\" DevicePath \"\"" Feb 27 08:08:07 crc kubenswrapper[4612]: I0227 08:08:07.652930 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-t9l25" Feb 27 08:08:07 crc kubenswrapper[4612]: I0227 08:08:07.654395 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-t9l25" event={"ID":"48835d1f-acd8-41db-a37f-7a82a08561bc","Type":"ContainerDied","Data":"f019225a20ece805697f3cee50c4f4d12330bd6494e6e98b527a6f6484fecc22"} Feb 27 08:08:07 crc kubenswrapper[4612]: I0227 08:08:07.654561 4612 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f019225a20ece805697f3cee50c4f4d12330bd6494e6e98b527a6f6484fecc22" Feb 27 08:08:07 crc kubenswrapper[4612]: I0227 08:08:07.657423 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-a904-account-create-update-hsnsq" Feb 27 08:08:07 crc kubenswrapper[4612]: I0227 08:08:07.657653 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-a904-account-create-update-hsnsq" event={"ID":"395f27e6-3047-48b3-a63f-7931bd177ccb","Type":"ContainerDied","Data":"35c80ed1f8f3c82bd3aa0b4ec5ee1313a6684ee2f21aa736d4d88e5cb69f16bd"} Feb 27 08:08:07 crc kubenswrapper[4612]: I0227 08:08:07.657870 4612 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="35c80ed1f8f3c82bd3aa0b4ec5ee1313a6684ee2f21aa736d4d88e5cb69f16bd" Feb 27 08:08:07 crc kubenswrapper[4612]: I0227 08:08:07.709715 4612 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Feb 27 08:08:07 crc kubenswrapper[4612]: I0227 08:08:07.709781 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Feb 27 08:08:08 crc kubenswrapper[4612]: I0227 08:08:08.024040 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-6rx64" Feb 27 08:08:08 crc kubenswrapper[4612]: I0227 08:08:08.059940 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a21d121a-8f63-428c-81b9-cc4d1a52be44-operator-scripts\") pod \"a21d121a-8f63-428c-81b9-cc4d1a52be44\" (UID: \"a21d121a-8f63-428c-81b9-cc4d1a52be44\") " Feb 27 08:08:08 crc kubenswrapper[4612]: I0227 08:08:08.060084 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xhk4\" (UniqueName: \"kubernetes.io/projected/a21d121a-8f63-428c-81b9-cc4d1a52be44-kube-api-access-9xhk4\") pod \"a21d121a-8f63-428c-81b9-cc4d1a52be44\" (UID: \"a21d121a-8f63-428c-81b9-cc4d1a52be44\") " Feb 27 08:08:08 crc kubenswrapper[4612]: I0227 08:08:08.060557 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a21d121a-8f63-428c-81b9-cc4d1a52be44-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "a21d121a-8f63-428c-81b9-cc4d1a52be44" (UID: "a21d121a-8f63-428c-81b9-cc4d1a52be44"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 08:08:08 crc kubenswrapper[4612]: I0227 08:08:08.066870 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a21d121a-8f63-428c-81b9-cc4d1a52be44-kube-api-access-9xhk4" (OuterVolumeSpecName: "kube-api-access-9xhk4") pod "a21d121a-8f63-428c-81b9-cc4d1a52be44" (UID: "a21d121a-8f63-428c-81b9-cc4d1a52be44"). InnerVolumeSpecName "kube-api-access-9xhk4". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 08:08:08 crc kubenswrapper[4612]: I0227 08:08:08.161978 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xhk4\" (UniqueName: \"kubernetes.io/projected/a21d121a-8f63-428c-81b9-cc4d1a52be44-kube-api-access-9xhk4\") on node \"crc\" DevicePath \"\"" Feb 27 08:08:08 crc kubenswrapper[4612]: I0227 08:08:08.162020 4612 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a21d121a-8f63-428c-81b9-cc4d1a52be44-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 27 08:08:08 crc kubenswrapper[4612]: I0227 08:08:08.273860 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6bc7876d45-k2v9x" Feb 27 08:08:08 crc kubenswrapper[4612]: I0227 08:08:08.629859 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-8554648995-md72g" Feb 27 08:08:08 crc kubenswrapper[4612]: I0227 08:08:08.665869 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-6rx64" event={"ID":"a21d121a-8f63-428c-81b9-cc4d1a52be44","Type":"ContainerDied","Data":"a9805d94413bcc5e2053ebec98234a9baab37507e8ae75c7a51128ce8435a19b"} Feb 27 08:08:08 crc kubenswrapper[4612]: I0227 08:08:08.665905 4612 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a9805d94413bcc5e2053ebec98234a9baab37507e8ae75c7a51128ce8435a19b" Feb 27 08:08:08 crc kubenswrapper[4612]: I0227 08:08:08.665970 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-6rx64" Feb 27 08:08:08 crc kubenswrapper[4612]: I0227 08:08:08.676929 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-4ndtf"] Feb 27 08:08:08 crc kubenswrapper[4612]: E0227 08:08:08.677555 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="48835d1f-acd8-41db-a37f-7a82a08561bc" containerName="mariadb-database-create" Feb 27 08:08:08 crc kubenswrapper[4612]: I0227 08:08:08.677661 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="48835d1f-acd8-41db-a37f-7a82a08561bc" containerName="mariadb-database-create" Feb 27 08:08:08 crc kubenswrapper[4612]: E0227 08:08:08.677804 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="395f27e6-3047-48b3-a63f-7931bd177ccb" containerName="mariadb-account-create-update" Feb 27 08:08:08 crc kubenswrapper[4612]: I0227 08:08:08.677898 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="395f27e6-3047-48b3-a63f-7931bd177ccb" containerName="mariadb-account-create-update" Feb 27 08:08:08 crc kubenswrapper[4612]: E0227 08:08:08.677991 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a21d121a-8f63-428c-81b9-cc4d1a52be44" containerName="mariadb-account-create-update" Feb 27 08:08:08 crc kubenswrapper[4612]: I0227 08:08:08.678073 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="a21d121a-8f63-428c-81b9-cc4d1a52be44" containerName="mariadb-account-create-update" Feb 27 08:08:08 crc kubenswrapper[4612]: I0227 08:08:08.678366 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="a21d121a-8f63-428c-81b9-cc4d1a52be44" containerName="mariadb-account-create-update" Feb 27 08:08:08 crc kubenswrapper[4612]: I0227 08:08:08.678461 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="48835d1f-acd8-41db-a37f-7a82a08561bc" containerName="mariadb-database-create" Feb 27 08:08:08 crc kubenswrapper[4612]: I0227 08:08:08.678548 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="395f27e6-3047-48b3-a63f-7931bd177ccb" containerName="mariadb-account-create-update" Feb 27 08:08:08 crc kubenswrapper[4612]: I0227 08:08:08.679283 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-4ndtf" Feb 27 08:08:08 crc kubenswrapper[4612]: I0227 08:08:08.686648 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-55plp" Feb 27 08:08:08 crc kubenswrapper[4612]: I0227 08:08:08.686933 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Feb 27 08:08:08 crc kubenswrapper[4612]: I0227 08:08:08.702525 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-4ndtf"] Feb 27 08:08:08 crc kubenswrapper[4612]: I0227 08:08:08.739176 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6bc7876d45-k2v9x"] Feb 27 08:08:08 crc kubenswrapper[4612]: I0227 08:08:08.739425 4612 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6bc7876d45-k2v9x" podUID="15aa4339-acbf-4d66-b8d8-ca4ef14bad44" containerName="dnsmasq-dns" containerID="cri-o://06474ec45acf96a16a692d4e6bd2ac7bac092af130f80483e085703beac8dec7" gracePeriod=10 Feb 27 08:08:08 crc kubenswrapper[4612]: I0227 08:08:08.773198 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca30e354-dd13-4fa2-96a9-bbd6c605b1f8-combined-ca-bundle\") pod \"glance-db-sync-4ndtf\" (UID: \"ca30e354-dd13-4fa2-96a9-bbd6c605b1f8\") " pod="openstack/glance-db-sync-4ndtf" Feb 27 08:08:08 crc kubenswrapper[4612]: I0227 08:08:08.773259 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ca30e354-dd13-4fa2-96a9-bbd6c605b1f8-config-data\") pod \"glance-db-sync-4ndtf\" (UID: \"ca30e354-dd13-4fa2-96a9-bbd6c605b1f8\") " pod="openstack/glance-db-sync-4ndtf" Feb 27 08:08:08 crc kubenswrapper[4612]: I0227 08:08:08.773296 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v7fp5\" (UniqueName: \"kubernetes.io/projected/ca30e354-dd13-4fa2-96a9-bbd6c605b1f8-kube-api-access-v7fp5\") pod \"glance-db-sync-4ndtf\" (UID: \"ca30e354-dd13-4fa2-96a9-bbd6c605b1f8\") " pod="openstack/glance-db-sync-4ndtf" Feb 27 08:08:08 crc kubenswrapper[4612]: I0227 08:08:08.773333 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/ca30e354-dd13-4fa2-96a9-bbd6c605b1f8-db-sync-config-data\") pod \"glance-db-sync-4ndtf\" (UID: \"ca30e354-dd13-4fa2-96a9-bbd6c605b1f8\") " pod="openstack/glance-db-sync-4ndtf" Feb 27 08:08:08 crc kubenswrapper[4612]: I0227 08:08:08.877026 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca30e354-dd13-4fa2-96a9-bbd6c605b1f8-combined-ca-bundle\") pod \"glance-db-sync-4ndtf\" (UID: \"ca30e354-dd13-4fa2-96a9-bbd6c605b1f8\") " pod="openstack/glance-db-sync-4ndtf" Feb 27 08:08:08 crc kubenswrapper[4612]: I0227 08:08:08.877119 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ca30e354-dd13-4fa2-96a9-bbd6c605b1f8-config-data\") pod \"glance-db-sync-4ndtf\" (UID: \"ca30e354-dd13-4fa2-96a9-bbd6c605b1f8\") " pod="openstack/glance-db-sync-4ndtf" Feb 27 08:08:08 crc kubenswrapper[4612]: I0227 08:08:08.877178 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v7fp5\" (UniqueName: \"kubernetes.io/projected/ca30e354-dd13-4fa2-96a9-bbd6c605b1f8-kube-api-access-v7fp5\") pod \"glance-db-sync-4ndtf\" (UID: \"ca30e354-dd13-4fa2-96a9-bbd6c605b1f8\") " pod="openstack/glance-db-sync-4ndtf" Feb 27 08:08:08 crc kubenswrapper[4612]: I0227 08:08:08.877237 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/ca30e354-dd13-4fa2-96a9-bbd6c605b1f8-db-sync-config-data\") pod \"glance-db-sync-4ndtf\" (UID: \"ca30e354-dd13-4fa2-96a9-bbd6c605b1f8\") " pod="openstack/glance-db-sync-4ndtf" Feb 27 08:08:08 crc kubenswrapper[4612]: I0227 08:08:08.880945 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/ca30e354-dd13-4fa2-96a9-bbd6c605b1f8-db-sync-config-data\") pod \"glance-db-sync-4ndtf\" (UID: \"ca30e354-dd13-4fa2-96a9-bbd6c605b1f8\") " pod="openstack/glance-db-sync-4ndtf" Feb 27 08:08:08 crc kubenswrapper[4612]: I0227 08:08:08.921952 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ca30e354-dd13-4fa2-96a9-bbd6c605b1f8-config-data\") pod \"glance-db-sync-4ndtf\" (UID: \"ca30e354-dd13-4fa2-96a9-bbd6c605b1f8\") " pod="openstack/glance-db-sync-4ndtf" Feb 27 08:08:08 crc kubenswrapper[4612]: I0227 08:08:08.929292 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v7fp5\" (UniqueName: \"kubernetes.io/projected/ca30e354-dd13-4fa2-96a9-bbd6c605b1f8-kube-api-access-v7fp5\") pod \"glance-db-sync-4ndtf\" (UID: \"ca30e354-dd13-4fa2-96a9-bbd6c605b1f8\") " pod="openstack/glance-db-sync-4ndtf" Feb 27 08:08:08 crc kubenswrapper[4612]: I0227 08:08:08.933440 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca30e354-dd13-4fa2-96a9-bbd6c605b1f8-combined-ca-bundle\") pod \"glance-db-sync-4ndtf\" (UID: \"ca30e354-dd13-4fa2-96a9-bbd6c605b1f8\") " pod="openstack/glance-db-sync-4ndtf" Feb 27 08:08:08 crc kubenswrapper[4612]: I0227 08:08:08.940814 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-lkbpz"] Feb 27 08:08:08 crc kubenswrapper[4612]: I0227 08:08:08.944197 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-lkbpz" Feb 27 08:08:08 crc kubenswrapper[4612]: I0227 08:08:08.954055 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-lkbpz"] Feb 27 08:08:08 crc kubenswrapper[4612]: I0227 08:08:08.986946 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/22a1a3d2-6c5d-4102-8fd4-07f953e402f7-operator-scripts\") pod \"keystone-db-create-lkbpz\" (UID: \"22a1a3d2-6c5d-4102-8fd4-07f953e402f7\") " pod="openstack/keystone-db-create-lkbpz" Feb 27 08:08:08 crc kubenswrapper[4612]: I0227 08:08:08.987067 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5jh9x\" (UniqueName: \"kubernetes.io/projected/22a1a3d2-6c5d-4102-8fd4-07f953e402f7-kube-api-access-5jh9x\") pod \"keystone-db-create-lkbpz\" (UID: \"22a1a3d2-6c5d-4102-8fd4-07f953e402f7\") " pod="openstack/keystone-db-create-lkbpz" Feb 27 08:08:09 crc kubenswrapper[4612]: I0227 08:08:09.021076 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-4ndtf" Feb 27 08:08:09 crc kubenswrapper[4612]: I0227 08:08:09.073589 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-7819-account-create-update-lppdh"] Feb 27 08:08:09 crc kubenswrapper[4612]: I0227 08:08:09.075154 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-7819-account-create-update-lppdh" Feb 27 08:08:09 crc kubenswrapper[4612]: I0227 08:08:09.079762 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Feb 27 08:08:09 crc kubenswrapper[4612]: I0227 08:08:09.080817 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-7819-account-create-update-lppdh"] Feb 27 08:08:09 crc kubenswrapper[4612]: I0227 08:08:09.093474 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/22a1a3d2-6c5d-4102-8fd4-07f953e402f7-operator-scripts\") pod \"keystone-db-create-lkbpz\" (UID: \"22a1a3d2-6c5d-4102-8fd4-07f953e402f7\") " pod="openstack/keystone-db-create-lkbpz" Feb 27 08:08:09 crc kubenswrapper[4612]: I0227 08:08:09.093541 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5jh9x\" (UniqueName: \"kubernetes.io/projected/22a1a3d2-6c5d-4102-8fd4-07f953e402f7-kube-api-access-5jh9x\") pod \"keystone-db-create-lkbpz\" (UID: \"22a1a3d2-6c5d-4102-8fd4-07f953e402f7\") " pod="openstack/keystone-db-create-lkbpz" Feb 27 08:08:09 crc kubenswrapper[4612]: I0227 08:08:09.094753 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/22a1a3d2-6c5d-4102-8fd4-07f953e402f7-operator-scripts\") pod \"keystone-db-create-lkbpz\" (UID: \"22a1a3d2-6c5d-4102-8fd4-07f953e402f7\") " pod="openstack/keystone-db-create-lkbpz" Feb 27 08:08:09 crc kubenswrapper[4612]: I0227 08:08:09.165425 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5jh9x\" (UniqueName: \"kubernetes.io/projected/22a1a3d2-6c5d-4102-8fd4-07f953e402f7-kube-api-access-5jh9x\") pod \"keystone-db-create-lkbpz\" (UID: \"22a1a3d2-6c5d-4102-8fd4-07f953e402f7\") " pod="openstack/keystone-db-create-lkbpz" Feb 27 08:08:09 crc kubenswrapper[4612]: I0227 08:08:09.184733 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-r47j8"] Feb 27 08:08:09 crc kubenswrapper[4612]: I0227 08:08:09.185990 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-r47j8" Feb 27 08:08:09 crc kubenswrapper[4612]: I0227 08:08:09.195848 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/03d154b1-4114-4c56-82b0-5bf148511e5f-operator-scripts\") pod \"keystone-7819-account-create-update-lppdh\" (UID: \"03d154b1-4114-4c56-82b0-5bf148511e5f\") " pod="openstack/keystone-7819-account-create-update-lppdh" Feb 27 08:08:09 crc kubenswrapper[4612]: I0227 08:08:09.196213 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v4mtb\" (UniqueName: \"kubernetes.io/projected/03d154b1-4114-4c56-82b0-5bf148511e5f-kube-api-access-v4mtb\") pod \"keystone-7819-account-create-update-lppdh\" (UID: \"03d154b1-4114-4c56-82b0-5bf148511e5f\") " pod="openstack/keystone-7819-account-create-update-lppdh" Feb 27 08:08:09 crc kubenswrapper[4612]: I0227 08:08:09.196269 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-r47j8"] Feb 27 08:08:09 crc kubenswrapper[4612]: I0227 08:08:09.289248 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-lkbpz" Feb 27 08:08:09 crc kubenswrapper[4612]: I0227 08:08:09.297887 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v4mtb\" (UniqueName: \"kubernetes.io/projected/03d154b1-4114-4c56-82b0-5bf148511e5f-kube-api-access-v4mtb\") pod \"keystone-7819-account-create-update-lppdh\" (UID: \"03d154b1-4114-4c56-82b0-5bf148511e5f\") " pod="openstack/keystone-7819-account-create-update-lppdh" Feb 27 08:08:09 crc kubenswrapper[4612]: I0227 08:08:09.297943 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sp6bh\" (UniqueName: \"kubernetes.io/projected/1551171a-913f-4d99-b4da-148ea454f7f6-kube-api-access-sp6bh\") pod \"placement-db-create-r47j8\" (UID: \"1551171a-913f-4d99-b4da-148ea454f7f6\") " pod="openstack/placement-db-create-r47j8" Feb 27 08:08:09 crc kubenswrapper[4612]: I0227 08:08:09.297984 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/03d154b1-4114-4c56-82b0-5bf148511e5f-operator-scripts\") pod \"keystone-7819-account-create-update-lppdh\" (UID: \"03d154b1-4114-4c56-82b0-5bf148511e5f\") " pod="openstack/keystone-7819-account-create-update-lppdh" Feb 27 08:08:09 crc kubenswrapper[4612]: I0227 08:08:09.298659 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1551171a-913f-4d99-b4da-148ea454f7f6-operator-scripts\") pod \"placement-db-create-r47j8\" (UID: \"1551171a-913f-4d99-b4da-148ea454f7f6\") " pod="openstack/placement-db-create-r47j8" Feb 27 08:08:09 crc kubenswrapper[4612]: I0227 08:08:09.298763 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/03d154b1-4114-4c56-82b0-5bf148511e5f-operator-scripts\") pod \"keystone-7819-account-create-update-lppdh\" (UID: \"03d154b1-4114-4c56-82b0-5bf148511e5f\") " pod="openstack/keystone-7819-account-create-update-lppdh" Feb 27 08:08:09 crc kubenswrapper[4612]: I0227 08:08:09.314369 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v4mtb\" (UniqueName: \"kubernetes.io/projected/03d154b1-4114-4c56-82b0-5bf148511e5f-kube-api-access-v4mtb\") pod \"keystone-7819-account-create-update-lppdh\" (UID: \"03d154b1-4114-4c56-82b0-5bf148511e5f\") " pod="openstack/keystone-7819-account-create-update-lppdh" Feb 27 08:08:09 crc kubenswrapper[4612]: I0227 08:08:09.378451 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-3e89-account-create-update-crnl5"] Feb 27 08:08:09 crc kubenswrapper[4612]: I0227 08:08:09.379518 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-3e89-account-create-update-crnl5" Feb 27 08:08:09 crc kubenswrapper[4612]: I0227 08:08:09.397903 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Feb 27 08:08:09 crc kubenswrapper[4612]: I0227 08:08:09.405814 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1551171a-913f-4d99-b4da-148ea454f7f6-operator-scripts\") pod \"placement-db-create-r47j8\" (UID: \"1551171a-913f-4d99-b4da-148ea454f7f6\") " pod="openstack/placement-db-create-r47j8" Feb 27 08:08:09 crc kubenswrapper[4612]: I0227 08:08:09.405956 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sp6bh\" (UniqueName: \"kubernetes.io/projected/1551171a-913f-4d99-b4da-148ea454f7f6-kube-api-access-sp6bh\") pod \"placement-db-create-r47j8\" (UID: \"1551171a-913f-4d99-b4da-148ea454f7f6\") " pod="openstack/placement-db-create-r47j8" Feb 27 08:08:09 crc kubenswrapper[4612]: I0227 08:08:09.408086 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1551171a-913f-4d99-b4da-148ea454f7f6-operator-scripts\") pod \"placement-db-create-r47j8\" (UID: \"1551171a-913f-4d99-b4da-148ea454f7f6\") " pod="openstack/placement-db-create-r47j8" Feb 27 08:08:09 crc kubenswrapper[4612]: I0227 08:08:09.431406 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-3e89-account-create-update-crnl5"] Feb 27 08:08:09 crc kubenswrapper[4612]: I0227 08:08:09.434541 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-7819-account-create-update-lppdh" Feb 27 08:08:09 crc kubenswrapper[4612]: I0227 08:08:09.443389 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bc7876d45-k2v9x" Feb 27 08:08:09 crc kubenswrapper[4612]: I0227 08:08:09.462096 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sp6bh\" (UniqueName: \"kubernetes.io/projected/1551171a-913f-4d99-b4da-148ea454f7f6-kube-api-access-sp6bh\") pod \"placement-db-create-r47j8\" (UID: \"1551171a-913f-4d99-b4da-148ea454f7f6\") " pod="openstack/placement-db-create-r47j8" Feb 27 08:08:09 crc kubenswrapper[4612]: I0227 08:08:09.507375 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c7tbx\" (UniqueName: \"kubernetes.io/projected/15aa4339-acbf-4d66-b8d8-ca4ef14bad44-kube-api-access-c7tbx\") pod \"15aa4339-acbf-4d66-b8d8-ca4ef14bad44\" (UID: \"15aa4339-acbf-4d66-b8d8-ca4ef14bad44\") " Feb 27 08:08:09 crc kubenswrapper[4612]: I0227 08:08:09.507438 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/15aa4339-acbf-4d66-b8d8-ca4ef14bad44-ovsdbserver-sb\") pod \"15aa4339-acbf-4d66-b8d8-ca4ef14bad44\" (UID: \"15aa4339-acbf-4d66-b8d8-ca4ef14bad44\") " Feb 27 08:08:09 crc kubenswrapper[4612]: I0227 08:08:09.507787 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/15aa4339-acbf-4d66-b8d8-ca4ef14bad44-config\") pod \"15aa4339-acbf-4d66-b8d8-ca4ef14bad44\" (UID: \"15aa4339-acbf-4d66-b8d8-ca4ef14bad44\") " Feb 27 08:08:09 crc kubenswrapper[4612]: I0227 08:08:09.507843 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/15aa4339-acbf-4d66-b8d8-ca4ef14bad44-dns-svc\") pod \"15aa4339-acbf-4d66-b8d8-ca4ef14bad44\" (UID: \"15aa4339-acbf-4d66-b8d8-ca4ef14bad44\") " Feb 27 08:08:09 crc kubenswrapper[4612]: I0227 08:08:09.509923 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jmqz8\" (UniqueName: \"kubernetes.io/projected/6ad4524a-e07e-4a6e-93b1-5e5e2d7b251c-kube-api-access-jmqz8\") pod \"placement-3e89-account-create-update-crnl5\" (UID: \"6ad4524a-e07e-4a6e-93b1-5e5e2d7b251c\") " pod="openstack/placement-3e89-account-create-update-crnl5" Feb 27 08:08:09 crc kubenswrapper[4612]: I0227 08:08:09.510007 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6ad4524a-e07e-4a6e-93b1-5e5e2d7b251c-operator-scripts\") pod \"placement-3e89-account-create-update-crnl5\" (UID: \"6ad4524a-e07e-4a6e-93b1-5e5e2d7b251c\") " pod="openstack/placement-3e89-account-create-update-crnl5" Feb 27 08:08:09 crc kubenswrapper[4612]: I0227 08:08:09.511586 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-r47j8" Feb 27 08:08:09 crc kubenswrapper[4612]: I0227 08:08:09.519757 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/15aa4339-acbf-4d66-b8d8-ca4ef14bad44-kube-api-access-c7tbx" (OuterVolumeSpecName: "kube-api-access-c7tbx") pod "15aa4339-acbf-4d66-b8d8-ca4ef14bad44" (UID: "15aa4339-acbf-4d66-b8d8-ca4ef14bad44"). InnerVolumeSpecName "kube-api-access-c7tbx". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 08:08:09 crc kubenswrapper[4612]: I0227 08:08:09.577668 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/15aa4339-acbf-4d66-b8d8-ca4ef14bad44-config" (OuterVolumeSpecName: "config") pod "15aa4339-acbf-4d66-b8d8-ca4ef14bad44" (UID: "15aa4339-acbf-4d66-b8d8-ca4ef14bad44"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 08:08:09 crc kubenswrapper[4612]: I0227 08:08:09.585662 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/15aa4339-acbf-4d66-b8d8-ca4ef14bad44-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "15aa4339-acbf-4d66-b8d8-ca4ef14bad44" (UID: "15aa4339-acbf-4d66-b8d8-ca4ef14bad44"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 08:08:09 crc kubenswrapper[4612]: I0227 08:08:09.595460 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/15aa4339-acbf-4d66-b8d8-ca4ef14bad44-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "15aa4339-acbf-4d66-b8d8-ca4ef14bad44" (UID: "15aa4339-acbf-4d66-b8d8-ca4ef14bad44"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 08:08:09 crc kubenswrapper[4612]: I0227 08:08:09.611504 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jmqz8\" (UniqueName: \"kubernetes.io/projected/6ad4524a-e07e-4a6e-93b1-5e5e2d7b251c-kube-api-access-jmqz8\") pod \"placement-3e89-account-create-update-crnl5\" (UID: \"6ad4524a-e07e-4a6e-93b1-5e5e2d7b251c\") " pod="openstack/placement-3e89-account-create-update-crnl5" Feb 27 08:08:09 crc kubenswrapper[4612]: I0227 08:08:09.612097 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6ad4524a-e07e-4a6e-93b1-5e5e2d7b251c-operator-scripts\") pod \"placement-3e89-account-create-update-crnl5\" (UID: \"6ad4524a-e07e-4a6e-93b1-5e5e2d7b251c\") " pod="openstack/placement-3e89-account-create-update-crnl5" Feb 27 08:08:09 crc kubenswrapper[4612]: I0227 08:08:09.612250 4612 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/15aa4339-acbf-4d66-b8d8-ca4ef14bad44-config\") on node \"crc\" DevicePath \"\"" Feb 27 08:08:09 crc kubenswrapper[4612]: I0227 08:08:09.612325 4612 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/15aa4339-acbf-4d66-b8d8-ca4ef14bad44-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 27 08:08:09 crc kubenswrapper[4612]: I0227 08:08:09.612384 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c7tbx\" (UniqueName: \"kubernetes.io/projected/15aa4339-acbf-4d66-b8d8-ca4ef14bad44-kube-api-access-c7tbx\") on node \"crc\" DevicePath \"\"" Feb 27 08:08:09 crc kubenswrapper[4612]: I0227 08:08:09.612442 4612 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/15aa4339-acbf-4d66-b8d8-ca4ef14bad44-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 27 08:08:09 crc kubenswrapper[4612]: I0227 08:08:09.612737 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6ad4524a-e07e-4a6e-93b1-5e5e2d7b251c-operator-scripts\") pod \"placement-3e89-account-create-update-crnl5\" (UID: \"6ad4524a-e07e-4a6e-93b1-5e5e2d7b251c\") " pod="openstack/placement-3e89-account-create-update-crnl5" Feb 27 08:08:09 crc kubenswrapper[4612]: I0227 08:08:09.656498 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jmqz8\" (UniqueName: \"kubernetes.io/projected/6ad4524a-e07e-4a6e-93b1-5e5e2d7b251c-kube-api-access-jmqz8\") pod \"placement-3e89-account-create-update-crnl5\" (UID: \"6ad4524a-e07e-4a6e-93b1-5e5e2d7b251c\") " pod="openstack/placement-3e89-account-create-update-crnl5" Feb 27 08:08:09 crc kubenswrapper[4612]: I0227 08:08:09.684283 4612 generic.go:334] "Generic (PLEG): container finished" podID="15aa4339-acbf-4d66-b8d8-ca4ef14bad44" containerID="06474ec45acf96a16a692d4e6bd2ac7bac092af130f80483e085703beac8dec7" exitCode=0 Feb 27 08:08:09 crc kubenswrapper[4612]: I0227 08:08:09.684548 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bc7876d45-k2v9x" event={"ID":"15aa4339-acbf-4d66-b8d8-ca4ef14bad44","Type":"ContainerDied","Data":"06474ec45acf96a16a692d4e6bd2ac7bac092af130f80483e085703beac8dec7"} Feb 27 08:08:09 crc kubenswrapper[4612]: I0227 08:08:09.684586 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bc7876d45-k2v9x" event={"ID":"15aa4339-acbf-4d66-b8d8-ca4ef14bad44","Type":"ContainerDied","Data":"d56d40a9f9e3aac6a8261f759cc6ecd7837bf6cbc9d1bd97f7fee3146dec8b72"} Feb 27 08:08:09 crc kubenswrapper[4612]: I0227 08:08:09.684632 4612 scope.go:117] "RemoveContainer" containerID="06474ec45acf96a16a692d4e6bd2ac7bac092af130f80483e085703beac8dec7" Feb 27 08:08:09 crc kubenswrapper[4612]: I0227 08:08:09.684789 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bc7876d45-k2v9x" Feb 27 08:08:09 crc kubenswrapper[4612]: I0227 08:08:09.728100 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6bc7876d45-k2v9x"] Feb 27 08:08:09 crc kubenswrapper[4612]: I0227 08:08:09.738246 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-3e89-account-create-update-crnl5" Feb 27 08:08:09 crc kubenswrapper[4612]: I0227 08:08:09.742519 4612 scope.go:117] "RemoveContainer" containerID="7cd099f6ddd6f47437ac0ab4a9df73420ed2890fe4fa713c2ee35e8a52b44d47" Feb 27 08:08:09 crc kubenswrapper[4612]: I0227 08:08:09.760739 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6bc7876d45-k2v9x"] Feb 27 08:08:09 crc kubenswrapper[4612]: I0227 08:08:09.800117 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-4ndtf"] Feb 27 08:08:09 crc kubenswrapper[4612]: I0227 08:08:09.804276 4612 scope.go:117] "RemoveContainer" containerID="06474ec45acf96a16a692d4e6bd2ac7bac092af130f80483e085703beac8dec7" Feb 27 08:08:09 crc kubenswrapper[4612]: E0227 08:08:09.804614 4612 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"06474ec45acf96a16a692d4e6bd2ac7bac092af130f80483e085703beac8dec7\": container with ID starting with 06474ec45acf96a16a692d4e6bd2ac7bac092af130f80483e085703beac8dec7 not found: ID does not exist" containerID="06474ec45acf96a16a692d4e6bd2ac7bac092af130f80483e085703beac8dec7" Feb 27 08:08:09 crc kubenswrapper[4612]: I0227 08:08:09.804653 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"06474ec45acf96a16a692d4e6bd2ac7bac092af130f80483e085703beac8dec7"} err="failed to get container status \"06474ec45acf96a16a692d4e6bd2ac7bac092af130f80483e085703beac8dec7\": rpc error: code = NotFound desc = could not find container \"06474ec45acf96a16a692d4e6bd2ac7bac092af130f80483e085703beac8dec7\": container with ID starting with 06474ec45acf96a16a692d4e6bd2ac7bac092af130f80483e085703beac8dec7 not found: ID does not exist" Feb 27 08:08:09 crc kubenswrapper[4612]: I0227 08:08:09.804705 4612 scope.go:117] "RemoveContainer" containerID="7cd099f6ddd6f47437ac0ab4a9df73420ed2890fe4fa713c2ee35e8a52b44d47" Feb 27 08:08:09 crc kubenswrapper[4612]: E0227 08:08:09.804946 4612 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7cd099f6ddd6f47437ac0ab4a9df73420ed2890fe4fa713c2ee35e8a52b44d47\": container with ID starting with 7cd099f6ddd6f47437ac0ab4a9df73420ed2890fe4fa713c2ee35e8a52b44d47 not found: ID does not exist" containerID="7cd099f6ddd6f47437ac0ab4a9df73420ed2890fe4fa713c2ee35e8a52b44d47" Feb 27 08:08:09 crc kubenswrapper[4612]: I0227 08:08:09.804971 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7cd099f6ddd6f47437ac0ab4a9df73420ed2890fe4fa713c2ee35e8a52b44d47"} err="failed to get container status \"7cd099f6ddd6f47437ac0ab4a9df73420ed2890fe4fa713c2ee35e8a52b44d47\": rpc error: code = NotFound desc = could not find container \"7cd099f6ddd6f47437ac0ab4a9df73420ed2890fe4fa713c2ee35e8a52b44d47\": container with ID starting with 7cd099f6ddd6f47437ac0ab4a9df73420ed2890fe4fa713c2ee35e8a52b44d47 not found: ID does not exist" Feb 27 08:08:09 crc kubenswrapper[4612]: W0227 08:08:09.808878 4612 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podca30e354_dd13_4fa2_96a9_bbd6c605b1f8.slice/crio-f2ed3c9fe0c68d2c39512e5c74abd7b79f70e6b1c32ce725ee6d8f9735bae323 WatchSource:0}: Error finding container f2ed3c9fe0c68d2c39512e5c74abd7b79f70e6b1c32ce725ee6d8f9735bae323: Status 404 returned error can't find the container with id f2ed3c9fe0c68d2c39512e5c74abd7b79f70e6b1c32ce725ee6d8f9735bae323 Feb 27 08:08:09 crc kubenswrapper[4612]: I0227 08:08:09.814293 4612 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 27 08:08:10 crc kubenswrapper[4612]: I0227 08:08:09.874188 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-lkbpz"] Feb 27 08:08:10 crc kubenswrapper[4612]: W0227 08:08:09.885174 4612 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod22a1a3d2_6c5d_4102_8fd4_07f953e402f7.slice/crio-fb09e7d0368ebacdf9b67eabd2e0d6041b8446a4666ab50b12542dcc1916e7ab WatchSource:0}: Error finding container fb09e7d0368ebacdf9b67eabd2e0d6041b8446a4666ab50b12542dcc1916e7ab: Status 404 returned error can't find the container with id fb09e7d0368ebacdf9b67eabd2e0d6041b8446a4666ab50b12542dcc1916e7ab Feb 27 08:08:10 crc kubenswrapper[4612]: I0227 08:08:10.057774 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-r47j8"] Feb 27 08:08:10 crc kubenswrapper[4612]: W0227 08:08:10.082322 4612 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1551171a_913f_4d99_b4da_148ea454f7f6.slice/crio-45b2539534cde41a27c7e0a0bc93a6e92ca4f14c7a844de62f891cd3eb69a4c3 WatchSource:0}: Error finding container 45b2539534cde41a27c7e0a0bc93a6e92ca4f14c7a844de62f891cd3eb69a4c3: Status 404 returned error can't find the container with id 45b2539534cde41a27c7e0a0bc93a6e92ca4f14c7a844de62f891cd3eb69a4c3 Feb 27 08:08:10 crc kubenswrapper[4612]: I0227 08:08:10.683073 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-s2x4c"] Feb 27 08:08:10 crc kubenswrapper[4612]: E0227 08:08:10.683943 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="15aa4339-acbf-4d66-b8d8-ca4ef14bad44" containerName="init" Feb 27 08:08:10 crc kubenswrapper[4612]: I0227 08:08:10.683976 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="15aa4339-acbf-4d66-b8d8-ca4ef14bad44" containerName="init" Feb 27 08:08:10 crc kubenswrapper[4612]: E0227 08:08:10.684007 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="15aa4339-acbf-4d66-b8d8-ca4ef14bad44" containerName="dnsmasq-dns" Feb 27 08:08:10 crc kubenswrapper[4612]: I0227 08:08:10.684016 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="15aa4339-acbf-4d66-b8d8-ca4ef14bad44" containerName="dnsmasq-dns" Feb 27 08:08:10 crc kubenswrapper[4612]: I0227 08:08:10.684283 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="15aa4339-acbf-4d66-b8d8-ca4ef14bad44" containerName="dnsmasq-dns" Feb 27 08:08:10 crc kubenswrapper[4612]: I0227 08:08:10.685493 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8fbc5445-s2x4c" Feb 27 08:08:10 crc kubenswrapper[4612]: I0227 08:08:10.729660 4612 generic.go:334] "Generic (PLEG): container finished" podID="1551171a-913f-4d99-b4da-148ea454f7f6" containerID="ccaff13d789c8a216b77ec37c3f21f72a4ebf4890273ea56836de2332d1e0623" exitCode=0 Feb 27 08:08:10 crc kubenswrapper[4612]: I0227 08:08:10.730108 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-r47j8" event={"ID":"1551171a-913f-4d99-b4da-148ea454f7f6","Type":"ContainerDied","Data":"ccaff13d789c8a216b77ec37c3f21f72a4ebf4890273ea56836de2332d1e0623"} Feb 27 08:08:10 crc kubenswrapper[4612]: I0227 08:08:10.730151 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-r47j8" event={"ID":"1551171a-913f-4d99-b4da-148ea454f7f6","Type":"ContainerStarted","Data":"45b2539534cde41a27c7e0a0bc93a6e92ca4f14c7a844de62f891cd3eb69a4c3"} Feb 27 08:08:10 crc kubenswrapper[4612]: I0227 08:08:10.741534 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-4ndtf" event={"ID":"ca30e354-dd13-4fa2-96a9-bbd6c605b1f8","Type":"ContainerStarted","Data":"f2ed3c9fe0c68d2c39512e5c74abd7b79f70e6b1c32ce725ee6d8f9735bae323"} Feb 27 08:08:10 crc kubenswrapper[4612]: I0227 08:08:10.743506 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c7a8adaa-cf40-4c58-964e-4bc43934baa1-ovsdbserver-sb\") pod \"dnsmasq-dns-b8fbc5445-s2x4c\" (UID: \"c7a8adaa-cf40-4c58-964e-4bc43934baa1\") " pod="openstack/dnsmasq-dns-b8fbc5445-s2x4c" Feb 27 08:08:10 crc kubenswrapper[4612]: I0227 08:08:10.743573 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c7a8adaa-cf40-4c58-964e-4bc43934baa1-ovsdbserver-nb\") pod \"dnsmasq-dns-b8fbc5445-s2x4c\" (UID: \"c7a8adaa-cf40-4c58-964e-4bc43934baa1\") " pod="openstack/dnsmasq-dns-b8fbc5445-s2x4c" Feb 27 08:08:10 crc kubenswrapper[4612]: I0227 08:08:10.743601 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5lxnn\" (UniqueName: \"kubernetes.io/projected/c7a8adaa-cf40-4c58-964e-4bc43934baa1-kube-api-access-5lxnn\") pod \"dnsmasq-dns-b8fbc5445-s2x4c\" (UID: \"c7a8adaa-cf40-4c58-964e-4bc43934baa1\") " pod="openstack/dnsmasq-dns-b8fbc5445-s2x4c" Feb 27 08:08:10 crc kubenswrapper[4612]: I0227 08:08:10.743660 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c7a8adaa-cf40-4c58-964e-4bc43934baa1-dns-svc\") pod \"dnsmasq-dns-b8fbc5445-s2x4c\" (UID: \"c7a8adaa-cf40-4c58-964e-4bc43934baa1\") " pod="openstack/dnsmasq-dns-b8fbc5445-s2x4c" Feb 27 08:08:10 crc kubenswrapper[4612]: I0227 08:08:10.743710 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c7a8adaa-cf40-4c58-964e-4bc43934baa1-config\") pod \"dnsmasq-dns-b8fbc5445-s2x4c\" (UID: \"c7a8adaa-cf40-4c58-964e-4bc43934baa1\") " pod="openstack/dnsmasq-dns-b8fbc5445-s2x4c" Feb 27 08:08:10 crc kubenswrapper[4612]: I0227 08:08:10.773372 4612 generic.go:334] "Generic (PLEG): container finished" podID="22a1a3d2-6c5d-4102-8fd4-07f953e402f7" containerID="97ff21863c2120601b515d9bf34ce20c3286d9f973a71672a8251e430cf070dd" exitCode=0 Feb 27 08:08:10 crc kubenswrapper[4612]: I0227 08:08:10.773478 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-lkbpz" event={"ID":"22a1a3d2-6c5d-4102-8fd4-07f953e402f7","Type":"ContainerDied","Data":"97ff21863c2120601b515d9bf34ce20c3286d9f973a71672a8251e430cf070dd"} Feb 27 08:08:10 crc kubenswrapper[4612]: I0227 08:08:10.773515 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-lkbpz" event={"ID":"22a1a3d2-6c5d-4102-8fd4-07f953e402f7","Type":"ContainerStarted","Data":"fb09e7d0368ebacdf9b67eabd2e0d6041b8446a4666ab50b12542dcc1916e7ab"} Feb 27 08:08:10 crc kubenswrapper[4612]: I0227 08:08:10.792933 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-s2x4c"] Feb 27 08:08:10 crc kubenswrapper[4612]: I0227 08:08:10.844677 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c7a8adaa-cf40-4c58-964e-4bc43934baa1-ovsdbserver-nb\") pod \"dnsmasq-dns-b8fbc5445-s2x4c\" (UID: \"c7a8adaa-cf40-4c58-964e-4bc43934baa1\") " pod="openstack/dnsmasq-dns-b8fbc5445-s2x4c" Feb 27 08:08:10 crc kubenswrapper[4612]: I0227 08:08:10.844740 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5lxnn\" (UniqueName: \"kubernetes.io/projected/c7a8adaa-cf40-4c58-964e-4bc43934baa1-kube-api-access-5lxnn\") pod \"dnsmasq-dns-b8fbc5445-s2x4c\" (UID: \"c7a8adaa-cf40-4c58-964e-4bc43934baa1\") " pod="openstack/dnsmasq-dns-b8fbc5445-s2x4c" Feb 27 08:08:10 crc kubenswrapper[4612]: I0227 08:08:10.844808 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c7a8adaa-cf40-4c58-964e-4bc43934baa1-dns-svc\") pod \"dnsmasq-dns-b8fbc5445-s2x4c\" (UID: \"c7a8adaa-cf40-4c58-964e-4bc43934baa1\") " pod="openstack/dnsmasq-dns-b8fbc5445-s2x4c" Feb 27 08:08:10 crc kubenswrapper[4612]: I0227 08:08:10.844835 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c7a8adaa-cf40-4c58-964e-4bc43934baa1-config\") pod \"dnsmasq-dns-b8fbc5445-s2x4c\" (UID: \"c7a8adaa-cf40-4c58-964e-4bc43934baa1\") " pod="openstack/dnsmasq-dns-b8fbc5445-s2x4c" Feb 27 08:08:10 crc kubenswrapper[4612]: I0227 08:08:10.844888 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c7a8adaa-cf40-4c58-964e-4bc43934baa1-ovsdbserver-sb\") pod \"dnsmasq-dns-b8fbc5445-s2x4c\" (UID: \"c7a8adaa-cf40-4c58-964e-4bc43934baa1\") " pod="openstack/dnsmasq-dns-b8fbc5445-s2x4c" Feb 27 08:08:10 crc kubenswrapper[4612]: I0227 08:08:10.854746 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c7a8adaa-cf40-4c58-964e-4bc43934baa1-ovsdbserver-nb\") pod \"dnsmasq-dns-b8fbc5445-s2x4c\" (UID: \"c7a8adaa-cf40-4c58-964e-4bc43934baa1\") " pod="openstack/dnsmasq-dns-b8fbc5445-s2x4c" Feb 27 08:08:10 crc kubenswrapper[4612]: I0227 08:08:10.855242 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c7a8adaa-cf40-4c58-964e-4bc43934baa1-ovsdbserver-sb\") pod \"dnsmasq-dns-b8fbc5445-s2x4c\" (UID: \"c7a8adaa-cf40-4c58-964e-4bc43934baa1\") " pod="openstack/dnsmasq-dns-b8fbc5445-s2x4c" Feb 27 08:08:10 crc kubenswrapper[4612]: I0227 08:08:10.856656 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c7a8adaa-cf40-4c58-964e-4bc43934baa1-dns-svc\") pod \"dnsmasq-dns-b8fbc5445-s2x4c\" (UID: \"c7a8adaa-cf40-4c58-964e-4bc43934baa1\") " pod="openstack/dnsmasq-dns-b8fbc5445-s2x4c" Feb 27 08:08:10 crc kubenswrapper[4612]: I0227 08:08:10.858366 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c7a8adaa-cf40-4c58-964e-4bc43934baa1-config\") pod \"dnsmasq-dns-b8fbc5445-s2x4c\" (UID: \"c7a8adaa-cf40-4c58-964e-4bc43934baa1\") " pod="openstack/dnsmasq-dns-b8fbc5445-s2x4c" Feb 27 08:08:10 crc kubenswrapper[4612]: I0227 08:08:10.879205 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="15aa4339-acbf-4d66-b8d8-ca4ef14bad44" path="/var/lib/kubelet/pods/15aa4339-acbf-4d66-b8d8-ca4ef14bad44/volumes" Feb 27 08:08:10 crc kubenswrapper[4612]: I0227 08:08:10.899325 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5lxnn\" (UniqueName: \"kubernetes.io/projected/c7a8adaa-cf40-4c58-964e-4bc43934baa1-kube-api-access-5lxnn\") pod \"dnsmasq-dns-b8fbc5445-s2x4c\" (UID: \"c7a8adaa-cf40-4c58-964e-4bc43934baa1\") " pod="openstack/dnsmasq-dns-b8fbc5445-s2x4c" Feb 27 08:08:10 crc kubenswrapper[4612]: I0227 08:08:10.912364 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-7819-account-create-update-lppdh"] Feb 27 08:08:10 crc kubenswrapper[4612]: W0227 08:08:10.919103 4612 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod03d154b1_4114_4c56_82b0_5bf148511e5f.slice/crio-2d3a191afc4bc62db37584755712098fc0f1aeedc44e40c56f6156f0fbecbaf1 WatchSource:0}: Error finding container 2d3a191afc4bc62db37584755712098fc0f1aeedc44e40c56f6156f0fbecbaf1: Status 404 returned error can't find the container with id 2d3a191afc4bc62db37584755712098fc0f1aeedc44e40c56f6156f0fbecbaf1 Feb 27 08:08:11 crc kubenswrapper[4612]: I0227 08:08:11.036311 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-3e89-account-create-update-crnl5"] Feb 27 08:08:11 crc kubenswrapper[4612]: I0227 08:08:11.076269 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8fbc5445-s2x4c" Feb 27 08:08:11 crc kubenswrapper[4612]: I0227 08:08:11.641834 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-s2x4c"] Feb 27 08:08:11 crc kubenswrapper[4612]: I0227 08:08:11.805624 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-s2x4c" event={"ID":"c7a8adaa-cf40-4c58-964e-4bc43934baa1","Type":"ContainerStarted","Data":"34641ee5ef5076347513a394ae91cb085c75f5f357ab620d54332604c68b8bad"} Feb 27 08:08:11 crc kubenswrapper[4612]: I0227 08:08:11.806929 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-7819-account-create-update-lppdh" event={"ID":"03d154b1-4114-4c56-82b0-5bf148511e5f","Type":"ContainerStarted","Data":"b11b057c5ff66e783eb8beacc2972c189703b113319ce71704cf6108ad5c60d9"} Feb 27 08:08:11 crc kubenswrapper[4612]: I0227 08:08:11.806970 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-7819-account-create-update-lppdh" event={"ID":"03d154b1-4114-4c56-82b0-5bf148511e5f","Type":"ContainerStarted","Data":"2d3a191afc4bc62db37584755712098fc0f1aeedc44e40c56f6156f0fbecbaf1"} Feb 27 08:08:11 crc kubenswrapper[4612]: I0227 08:08:11.809708 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-3e89-account-create-update-crnl5" event={"ID":"6ad4524a-e07e-4a6e-93b1-5e5e2d7b251c","Type":"ContainerStarted","Data":"7293eec8c40e09308544339c5238b1ec2a2fda76a771a6a729697572a772cb65"} Feb 27 08:08:11 crc kubenswrapper[4612]: I0227 08:08:11.809732 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-3e89-account-create-update-crnl5" event={"ID":"6ad4524a-e07e-4a6e-93b1-5e5e2d7b251c","Type":"ContainerStarted","Data":"33f9559e847061c7c265ff46b1d08a56e2a04e09539552175f6a1ba35f0dc4ce"} Feb 27 08:08:11 crc kubenswrapper[4612]: I0227 08:08:11.829218 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-7819-account-create-update-lppdh" podStartSLOduration=2.829202918 podStartE2EDuration="2.829202918s" podCreationTimestamp="2026-02-27 08:08:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 08:08:11.827824629 +0000 UTC m=+1149.681754627" watchObservedRunningTime="2026-02-27 08:08:11.829202918 +0000 UTC m=+1149.683132916" Feb 27 08:08:11 crc kubenswrapper[4612]: I0227 08:08:11.847209 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-3e89-account-create-update-crnl5" podStartSLOduration=2.847195444 podStartE2EDuration="2.847195444s" podCreationTimestamp="2026-02-27 08:08:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 08:08:11.847094501 +0000 UTC m=+1149.701024499" watchObservedRunningTime="2026-02-27 08:08:11.847195444 +0000 UTC m=+1149.701125442" Feb 27 08:08:12 crc kubenswrapper[4612]: I0227 08:08:12.021913 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-storage-0"] Feb 27 08:08:12 crc kubenswrapper[4612]: I0227 08:08:12.041585 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Feb 27 08:08:12 crc kubenswrapper[4612]: I0227 08:08:12.055738 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-storage-config-data" Feb 27 08:08:12 crc kubenswrapper[4612]: I0227 08:08:12.055781 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-swift-dockercfg-phft2" Feb 27 08:08:12 crc kubenswrapper[4612]: I0227 08:08:12.056286 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-conf" Feb 27 08:08:12 crc kubenswrapper[4612]: I0227 08:08:12.056391 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-files" Feb 27 08:08:12 crc kubenswrapper[4612]: I0227 08:08:12.069351 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"swift-storage-0\" (UID: \"77288860-cf3c-47e6-a752-5b41aa1967fa\") " pod="openstack/swift-storage-0" Feb 27 08:08:12 crc kubenswrapper[4612]: I0227 08:08:12.069390 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/77288860-cf3c-47e6-a752-5b41aa1967fa-etc-swift\") pod \"swift-storage-0\" (UID: \"77288860-cf3c-47e6-a752-5b41aa1967fa\") " pod="openstack/swift-storage-0" Feb 27 08:08:12 crc kubenswrapper[4612]: I0227 08:08:12.069413 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/77288860-cf3c-47e6-a752-5b41aa1967fa-cache\") pod \"swift-storage-0\" (UID: \"77288860-cf3c-47e6-a752-5b41aa1967fa\") " pod="openstack/swift-storage-0" Feb 27 08:08:12 crc kubenswrapper[4612]: I0227 08:08:12.069452 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/77288860-cf3c-47e6-a752-5b41aa1967fa-lock\") pod \"swift-storage-0\" (UID: \"77288860-cf3c-47e6-a752-5b41aa1967fa\") " pod="openstack/swift-storage-0" Feb 27 08:08:12 crc kubenswrapper[4612]: I0227 08:08:12.069504 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kkpnj\" (UniqueName: \"kubernetes.io/projected/77288860-cf3c-47e6-a752-5b41aa1967fa-kube-api-access-kkpnj\") pod \"swift-storage-0\" (UID: \"77288860-cf3c-47e6-a752-5b41aa1967fa\") " pod="openstack/swift-storage-0" Feb 27 08:08:12 crc kubenswrapper[4612]: I0227 08:08:12.069547 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77288860-cf3c-47e6-a752-5b41aa1967fa-combined-ca-bundle\") pod \"swift-storage-0\" (UID: \"77288860-cf3c-47e6-a752-5b41aa1967fa\") " pod="openstack/swift-storage-0" Feb 27 08:08:12 crc kubenswrapper[4612]: I0227 08:08:12.074724 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Feb 27 08:08:12 crc kubenswrapper[4612]: I0227 08:08:12.171125 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/77288860-cf3c-47e6-a752-5b41aa1967fa-lock\") pod \"swift-storage-0\" (UID: \"77288860-cf3c-47e6-a752-5b41aa1967fa\") " pod="openstack/swift-storage-0" Feb 27 08:08:12 crc kubenswrapper[4612]: I0227 08:08:12.171443 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kkpnj\" (UniqueName: \"kubernetes.io/projected/77288860-cf3c-47e6-a752-5b41aa1967fa-kube-api-access-kkpnj\") pod \"swift-storage-0\" (UID: \"77288860-cf3c-47e6-a752-5b41aa1967fa\") " pod="openstack/swift-storage-0" Feb 27 08:08:12 crc kubenswrapper[4612]: I0227 08:08:12.171527 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77288860-cf3c-47e6-a752-5b41aa1967fa-combined-ca-bundle\") pod \"swift-storage-0\" (UID: \"77288860-cf3c-47e6-a752-5b41aa1967fa\") " pod="openstack/swift-storage-0" Feb 27 08:08:12 crc kubenswrapper[4612]: I0227 08:08:12.171569 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"swift-storage-0\" (UID: \"77288860-cf3c-47e6-a752-5b41aa1967fa\") " pod="openstack/swift-storage-0" Feb 27 08:08:12 crc kubenswrapper[4612]: I0227 08:08:12.171596 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/77288860-cf3c-47e6-a752-5b41aa1967fa-etc-swift\") pod \"swift-storage-0\" (UID: \"77288860-cf3c-47e6-a752-5b41aa1967fa\") " pod="openstack/swift-storage-0" Feb 27 08:08:12 crc kubenswrapper[4612]: I0227 08:08:12.171626 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/77288860-cf3c-47e6-a752-5b41aa1967fa-cache\") pod \"swift-storage-0\" (UID: \"77288860-cf3c-47e6-a752-5b41aa1967fa\") " pod="openstack/swift-storage-0" Feb 27 08:08:12 crc kubenswrapper[4612]: I0227 08:08:12.172062 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/77288860-cf3c-47e6-a752-5b41aa1967fa-cache\") pod \"swift-storage-0\" (UID: \"77288860-cf3c-47e6-a752-5b41aa1967fa\") " pod="openstack/swift-storage-0" Feb 27 08:08:12 crc kubenswrapper[4612]: E0227 08:08:12.173869 4612 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Feb 27 08:08:12 crc kubenswrapper[4612]: E0227 08:08:12.173888 4612 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Feb 27 08:08:12 crc kubenswrapper[4612]: E0227 08:08:12.173927 4612 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/77288860-cf3c-47e6-a752-5b41aa1967fa-etc-swift podName:77288860-cf3c-47e6-a752-5b41aa1967fa nodeName:}" failed. No retries permitted until 2026-02-27 08:08:12.673911325 +0000 UTC m=+1150.527841323 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/77288860-cf3c-47e6-a752-5b41aa1967fa-etc-swift") pod "swift-storage-0" (UID: "77288860-cf3c-47e6-a752-5b41aa1967fa") : configmap "swift-ring-files" not found Feb 27 08:08:12 crc kubenswrapper[4612]: I0227 08:08:12.174082 4612 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"swift-storage-0\" (UID: \"77288860-cf3c-47e6-a752-5b41aa1967fa\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/swift-storage-0" Feb 27 08:08:12 crc kubenswrapper[4612]: I0227 08:08:12.174275 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/77288860-cf3c-47e6-a752-5b41aa1967fa-lock\") pod \"swift-storage-0\" (UID: \"77288860-cf3c-47e6-a752-5b41aa1967fa\") " pod="openstack/swift-storage-0" Feb 27 08:08:12 crc kubenswrapper[4612]: I0227 08:08:12.180015 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77288860-cf3c-47e6-a752-5b41aa1967fa-combined-ca-bundle\") pod \"swift-storage-0\" (UID: \"77288860-cf3c-47e6-a752-5b41aa1967fa\") " pod="openstack/swift-storage-0" Feb 27 08:08:12 crc kubenswrapper[4612]: I0227 08:08:12.198411 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-gp8mt"] Feb 27 08:08:12 crc kubenswrapper[4612]: I0227 08:08:12.199709 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-gp8mt" Feb 27 08:08:12 crc kubenswrapper[4612]: I0227 08:08:12.214461 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kkpnj\" (UniqueName: \"kubernetes.io/projected/77288860-cf3c-47e6-a752-5b41aa1967fa-kube-api-access-kkpnj\") pod \"swift-storage-0\" (UID: \"77288860-cf3c-47e6-a752-5b41aa1967fa\") " pod="openstack/swift-storage-0" Feb 27 08:08:12 crc kubenswrapper[4612]: I0227 08:08:12.230564 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Feb 27 08:08:12 crc kubenswrapper[4612]: I0227 08:08:12.230801 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Feb 27 08:08:12 crc kubenswrapper[4612]: I0227 08:08:12.230907 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Feb 27 08:08:12 crc kubenswrapper[4612]: I0227 08:08:12.255457 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-gp8mt"] Feb 27 08:08:12 crc kubenswrapper[4612]: I0227 08:08:12.270272 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-r47j8" Feb 27 08:08:12 crc kubenswrapper[4612]: I0227 08:08:12.272807 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/76db835b-ff4b-4d34-9578-38c5e3a605ef-dispersionconf\") pod \"swift-ring-rebalance-gp8mt\" (UID: \"76db835b-ff4b-4d34-9578-38c5e3a605ef\") " pod="openstack/swift-ring-rebalance-gp8mt" Feb 27 08:08:12 crc kubenswrapper[4612]: I0227 08:08:12.272854 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76db835b-ff4b-4d34-9578-38c5e3a605ef-combined-ca-bundle\") pod \"swift-ring-rebalance-gp8mt\" (UID: \"76db835b-ff4b-4d34-9578-38c5e3a605ef\") " pod="openstack/swift-ring-rebalance-gp8mt" Feb 27 08:08:12 crc kubenswrapper[4612]: I0227 08:08:12.272879 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/76db835b-ff4b-4d34-9578-38c5e3a605ef-ring-data-devices\") pod \"swift-ring-rebalance-gp8mt\" (UID: \"76db835b-ff4b-4d34-9578-38c5e3a605ef\") " pod="openstack/swift-ring-rebalance-gp8mt" Feb 27 08:08:12 crc kubenswrapper[4612]: I0227 08:08:12.272925 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/76db835b-ff4b-4d34-9578-38c5e3a605ef-swiftconf\") pod \"swift-ring-rebalance-gp8mt\" (UID: \"76db835b-ff4b-4d34-9578-38c5e3a605ef\") " pod="openstack/swift-ring-rebalance-gp8mt" Feb 27 08:08:12 crc kubenswrapper[4612]: I0227 08:08:12.272953 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4g5j5\" (UniqueName: \"kubernetes.io/projected/76db835b-ff4b-4d34-9578-38c5e3a605ef-kube-api-access-4g5j5\") pod \"swift-ring-rebalance-gp8mt\" (UID: \"76db835b-ff4b-4d34-9578-38c5e3a605ef\") " pod="openstack/swift-ring-rebalance-gp8mt" Feb 27 08:08:12 crc kubenswrapper[4612]: I0227 08:08:12.273035 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/76db835b-ff4b-4d34-9578-38c5e3a605ef-scripts\") pod \"swift-ring-rebalance-gp8mt\" (UID: \"76db835b-ff4b-4d34-9578-38c5e3a605ef\") " pod="openstack/swift-ring-rebalance-gp8mt" Feb 27 08:08:12 crc kubenswrapper[4612]: I0227 08:08:12.273074 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/76db835b-ff4b-4d34-9578-38c5e3a605ef-etc-swift\") pod \"swift-ring-rebalance-gp8mt\" (UID: \"76db835b-ff4b-4d34-9578-38c5e3a605ef\") " pod="openstack/swift-ring-rebalance-gp8mt" Feb 27 08:08:12 crc kubenswrapper[4612]: I0227 08:08:12.301945 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"swift-storage-0\" (UID: \"77288860-cf3c-47e6-a752-5b41aa1967fa\") " pod="openstack/swift-storage-0" Feb 27 08:08:12 crc kubenswrapper[4612]: I0227 08:08:12.363101 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-lkbpz" Feb 27 08:08:12 crc kubenswrapper[4612]: I0227 08:08:12.378375 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/76db835b-ff4b-4d34-9578-38c5e3a605ef-swiftconf\") pod \"swift-ring-rebalance-gp8mt\" (UID: \"76db835b-ff4b-4d34-9578-38c5e3a605ef\") " pod="openstack/swift-ring-rebalance-gp8mt" Feb 27 08:08:12 crc kubenswrapper[4612]: I0227 08:08:12.378425 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4g5j5\" (UniqueName: \"kubernetes.io/projected/76db835b-ff4b-4d34-9578-38c5e3a605ef-kube-api-access-4g5j5\") pod \"swift-ring-rebalance-gp8mt\" (UID: \"76db835b-ff4b-4d34-9578-38c5e3a605ef\") " pod="openstack/swift-ring-rebalance-gp8mt" Feb 27 08:08:12 crc kubenswrapper[4612]: I0227 08:08:12.378492 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/76db835b-ff4b-4d34-9578-38c5e3a605ef-scripts\") pod \"swift-ring-rebalance-gp8mt\" (UID: \"76db835b-ff4b-4d34-9578-38c5e3a605ef\") " pod="openstack/swift-ring-rebalance-gp8mt" Feb 27 08:08:12 crc kubenswrapper[4612]: I0227 08:08:12.378528 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/76db835b-ff4b-4d34-9578-38c5e3a605ef-etc-swift\") pod \"swift-ring-rebalance-gp8mt\" (UID: \"76db835b-ff4b-4d34-9578-38c5e3a605ef\") " pod="openstack/swift-ring-rebalance-gp8mt" Feb 27 08:08:12 crc kubenswrapper[4612]: I0227 08:08:12.378552 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/76db835b-ff4b-4d34-9578-38c5e3a605ef-dispersionconf\") pod \"swift-ring-rebalance-gp8mt\" (UID: \"76db835b-ff4b-4d34-9578-38c5e3a605ef\") " pod="openstack/swift-ring-rebalance-gp8mt" Feb 27 08:08:12 crc kubenswrapper[4612]: I0227 08:08:12.378582 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76db835b-ff4b-4d34-9578-38c5e3a605ef-combined-ca-bundle\") pod \"swift-ring-rebalance-gp8mt\" (UID: \"76db835b-ff4b-4d34-9578-38c5e3a605ef\") " pod="openstack/swift-ring-rebalance-gp8mt" Feb 27 08:08:12 crc kubenswrapper[4612]: I0227 08:08:12.378600 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/76db835b-ff4b-4d34-9578-38c5e3a605ef-ring-data-devices\") pod \"swift-ring-rebalance-gp8mt\" (UID: \"76db835b-ff4b-4d34-9578-38c5e3a605ef\") " pod="openstack/swift-ring-rebalance-gp8mt" Feb 27 08:08:12 crc kubenswrapper[4612]: I0227 08:08:12.382001 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/76db835b-ff4b-4d34-9578-38c5e3a605ef-ring-data-devices\") pod \"swift-ring-rebalance-gp8mt\" (UID: \"76db835b-ff4b-4d34-9578-38c5e3a605ef\") " pod="openstack/swift-ring-rebalance-gp8mt" Feb 27 08:08:12 crc kubenswrapper[4612]: I0227 08:08:12.383133 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/76db835b-ff4b-4d34-9578-38c5e3a605ef-etc-swift\") pod \"swift-ring-rebalance-gp8mt\" (UID: \"76db835b-ff4b-4d34-9578-38c5e3a605ef\") " pod="openstack/swift-ring-rebalance-gp8mt" Feb 27 08:08:12 crc kubenswrapper[4612]: I0227 08:08:12.383899 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/76db835b-ff4b-4d34-9578-38c5e3a605ef-scripts\") pod \"swift-ring-rebalance-gp8mt\" (UID: \"76db835b-ff4b-4d34-9578-38c5e3a605ef\") " pod="openstack/swift-ring-rebalance-gp8mt" Feb 27 08:08:12 crc kubenswrapper[4612]: I0227 08:08:12.386142 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/76db835b-ff4b-4d34-9578-38c5e3a605ef-swiftconf\") pod \"swift-ring-rebalance-gp8mt\" (UID: \"76db835b-ff4b-4d34-9578-38c5e3a605ef\") " pod="openstack/swift-ring-rebalance-gp8mt" Feb 27 08:08:12 crc kubenswrapper[4612]: I0227 08:08:12.388259 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/76db835b-ff4b-4d34-9578-38c5e3a605ef-dispersionconf\") pod \"swift-ring-rebalance-gp8mt\" (UID: \"76db835b-ff4b-4d34-9578-38c5e3a605ef\") " pod="openstack/swift-ring-rebalance-gp8mt" Feb 27 08:08:12 crc kubenswrapper[4612]: I0227 08:08:12.395734 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76db835b-ff4b-4d34-9578-38c5e3a605ef-combined-ca-bundle\") pod \"swift-ring-rebalance-gp8mt\" (UID: \"76db835b-ff4b-4d34-9578-38c5e3a605ef\") " pod="openstack/swift-ring-rebalance-gp8mt" Feb 27 08:08:12 crc kubenswrapper[4612]: I0227 08:08:12.404292 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4g5j5\" (UniqueName: \"kubernetes.io/projected/76db835b-ff4b-4d34-9578-38c5e3a605ef-kube-api-access-4g5j5\") pod \"swift-ring-rebalance-gp8mt\" (UID: \"76db835b-ff4b-4d34-9578-38c5e3a605ef\") " pod="openstack/swift-ring-rebalance-gp8mt" Feb 27 08:08:12 crc kubenswrapper[4612]: I0227 08:08:12.483608 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5jh9x\" (UniqueName: \"kubernetes.io/projected/22a1a3d2-6c5d-4102-8fd4-07f953e402f7-kube-api-access-5jh9x\") pod \"22a1a3d2-6c5d-4102-8fd4-07f953e402f7\" (UID: \"22a1a3d2-6c5d-4102-8fd4-07f953e402f7\") " Feb 27 08:08:12 crc kubenswrapper[4612]: I0227 08:08:12.483791 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1551171a-913f-4d99-b4da-148ea454f7f6-operator-scripts\") pod \"1551171a-913f-4d99-b4da-148ea454f7f6\" (UID: \"1551171a-913f-4d99-b4da-148ea454f7f6\") " Feb 27 08:08:12 crc kubenswrapper[4612]: I0227 08:08:12.483828 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/22a1a3d2-6c5d-4102-8fd4-07f953e402f7-operator-scripts\") pod \"22a1a3d2-6c5d-4102-8fd4-07f953e402f7\" (UID: \"22a1a3d2-6c5d-4102-8fd4-07f953e402f7\") " Feb 27 08:08:12 crc kubenswrapper[4612]: I0227 08:08:12.483944 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sp6bh\" (UniqueName: \"kubernetes.io/projected/1551171a-913f-4d99-b4da-148ea454f7f6-kube-api-access-sp6bh\") pod \"1551171a-913f-4d99-b4da-148ea454f7f6\" (UID: \"1551171a-913f-4d99-b4da-148ea454f7f6\") " Feb 27 08:08:12 crc kubenswrapper[4612]: I0227 08:08:12.485088 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22a1a3d2-6c5d-4102-8fd4-07f953e402f7-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "22a1a3d2-6c5d-4102-8fd4-07f953e402f7" (UID: "22a1a3d2-6c5d-4102-8fd4-07f953e402f7"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 08:08:12 crc kubenswrapper[4612]: I0227 08:08:12.485258 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1551171a-913f-4d99-b4da-148ea454f7f6-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "1551171a-913f-4d99-b4da-148ea454f7f6" (UID: "1551171a-913f-4d99-b4da-148ea454f7f6"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 08:08:12 crc kubenswrapper[4612]: I0227 08:08:12.488072 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22a1a3d2-6c5d-4102-8fd4-07f953e402f7-kube-api-access-5jh9x" (OuterVolumeSpecName: "kube-api-access-5jh9x") pod "22a1a3d2-6c5d-4102-8fd4-07f953e402f7" (UID: "22a1a3d2-6c5d-4102-8fd4-07f953e402f7"). InnerVolumeSpecName "kube-api-access-5jh9x". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 08:08:12 crc kubenswrapper[4612]: I0227 08:08:12.488540 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1551171a-913f-4d99-b4da-148ea454f7f6-kube-api-access-sp6bh" (OuterVolumeSpecName: "kube-api-access-sp6bh") pod "1551171a-913f-4d99-b4da-148ea454f7f6" (UID: "1551171a-913f-4d99-b4da-148ea454f7f6"). InnerVolumeSpecName "kube-api-access-sp6bh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 08:08:12 crc kubenswrapper[4612]: I0227 08:08:12.585981 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5jh9x\" (UniqueName: \"kubernetes.io/projected/22a1a3d2-6c5d-4102-8fd4-07f953e402f7-kube-api-access-5jh9x\") on node \"crc\" DevicePath \"\"" Feb 27 08:08:12 crc kubenswrapper[4612]: I0227 08:08:12.586009 4612 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1551171a-913f-4d99-b4da-148ea454f7f6-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 27 08:08:12 crc kubenswrapper[4612]: I0227 08:08:12.586018 4612 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/22a1a3d2-6c5d-4102-8fd4-07f953e402f7-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 27 08:08:12 crc kubenswrapper[4612]: I0227 08:08:12.586028 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sp6bh\" (UniqueName: \"kubernetes.io/projected/1551171a-913f-4d99-b4da-148ea454f7f6-kube-api-access-sp6bh\") on node \"crc\" DevicePath \"\"" Feb 27 08:08:12 crc kubenswrapper[4612]: I0227 08:08:12.592024 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-gp8mt" Feb 27 08:08:12 crc kubenswrapper[4612]: I0227 08:08:12.689851 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/77288860-cf3c-47e6-a752-5b41aa1967fa-etc-swift\") pod \"swift-storage-0\" (UID: \"77288860-cf3c-47e6-a752-5b41aa1967fa\") " pod="openstack/swift-storage-0" Feb 27 08:08:12 crc kubenswrapper[4612]: E0227 08:08:12.690180 4612 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Feb 27 08:08:12 crc kubenswrapper[4612]: E0227 08:08:12.690196 4612 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Feb 27 08:08:12 crc kubenswrapper[4612]: E0227 08:08:12.690245 4612 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/77288860-cf3c-47e6-a752-5b41aa1967fa-etc-swift podName:77288860-cf3c-47e6-a752-5b41aa1967fa nodeName:}" failed. No retries permitted until 2026-02-27 08:08:13.69022994 +0000 UTC m=+1151.544159938 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/77288860-cf3c-47e6-a752-5b41aa1967fa-etc-swift") pod "swift-storage-0" (UID: "77288860-cf3c-47e6-a752-5b41aa1967fa") : configmap "swift-ring-files" not found Feb 27 08:08:12 crc kubenswrapper[4612]: I0227 08:08:12.822342 4612 generic.go:334] "Generic (PLEG): container finished" podID="6ad4524a-e07e-4a6e-93b1-5e5e2d7b251c" containerID="7293eec8c40e09308544339c5238b1ec2a2fda76a771a6a729697572a772cb65" exitCode=0 Feb 27 08:08:12 crc kubenswrapper[4612]: I0227 08:08:12.822657 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-3e89-account-create-update-crnl5" event={"ID":"6ad4524a-e07e-4a6e-93b1-5e5e2d7b251c","Type":"ContainerDied","Data":"7293eec8c40e09308544339c5238b1ec2a2fda76a771a6a729697572a772cb65"} Feb 27 08:08:12 crc kubenswrapper[4612]: I0227 08:08:12.827319 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-lkbpz" event={"ID":"22a1a3d2-6c5d-4102-8fd4-07f953e402f7","Type":"ContainerDied","Data":"fb09e7d0368ebacdf9b67eabd2e0d6041b8446a4666ab50b12542dcc1916e7ab"} Feb 27 08:08:12 crc kubenswrapper[4612]: I0227 08:08:12.827352 4612 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fb09e7d0368ebacdf9b67eabd2e0d6041b8446a4666ab50b12542dcc1916e7ab" Feb 27 08:08:12 crc kubenswrapper[4612]: I0227 08:08:12.827388 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-lkbpz" Feb 27 08:08:12 crc kubenswrapper[4612]: I0227 08:08:12.829108 4612 generic.go:334] "Generic (PLEG): container finished" podID="c7a8adaa-cf40-4c58-964e-4bc43934baa1" containerID="5a3878c9a7846dbfb666822a446df679f70e9cabd148b9bb6c58aa17645290b9" exitCode=0 Feb 27 08:08:12 crc kubenswrapper[4612]: I0227 08:08:12.829173 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-s2x4c" event={"ID":"c7a8adaa-cf40-4c58-964e-4bc43934baa1","Type":"ContainerDied","Data":"5a3878c9a7846dbfb666822a446df679f70e9cabd148b9bb6c58aa17645290b9"} Feb 27 08:08:12 crc kubenswrapper[4612]: I0227 08:08:12.844504 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-r47j8" Feb 27 08:08:12 crc kubenswrapper[4612]: I0227 08:08:12.844520 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-r47j8" event={"ID":"1551171a-913f-4d99-b4da-148ea454f7f6","Type":"ContainerDied","Data":"45b2539534cde41a27c7e0a0bc93a6e92ca4f14c7a844de62f891cd3eb69a4c3"} Feb 27 08:08:12 crc kubenswrapper[4612]: I0227 08:08:12.844648 4612 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="45b2539534cde41a27c7e0a0bc93a6e92ca4f14c7a844de62f891cd3eb69a4c3" Feb 27 08:08:12 crc kubenswrapper[4612]: I0227 08:08:12.851913 4612 generic.go:334] "Generic (PLEG): container finished" podID="03d154b1-4114-4c56-82b0-5bf148511e5f" containerID="b11b057c5ff66e783eb8beacc2972c189703b113319ce71704cf6108ad5c60d9" exitCode=0 Feb 27 08:08:12 crc kubenswrapper[4612]: I0227 08:08:12.851967 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-7819-account-create-update-lppdh" event={"ID":"03d154b1-4114-4c56-82b0-5bf148511e5f","Type":"ContainerDied","Data":"b11b057c5ff66e783eb8beacc2972c189703b113319ce71704cf6108ad5c60d9"} Feb 27 08:08:13 crc kubenswrapper[4612]: I0227 08:08:13.094985 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-gp8mt"] Feb 27 08:08:13 crc kubenswrapper[4612]: I0227 08:08:13.705638 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/77288860-cf3c-47e6-a752-5b41aa1967fa-etc-swift\") pod \"swift-storage-0\" (UID: \"77288860-cf3c-47e6-a752-5b41aa1967fa\") " pod="openstack/swift-storage-0" Feb 27 08:08:13 crc kubenswrapper[4612]: E0227 08:08:13.705941 4612 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Feb 27 08:08:13 crc kubenswrapper[4612]: E0227 08:08:13.705967 4612 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Feb 27 08:08:13 crc kubenswrapper[4612]: E0227 08:08:13.706021 4612 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/77288860-cf3c-47e6-a752-5b41aa1967fa-etc-swift podName:77288860-cf3c-47e6-a752-5b41aa1967fa nodeName:}" failed. No retries permitted until 2026-02-27 08:08:15.706001115 +0000 UTC m=+1153.559931123 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/77288860-cf3c-47e6-a752-5b41aa1967fa-etc-swift") pod "swift-storage-0" (UID: "77288860-cf3c-47e6-a752-5b41aa1967fa") : configmap "swift-ring-files" not found Feb 27 08:08:13 crc kubenswrapper[4612]: I0227 08:08:13.870447 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-s2x4c" event={"ID":"c7a8adaa-cf40-4c58-964e-4bc43934baa1","Type":"ContainerStarted","Data":"f9dcd8a069d0a8cd4b275be183e3b0b8e7b97cfaa9dc697c783028f9c8d0564b"} Feb 27 08:08:13 crc kubenswrapper[4612]: I0227 08:08:13.871944 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-b8fbc5445-s2x4c" Feb 27 08:08:13 crc kubenswrapper[4612]: I0227 08:08:13.881127 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-gp8mt" event={"ID":"76db835b-ff4b-4d34-9578-38c5e3a605ef","Type":"ContainerStarted","Data":"a661d2d8b6154865ea1cc647ecbcb11463e54468b72800b04ed32606de00511f"} Feb 27 08:08:13 crc kubenswrapper[4612]: I0227 08:08:13.898581 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-b8fbc5445-s2x4c" podStartSLOduration=3.898536792 podStartE2EDuration="3.898536792s" podCreationTimestamp="2026-02-27 08:08:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 08:08:13.896624667 +0000 UTC m=+1151.750554665" watchObservedRunningTime="2026-02-27 08:08:13.898536792 +0000 UTC m=+1151.752466790" Feb 27 08:08:13 crc kubenswrapper[4612]: I0227 08:08:13.982505 4612 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Feb 27 08:08:14 crc kubenswrapper[4612]: I0227 08:08:14.122947 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Feb 27 08:08:14 crc kubenswrapper[4612]: I0227 08:08:14.451361 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-3e89-account-create-update-crnl5" Feb 27 08:08:14 crc kubenswrapper[4612]: I0227 08:08:14.453771 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-7819-account-create-update-lppdh" Feb 27 08:08:14 crc kubenswrapper[4612]: I0227 08:08:14.644806 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v4mtb\" (UniqueName: \"kubernetes.io/projected/03d154b1-4114-4c56-82b0-5bf148511e5f-kube-api-access-v4mtb\") pod \"03d154b1-4114-4c56-82b0-5bf148511e5f\" (UID: \"03d154b1-4114-4c56-82b0-5bf148511e5f\") " Feb 27 08:08:14 crc kubenswrapper[4612]: I0227 08:08:14.644893 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jmqz8\" (UniqueName: \"kubernetes.io/projected/6ad4524a-e07e-4a6e-93b1-5e5e2d7b251c-kube-api-access-jmqz8\") pod \"6ad4524a-e07e-4a6e-93b1-5e5e2d7b251c\" (UID: \"6ad4524a-e07e-4a6e-93b1-5e5e2d7b251c\") " Feb 27 08:08:14 crc kubenswrapper[4612]: I0227 08:08:14.644929 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6ad4524a-e07e-4a6e-93b1-5e5e2d7b251c-operator-scripts\") pod \"6ad4524a-e07e-4a6e-93b1-5e5e2d7b251c\" (UID: \"6ad4524a-e07e-4a6e-93b1-5e5e2d7b251c\") " Feb 27 08:08:14 crc kubenswrapper[4612]: I0227 08:08:14.645060 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/03d154b1-4114-4c56-82b0-5bf148511e5f-operator-scripts\") pod \"03d154b1-4114-4c56-82b0-5bf148511e5f\" (UID: \"03d154b1-4114-4c56-82b0-5bf148511e5f\") " Feb 27 08:08:14 crc kubenswrapper[4612]: I0227 08:08:14.645731 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/03d154b1-4114-4c56-82b0-5bf148511e5f-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "03d154b1-4114-4c56-82b0-5bf148511e5f" (UID: "03d154b1-4114-4c56-82b0-5bf148511e5f"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 08:08:14 crc kubenswrapper[4612]: I0227 08:08:14.645761 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ad4524a-e07e-4a6e-93b1-5e5e2d7b251c-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "6ad4524a-e07e-4a6e-93b1-5e5e2d7b251c" (UID: "6ad4524a-e07e-4a6e-93b1-5e5e2d7b251c"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 08:08:14 crc kubenswrapper[4612]: I0227 08:08:14.653440 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/03d154b1-4114-4c56-82b0-5bf148511e5f-kube-api-access-v4mtb" (OuterVolumeSpecName: "kube-api-access-v4mtb") pod "03d154b1-4114-4c56-82b0-5bf148511e5f" (UID: "03d154b1-4114-4c56-82b0-5bf148511e5f"). InnerVolumeSpecName "kube-api-access-v4mtb". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 08:08:14 crc kubenswrapper[4612]: I0227 08:08:14.653828 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ad4524a-e07e-4a6e-93b1-5e5e2d7b251c-kube-api-access-jmqz8" (OuterVolumeSpecName: "kube-api-access-jmqz8") pod "6ad4524a-e07e-4a6e-93b1-5e5e2d7b251c" (UID: "6ad4524a-e07e-4a6e-93b1-5e5e2d7b251c"). InnerVolumeSpecName "kube-api-access-jmqz8". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 08:08:14 crc kubenswrapper[4612]: I0227 08:08:14.747097 4612 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6ad4524a-e07e-4a6e-93b1-5e5e2d7b251c-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 27 08:08:14 crc kubenswrapper[4612]: I0227 08:08:14.747133 4612 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/03d154b1-4114-4c56-82b0-5bf148511e5f-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 27 08:08:14 crc kubenswrapper[4612]: I0227 08:08:14.747144 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v4mtb\" (UniqueName: \"kubernetes.io/projected/03d154b1-4114-4c56-82b0-5bf148511e5f-kube-api-access-v4mtb\") on node \"crc\" DevicePath \"\"" Feb 27 08:08:14 crc kubenswrapper[4612]: I0227 08:08:14.747157 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jmqz8\" (UniqueName: \"kubernetes.io/projected/6ad4524a-e07e-4a6e-93b1-5e5e2d7b251c-kube-api-access-jmqz8\") on node \"crc\" DevicePath \"\"" Feb 27 08:08:14 crc kubenswrapper[4612]: I0227 08:08:14.920120 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-7819-account-create-update-lppdh" event={"ID":"03d154b1-4114-4c56-82b0-5bf148511e5f","Type":"ContainerDied","Data":"2d3a191afc4bc62db37584755712098fc0f1aeedc44e40c56f6156f0fbecbaf1"} Feb 27 08:08:14 crc kubenswrapper[4612]: I0227 08:08:14.920164 4612 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2d3a191afc4bc62db37584755712098fc0f1aeedc44e40c56f6156f0fbecbaf1" Feb 27 08:08:14 crc kubenswrapper[4612]: I0227 08:08:14.920255 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-7819-account-create-update-lppdh" Feb 27 08:08:14 crc kubenswrapper[4612]: I0227 08:08:14.927278 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-3e89-account-create-update-crnl5" Feb 27 08:08:14 crc kubenswrapper[4612]: I0227 08:08:14.927327 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-3e89-account-create-update-crnl5" event={"ID":"6ad4524a-e07e-4a6e-93b1-5e5e2d7b251c","Type":"ContainerDied","Data":"33f9559e847061c7c265ff46b1d08a56e2a04e09539552175f6a1ba35f0dc4ce"} Feb 27 08:08:14 crc kubenswrapper[4612]: I0227 08:08:14.927632 4612 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="33f9559e847061c7c265ff46b1d08a56e2a04e09539552175f6a1ba35f0dc4ce" Feb 27 08:08:15 crc kubenswrapper[4612]: I0227 08:08:15.769645 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/77288860-cf3c-47e6-a752-5b41aa1967fa-etc-swift\") pod \"swift-storage-0\" (UID: \"77288860-cf3c-47e6-a752-5b41aa1967fa\") " pod="openstack/swift-storage-0" Feb 27 08:08:15 crc kubenswrapper[4612]: E0227 08:08:15.769866 4612 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Feb 27 08:08:15 crc kubenswrapper[4612]: E0227 08:08:15.770627 4612 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Feb 27 08:08:15 crc kubenswrapper[4612]: E0227 08:08:15.770712 4612 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/77288860-cf3c-47e6-a752-5b41aa1967fa-etc-swift podName:77288860-cf3c-47e6-a752-5b41aa1967fa nodeName:}" failed. No retries permitted until 2026-02-27 08:08:19.770679297 +0000 UTC m=+1157.624609295 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/77288860-cf3c-47e6-a752-5b41aa1967fa-etc-swift") pod "swift-storage-0" (UID: "77288860-cf3c-47e6-a752-5b41aa1967fa") : configmap "swift-ring-files" not found Feb 27 08:08:16 crc kubenswrapper[4612]: I0227 08:08:16.033959 4612 patch_prober.go:28] interesting pod/machine-config-daemon-924vb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 27 08:08:16 crc kubenswrapper[4612]: I0227 08:08:16.034021 4612 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 27 08:08:16 crc kubenswrapper[4612]: I0227 08:08:16.342543 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/root-account-create-update-6rx64"] Feb 27 08:08:16 crc kubenswrapper[4612]: I0227 08:08:16.356495 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/root-account-create-update-6rx64"] Feb 27 08:08:16 crc kubenswrapper[4612]: I0227 08:08:16.364162 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/root-account-create-update-khpvr"] Feb 27 08:08:16 crc kubenswrapper[4612]: E0227 08:08:16.364527 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ad4524a-e07e-4a6e-93b1-5e5e2d7b251c" containerName="mariadb-account-create-update" Feb 27 08:08:16 crc kubenswrapper[4612]: I0227 08:08:16.364545 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ad4524a-e07e-4a6e-93b1-5e5e2d7b251c" containerName="mariadb-account-create-update" Feb 27 08:08:16 crc kubenswrapper[4612]: E0227 08:08:16.364557 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1551171a-913f-4d99-b4da-148ea454f7f6" containerName="mariadb-database-create" Feb 27 08:08:16 crc kubenswrapper[4612]: I0227 08:08:16.364565 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="1551171a-913f-4d99-b4da-148ea454f7f6" containerName="mariadb-database-create" Feb 27 08:08:16 crc kubenswrapper[4612]: E0227 08:08:16.364583 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="22a1a3d2-6c5d-4102-8fd4-07f953e402f7" containerName="mariadb-database-create" Feb 27 08:08:16 crc kubenswrapper[4612]: I0227 08:08:16.364590 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="22a1a3d2-6c5d-4102-8fd4-07f953e402f7" containerName="mariadb-database-create" Feb 27 08:08:16 crc kubenswrapper[4612]: E0227 08:08:16.364615 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="03d154b1-4114-4c56-82b0-5bf148511e5f" containerName="mariadb-account-create-update" Feb 27 08:08:16 crc kubenswrapper[4612]: I0227 08:08:16.364621 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="03d154b1-4114-4c56-82b0-5bf148511e5f" containerName="mariadb-account-create-update" Feb 27 08:08:16 crc kubenswrapper[4612]: I0227 08:08:16.364777 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="1551171a-913f-4d99-b4da-148ea454f7f6" containerName="mariadb-database-create" Feb 27 08:08:16 crc kubenswrapper[4612]: I0227 08:08:16.364792 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="03d154b1-4114-4c56-82b0-5bf148511e5f" containerName="mariadb-account-create-update" Feb 27 08:08:16 crc kubenswrapper[4612]: I0227 08:08:16.364805 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="22a1a3d2-6c5d-4102-8fd4-07f953e402f7" containerName="mariadb-database-create" Feb 27 08:08:16 crc kubenswrapper[4612]: I0227 08:08:16.364813 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="6ad4524a-e07e-4a6e-93b1-5e5e2d7b251c" containerName="mariadb-account-create-update" Feb 27 08:08:16 crc kubenswrapper[4612]: I0227 08:08:16.365314 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-khpvr" Feb 27 08:08:16 crc kubenswrapper[4612]: I0227 08:08:16.368286 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-mariadb-root-db-secret" Feb 27 08:08:16 crc kubenswrapper[4612]: I0227 08:08:16.391871 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-khpvr"] Feb 27 08:08:16 crc kubenswrapper[4612]: I0227 08:08:16.485767 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f7qtj\" (UniqueName: \"kubernetes.io/projected/998cf80a-2cd6-4869-8967-5b0261ad2dd4-kube-api-access-f7qtj\") pod \"root-account-create-update-khpvr\" (UID: \"998cf80a-2cd6-4869-8967-5b0261ad2dd4\") " pod="openstack/root-account-create-update-khpvr" Feb 27 08:08:16 crc kubenswrapper[4612]: I0227 08:08:16.485893 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/998cf80a-2cd6-4869-8967-5b0261ad2dd4-operator-scripts\") pod \"root-account-create-update-khpvr\" (UID: \"998cf80a-2cd6-4869-8967-5b0261ad2dd4\") " pod="openstack/root-account-create-update-khpvr" Feb 27 08:08:16 crc kubenswrapper[4612]: I0227 08:08:16.587256 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f7qtj\" (UniqueName: \"kubernetes.io/projected/998cf80a-2cd6-4869-8967-5b0261ad2dd4-kube-api-access-f7qtj\") pod \"root-account-create-update-khpvr\" (UID: \"998cf80a-2cd6-4869-8967-5b0261ad2dd4\") " pod="openstack/root-account-create-update-khpvr" Feb 27 08:08:16 crc kubenswrapper[4612]: I0227 08:08:16.587403 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/998cf80a-2cd6-4869-8967-5b0261ad2dd4-operator-scripts\") pod \"root-account-create-update-khpvr\" (UID: \"998cf80a-2cd6-4869-8967-5b0261ad2dd4\") " pod="openstack/root-account-create-update-khpvr" Feb 27 08:08:16 crc kubenswrapper[4612]: I0227 08:08:16.588225 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/998cf80a-2cd6-4869-8967-5b0261ad2dd4-operator-scripts\") pod \"root-account-create-update-khpvr\" (UID: \"998cf80a-2cd6-4869-8967-5b0261ad2dd4\") " pod="openstack/root-account-create-update-khpvr" Feb 27 08:08:16 crc kubenswrapper[4612]: I0227 08:08:16.623505 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f7qtj\" (UniqueName: \"kubernetes.io/projected/998cf80a-2cd6-4869-8967-5b0261ad2dd4-kube-api-access-f7qtj\") pod \"root-account-create-update-khpvr\" (UID: \"998cf80a-2cd6-4869-8967-5b0261ad2dd4\") " pod="openstack/root-account-create-update-khpvr" Feb 27 08:08:16 crc kubenswrapper[4612]: I0227 08:08:16.691007 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-khpvr" Feb 27 08:08:16 crc kubenswrapper[4612]: I0227 08:08:16.870937 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a21d121a-8f63-428c-81b9-cc4d1a52be44" path="/var/lib/kubelet/pods/a21d121a-8f63-428c-81b9-cc4d1a52be44/volumes" Feb 27 08:08:18 crc kubenswrapper[4612]: I0227 08:08:18.761614 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-khpvr"] Feb 27 08:08:18 crc kubenswrapper[4612]: I0227 08:08:18.983364 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-khpvr" event={"ID":"998cf80a-2cd6-4869-8967-5b0261ad2dd4","Type":"ContainerStarted","Data":"4dd551c703fe08d89ae29f3a19ab83cb1dc49449642c22653f097f4a9e22bfe3"} Feb 27 08:08:18 crc kubenswrapper[4612]: I0227 08:08:18.983406 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-khpvr" event={"ID":"998cf80a-2cd6-4869-8967-5b0261ad2dd4","Type":"ContainerStarted","Data":"1ef76b21a8112b8cae3acb6dd5bbb2b928acbe3d1eac16855557d284f1dd611e"} Feb 27 08:08:18 crc kubenswrapper[4612]: I0227 08:08:18.985947 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-gp8mt" event={"ID":"76db835b-ff4b-4d34-9578-38c5e3a605ef","Type":"ContainerStarted","Data":"bea5733769b2fad138ab4910e2ff4a77de3ff34a152523bd650e52ce10b7f5aa"} Feb 27 08:08:19 crc kubenswrapper[4612]: I0227 08:08:19.024506 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/root-account-create-update-khpvr" podStartSLOduration=3.024491501 podStartE2EDuration="3.024491501s" podCreationTimestamp="2026-02-27 08:08:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 08:08:19.001955856 +0000 UTC m=+1156.855885854" watchObservedRunningTime="2026-02-27 08:08:19.024491501 +0000 UTC m=+1156.878421499" Feb 27 08:08:19 crc kubenswrapper[4612]: I0227 08:08:19.025262 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-gp8mt" podStartSLOduration=1.815952016 podStartE2EDuration="7.025254553s" podCreationTimestamp="2026-02-27 08:08:12 +0000 UTC" firstStartedPulling="2026-02-27 08:08:13.113000944 +0000 UTC m=+1150.966930942" lastFinishedPulling="2026-02-27 08:08:18.322303481 +0000 UTC m=+1156.176233479" observedRunningTime="2026-02-27 08:08:19.020763534 +0000 UTC m=+1156.874693532" watchObservedRunningTime="2026-02-27 08:08:19.025254553 +0000 UTC m=+1156.879184551" Feb 27 08:08:19 crc kubenswrapper[4612]: I0227 08:08:19.628390 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Feb 27 08:08:19 crc kubenswrapper[4612]: I0227 08:08:19.848295 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/77288860-cf3c-47e6-a752-5b41aa1967fa-etc-swift\") pod \"swift-storage-0\" (UID: \"77288860-cf3c-47e6-a752-5b41aa1967fa\") " pod="openstack/swift-storage-0" Feb 27 08:08:19 crc kubenswrapper[4612]: E0227 08:08:19.848549 4612 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Feb 27 08:08:19 crc kubenswrapper[4612]: E0227 08:08:19.848964 4612 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Feb 27 08:08:19 crc kubenswrapper[4612]: E0227 08:08:19.849075 4612 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/77288860-cf3c-47e6-a752-5b41aa1967fa-etc-swift podName:77288860-cf3c-47e6-a752-5b41aa1967fa nodeName:}" failed. No retries permitted until 2026-02-27 08:08:27.849050188 +0000 UTC m=+1165.702980186 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/77288860-cf3c-47e6-a752-5b41aa1967fa-etc-swift") pod "swift-storage-0" (UID: "77288860-cf3c-47e6-a752-5b41aa1967fa") : configmap "swift-ring-files" not found Feb 27 08:08:20 crc kubenswrapper[4612]: I0227 08:08:20.003607 4612 generic.go:334] "Generic (PLEG): container finished" podID="998cf80a-2cd6-4869-8967-5b0261ad2dd4" containerID="4dd551c703fe08d89ae29f3a19ab83cb1dc49449642c22653f097f4a9e22bfe3" exitCode=0 Feb 27 08:08:20 crc kubenswrapper[4612]: I0227 08:08:20.003701 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-khpvr" event={"ID":"998cf80a-2cd6-4869-8967-5b0261ad2dd4","Type":"ContainerDied","Data":"4dd551c703fe08d89ae29f3a19ab83cb1dc49449642c22653f097f4a9e22bfe3"} Feb 27 08:08:21 crc kubenswrapper[4612]: I0227 08:08:21.078950 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-b8fbc5445-s2x4c" Feb 27 08:08:21 crc kubenswrapper[4612]: I0227 08:08:21.158098 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8554648995-md72g"] Feb 27 08:08:21 crc kubenswrapper[4612]: I0227 08:08:21.190701 4612 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-8554648995-md72g" podUID="8fc4951a-451c-4194-8776-5e18a3a82479" containerName="dnsmasq-dns" containerID="cri-o://026e77480afb4661e5bdd221bfb4cb9ff8c2fd02e88b87d3293ec2748322f273" gracePeriod=10 Feb 27 08:08:22 crc kubenswrapper[4612]: I0227 08:08:22.022669 4612 generic.go:334] "Generic (PLEG): container finished" podID="8fc4951a-451c-4194-8776-5e18a3a82479" containerID="026e77480afb4661e5bdd221bfb4cb9ff8c2fd02e88b87d3293ec2748322f273" exitCode=0 Feb 27 08:08:22 crc kubenswrapper[4612]: I0227 08:08:22.022734 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-md72g" event={"ID":"8fc4951a-451c-4194-8776-5e18a3a82479","Type":"ContainerDied","Data":"026e77480afb4661e5bdd221bfb4cb9ff8c2fd02e88b87d3293ec2748322f273"} Feb 27 08:08:23 crc kubenswrapper[4612]: I0227 08:08:23.628821 4612 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-8554648995-md72g" podUID="8fc4951a-451c-4194-8776-5e18a3a82479" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.112:5353: connect: connection refused" Feb 27 08:08:24 crc kubenswrapper[4612]: I0227 08:08:24.394648 4612 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-z7f9r" podUID="1ef66a51-46e7-4b76-a70b-b475d47822fd" containerName="ovn-controller" probeResult="failure" output=< Feb 27 08:08:24 crc kubenswrapper[4612]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Feb 27 08:08:24 crc kubenswrapper[4612]: > Feb 27 08:08:24 crc kubenswrapper[4612]: I0227 08:08:24.730067 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-x255x" Feb 27 08:08:24 crc kubenswrapper[4612]: I0227 08:08:24.762531 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-x255x" Feb 27 08:08:24 crc kubenswrapper[4612]: I0227 08:08:24.982982 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-z7f9r-config-sfpd2"] Feb 27 08:08:24 crc kubenswrapper[4612]: I0227 08:08:24.987504 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-z7f9r-config-sfpd2" Feb 27 08:08:24 crc kubenswrapper[4612]: I0227 08:08:24.991824 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Feb 27 08:08:25 crc kubenswrapper[4612]: I0227 08:08:25.007854 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-z7f9r-config-sfpd2"] Feb 27 08:08:25 crc kubenswrapper[4612]: I0227 08:08:25.091430 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/5a1b6420-ccc5-4302-bfd7-be40d868e632-var-run\") pod \"ovn-controller-z7f9r-config-sfpd2\" (UID: \"5a1b6420-ccc5-4302-bfd7-be40d868e632\") " pod="openstack/ovn-controller-z7f9r-config-sfpd2" Feb 27 08:08:25 crc kubenswrapper[4612]: I0227 08:08:25.091486 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vvptn\" (UniqueName: \"kubernetes.io/projected/5a1b6420-ccc5-4302-bfd7-be40d868e632-kube-api-access-vvptn\") pod \"ovn-controller-z7f9r-config-sfpd2\" (UID: \"5a1b6420-ccc5-4302-bfd7-be40d868e632\") " pod="openstack/ovn-controller-z7f9r-config-sfpd2" Feb 27 08:08:25 crc kubenswrapper[4612]: I0227 08:08:25.091540 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/5a1b6420-ccc5-4302-bfd7-be40d868e632-additional-scripts\") pod \"ovn-controller-z7f9r-config-sfpd2\" (UID: \"5a1b6420-ccc5-4302-bfd7-be40d868e632\") " pod="openstack/ovn-controller-z7f9r-config-sfpd2" Feb 27 08:08:25 crc kubenswrapper[4612]: I0227 08:08:25.091619 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/5a1b6420-ccc5-4302-bfd7-be40d868e632-var-run-ovn\") pod \"ovn-controller-z7f9r-config-sfpd2\" (UID: \"5a1b6420-ccc5-4302-bfd7-be40d868e632\") " pod="openstack/ovn-controller-z7f9r-config-sfpd2" Feb 27 08:08:25 crc kubenswrapper[4612]: I0227 08:08:25.091708 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/5a1b6420-ccc5-4302-bfd7-be40d868e632-var-log-ovn\") pod \"ovn-controller-z7f9r-config-sfpd2\" (UID: \"5a1b6420-ccc5-4302-bfd7-be40d868e632\") " pod="openstack/ovn-controller-z7f9r-config-sfpd2" Feb 27 08:08:25 crc kubenswrapper[4612]: I0227 08:08:25.091728 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5a1b6420-ccc5-4302-bfd7-be40d868e632-scripts\") pod \"ovn-controller-z7f9r-config-sfpd2\" (UID: \"5a1b6420-ccc5-4302-bfd7-be40d868e632\") " pod="openstack/ovn-controller-z7f9r-config-sfpd2" Feb 27 08:08:25 crc kubenswrapper[4612]: I0227 08:08:25.192848 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/5a1b6420-ccc5-4302-bfd7-be40d868e632-var-run-ovn\") pod \"ovn-controller-z7f9r-config-sfpd2\" (UID: \"5a1b6420-ccc5-4302-bfd7-be40d868e632\") " pod="openstack/ovn-controller-z7f9r-config-sfpd2" Feb 27 08:08:25 crc kubenswrapper[4612]: I0227 08:08:25.192934 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/5a1b6420-ccc5-4302-bfd7-be40d868e632-var-log-ovn\") pod \"ovn-controller-z7f9r-config-sfpd2\" (UID: \"5a1b6420-ccc5-4302-bfd7-be40d868e632\") " pod="openstack/ovn-controller-z7f9r-config-sfpd2" Feb 27 08:08:25 crc kubenswrapper[4612]: I0227 08:08:25.192953 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5a1b6420-ccc5-4302-bfd7-be40d868e632-scripts\") pod \"ovn-controller-z7f9r-config-sfpd2\" (UID: \"5a1b6420-ccc5-4302-bfd7-be40d868e632\") " pod="openstack/ovn-controller-z7f9r-config-sfpd2" Feb 27 08:08:25 crc kubenswrapper[4612]: I0227 08:08:25.192999 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/5a1b6420-ccc5-4302-bfd7-be40d868e632-var-run\") pod \"ovn-controller-z7f9r-config-sfpd2\" (UID: \"5a1b6420-ccc5-4302-bfd7-be40d868e632\") " pod="openstack/ovn-controller-z7f9r-config-sfpd2" Feb 27 08:08:25 crc kubenswrapper[4612]: I0227 08:08:25.193018 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vvptn\" (UniqueName: \"kubernetes.io/projected/5a1b6420-ccc5-4302-bfd7-be40d868e632-kube-api-access-vvptn\") pod \"ovn-controller-z7f9r-config-sfpd2\" (UID: \"5a1b6420-ccc5-4302-bfd7-be40d868e632\") " pod="openstack/ovn-controller-z7f9r-config-sfpd2" Feb 27 08:08:25 crc kubenswrapper[4612]: I0227 08:08:25.193055 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/5a1b6420-ccc5-4302-bfd7-be40d868e632-additional-scripts\") pod \"ovn-controller-z7f9r-config-sfpd2\" (UID: \"5a1b6420-ccc5-4302-bfd7-be40d868e632\") " pod="openstack/ovn-controller-z7f9r-config-sfpd2" Feb 27 08:08:25 crc kubenswrapper[4612]: I0227 08:08:25.193667 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/5a1b6420-ccc5-4302-bfd7-be40d868e632-additional-scripts\") pod \"ovn-controller-z7f9r-config-sfpd2\" (UID: \"5a1b6420-ccc5-4302-bfd7-be40d868e632\") " pod="openstack/ovn-controller-z7f9r-config-sfpd2" Feb 27 08:08:25 crc kubenswrapper[4612]: I0227 08:08:25.194376 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/5a1b6420-ccc5-4302-bfd7-be40d868e632-var-run-ovn\") pod \"ovn-controller-z7f9r-config-sfpd2\" (UID: \"5a1b6420-ccc5-4302-bfd7-be40d868e632\") " pod="openstack/ovn-controller-z7f9r-config-sfpd2" Feb 27 08:08:25 crc kubenswrapper[4612]: I0227 08:08:25.194429 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/5a1b6420-ccc5-4302-bfd7-be40d868e632-var-log-ovn\") pod \"ovn-controller-z7f9r-config-sfpd2\" (UID: \"5a1b6420-ccc5-4302-bfd7-be40d868e632\") " pod="openstack/ovn-controller-z7f9r-config-sfpd2" Feb 27 08:08:25 crc kubenswrapper[4612]: I0227 08:08:25.195860 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/5a1b6420-ccc5-4302-bfd7-be40d868e632-var-run\") pod \"ovn-controller-z7f9r-config-sfpd2\" (UID: \"5a1b6420-ccc5-4302-bfd7-be40d868e632\") " pod="openstack/ovn-controller-z7f9r-config-sfpd2" Feb 27 08:08:25 crc kubenswrapper[4612]: I0227 08:08:25.196225 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5a1b6420-ccc5-4302-bfd7-be40d868e632-scripts\") pod \"ovn-controller-z7f9r-config-sfpd2\" (UID: \"5a1b6420-ccc5-4302-bfd7-be40d868e632\") " pod="openstack/ovn-controller-z7f9r-config-sfpd2" Feb 27 08:08:25 crc kubenswrapper[4612]: I0227 08:08:25.235602 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vvptn\" (UniqueName: \"kubernetes.io/projected/5a1b6420-ccc5-4302-bfd7-be40d868e632-kube-api-access-vvptn\") pod \"ovn-controller-z7f9r-config-sfpd2\" (UID: \"5a1b6420-ccc5-4302-bfd7-be40d868e632\") " pod="openstack/ovn-controller-z7f9r-config-sfpd2" Feb 27 08:08:25 crc kubenswrapper[4612]: I0227 08:08:25.311682 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-z7f9r-config-sfpd2" Feb 27 08:08:27 crc kubenswrapper[4612]: I0227 08:08:27.074480 4612 generic.go:334] "Generic (PLEG): container finished" podID="76db835b-ff4b-4d34-9578-38c5e3a605ef" containerID="bea5733769b2fad138ab4910e2ff4a77de3ff34a152523bd650e52ce10b7f5aa" exitCode=0 Feb 27 08:08:27 crc kubenswrapper[4612]: I0227 08:08:27.074586 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-gp8mt" event={"ID":"76db835b-ff4b-4d34-9578-38c5e3a605ef","Type":"ContainerDied","Data":"bea5733769b2fad138ab4910e2ff4a77de3ff34a152523bd650e52ce10b7f5aa"} Feb 27 08:08:27 crc kubenswrapper[4612]: I0227 08:08:27.939369 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/77288860-cf3c-47e6-a752-5b41aa1967fa-etc-swift\") pod \"swift-storage-0\" (UID: \"77288860-cf3c-47e6-a752-5b41aa1967fa\") " pod="openstack/swift-storage-0" Feb 27 08:08:27 crc kubenswrapper[4612]: I0227 08:08:27.948423 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/77288860-cf3c-47e6-a752-5b41aa1967fa-etc-swift\") pod \"swift-storage-0\" (UID: \"77288860-cf3c-47e6-a752-5b41aa1967fa\") " pod="openstack/swift-storage-0" Feb 27 08:08:28 crc kubenswrapper[4612]: I0227 08:08:28.061499 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Feb 27 08:08:28 crc kubenswrapper[4612]: I0227 08:08:28.629074 4612 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-8554648995-md72g" podUID="8fc4951a-451c-4194-8776-5e18a3a82479" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.112:5353: connect: connection refused" Feb 27 08:08:28 crc kubenswrapper[4612]: I0227 08:08:28.918050 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-gp8mt" Feb 27 08:08:28 crc kubenswrapper[4612]: I0227 08:08:28.927557 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-khpvr" Feb 27 08:08:28 crc kubenswrapper[4612]: I0227 08:08:28.976650 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/76db835b-ff4b-4d34-9578-38c5e3a605ef-swiftconf\") pod \"76db835b-ff4b-4d34-9578-38c5e3a605ef\" (UID: \"76db835b-ff4b-4d34-9578-38c5e3a605ef\") " Feb 27 08:08:28 crc kubenswrapper[4612]: I0227 08:08:28.976729 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/998cf80a-2cd6-4869-8967-5b0261ad2dd4-operator-scripts\") pod \"998cf80a-2cd6-4869-8967-5b0261ad2dd4\" (UID: \"998cf80a-2cd6-4869-8967-5b0261ad2dd4\") " Feb 27 08:08:28 crc kubenswrapper[4612]: I0227 08:08:28.976797 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/76db835b-ff4b-4d34-9578-38c5e3a605ef-ring-data-devices\") pod \"76db835b-ff4b-4d34-9578-38c5e3a605ef\" (UID: \"76db835b-ff4b-4d34-9578-38c5e3a605ef\") " Feb 27 08:08:28 crc kubenswrapper[4612]: I0227 08:08:28.976845 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/76db835b-ff4b-4d34-9578-38c5e3a605ef-dispersionconf\") pod \"76db835b-ff4b-4d34-9578-38c5e3a605ef\" (UID: \"76db835b-ff4b-4d34-9578-38c5e3a605ef\") " Feb 27 08:08:28 crc kubenswrapper[4612]: I0227 08:08:28.976871 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4g5j5\" (UniqueName: \"kubernetes.io/projected/76db835b-ff4b-4d34-9578-38c5e3a605ef-kube-api-access-4g5j5\") pod \"76db835b-ff4b-4d34-9578-38c5e3a605ef\" (UID: \"76db835b-ff4b-4d34-9578-38c5e3a605ef\") " Feb 27 08:08:28 crc kubenswrapper[4612]: I0227 08:08:28.976955 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/76db835b-ff4b-4d34-9578-38c5e3a605ef-scripts\") pod \"76db835b-ff4b-4d34-9578-38c5e3a605ef\" (UID: \"76db835b-ff4b-4d34-9578-38c5e3a605ef\") " Feb 27 08:08:28 crc kubenswrapper[4612]: I0227 08:08:28.976985 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76db835b-ff4b-4d34-9578-38c5e3a605ef-combined-ca-bundle\") pod \"76db835b-ff4b-4d34-9578-38c5e3a605ef\" (UID: \"76db835b-ff4b-4d34-9578-38c5e3a605ef\") " Feb 27 08:08:28 crc kubenswrapper[4612]: I0227 08:08:28.977016 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f7qtj\" (UniqueName: \"kubernetes.io/projected/998cf80a-2cd6-4869-8967-5b0261ad2dd4-kube-api-access-f7qtj\") pod \"998cf80a-2cd6-4869-8967-5b0261ad2dd4\" (UID: \"998cf80a-2cd6-4869-8967-5b0261ad2dd4\") " Feb 27 08:08:28 crc kubenswrapper[4612]: I0227 08:08:28.977058 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/76db835b-ff4b-4d34-9578-38c5e3a605ef-etc-swift\") pod \"76db835b-ff4b-4d34-9578-38c5e3a605ef\" (UID: \"76db835b-ff4b-4d34-9578-38c5e3a605ef\") " Feb 27 08:08:28 crc kubenswrapper[4612]: I0227 08:08:28.979277 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/76db835b-ff4b-4d34-9578-38c5e3a605ef-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "76db835b-ff4b-4d34-9578-38c5e3a605ef" (UID: "76db835b-ff4b-4d34-9578-38c5e3a605ef"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 08:08:28 crc kubenswrapper[4612]: I0227 08:08:28.979757 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/76db835b-ff4b-4d34-9578-38c5e3a605ef-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "76db835b-ff4b-4d34-9578-38c5e3a605ef" (UID: "76db835b-ff4b-4d34-9578-38c5e3a605ef"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 08:08:28 crc kubenswrapper[4612]: I0227 08:08:28.981371 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/998cf80a-2cd6-4869-8967-5b0261ad2dd4-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "998cf80a-2cd6-4869-8967-5b0261ad2dd4" (UID: "998cf80a-2cd6-4869-8967-5b0261ad2dd4"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 08:08:29 crc kubenswrapper[4612]: I0227 08:08:29.008993 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/998cf80a-2cd6-4869-8967-5b0261ad2dd4-kube-api-access-f7qtj" (OuterVolumeSpecName: "kube-api-access-f7qtj") pod "998cf80a-2cd6-4869-8967-5b0261ad2dd4" (UID: "998cf80a-2cd6-4869-8967-5b0261ad2dd4"). InnerVolumeSpecName "kube-api-access-f7qtj". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 08:08:29 crc kubenswrapper[4612]: I0227 08:08:29.011868 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/76db835b-ff4b-4d34-9578-38c5e3a605ef-kube-api-access-4g5j5" (OuterVolumeSpecName: "kube-api-access-4g5j5") pod "76db835b-ff4b-4d34-9578-38c5e3a605ef" (UID: "76db835b-ff4b-4d34-9578-38c5e3a605ef"). InnerVolumeSpecName "kube-api-access-4g5j5". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 08:08:29 crc kubenswrapper[4612]: I0227 08:08:29.039150 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/76db835b-ff4b-4d34-9578-38c5e3a605ef-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "76db835b-ff4b-4d34-9578-38c5e3a605ef" (UID: "76db835b-ff4b-4d34-9578-38c5e3a605ef"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:08:29 crc kubenswrapper[4612]: I0227 08:08:29.046902 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/76db835b-ff4b-4d34-9578-38c5e3a605ef-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "76db835b-ff4b-4d34-9578-38c5e3a605ef" (UID: "76db835b-ff4b-4d34-9578-38c5e3a605ef"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:08:29 crc kubenswrapper[4612]: I0227 08:08:29.050076 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/76db835b-ff4b-4d34-9578-38c5e3a605ef-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "76db835b-ff4b-4d34-9578-38c5e3a605ef" (UID: "76db835b-ff4b-4d34-9578-38c5e3a605ef"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:08:29 crc kubenswrapper[4612]: I0227 08:08:29.078816 4612 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/998cf80a-2cd6-4869-8967-5b0261ad2dd4-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 27 08:08:29 crc kubenswrapper[4612]: I0227 08:08:29.078842 4612 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/76db835b-ff4b-4d34-9578-38c5e3a605ef-ring-data-devices\") on node \"crc\" DevicePath \"\"" Feb 27 08:08:29 crc kubenswrapper[4612]: I0227 08:08:29.078851 4612 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/76db835b-ff4b-4d34-9578-38c5e3a605ef-dispersionconf\") on node \"crc\" DevicePath \"\"" Feb 27 08:08:29 crc kubenswrapper[4612]: I0227 08:08:29.078867 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4g5j5\" (UniqueName: \"kubernetes.io/projected/76db835b-ff4b-4d34-9578-38c5e3a605ef-kube-api-access-4g5j5\") on node \"crc\" DevicePath \"\"" Feb 27 08:08:29 crc kubenswrapper[4612]: I0227 08:08:29.078878 4612 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76db835b-ff4b-4d34-9578-38c5e3a605ef-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 08:08:29 crc kubenswrapper[4612]: I0227 08:08:29.078886 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f7qtj\" (UniqueName: \"kubernetes.io/projected/998cf80a-2cd6-4869-8967-5b0261ad2dd4-kube-api-access-f7qtj\") on node \"crc\" DevicePath \"\"" Feb 27 08:08:29 crc kubenswrapper[4612]: I0227 08:08:29.078893 4612 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/76db835b-ff4b-4d34-9578-38c5e3a605ef-etc-swift\") on node \"crc\" DevicePath \"\"" Feb 27 08:08:29 crc kubenswrapper[4612]: I0227 08:08:29.078903 4612 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/76db835b-ff4b-4d34-9578-38c5e3a605ef-swiftconf\") on node \"crc\" DevicePath \"\"" Feb 27 08:08:29 crc kubenswrapper[4612]: I0227 08:08:29.096838 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/76db835b-ff4b-4d34-9578-38c5e3a605ef-scripts" (OuterVolumeSpecName: "scripts") pod "76db835b-ff4b-4d34-9578-38c5e3a605ef" (UID: "76db835b-ff4b-4d34-9578-38c5e3a605ef"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 08:08:29 crc kubenswrapper[4612]: I0227 08:08:29.122595 4612 generic.go:334] "Generic (PLEG): container finished" podID="6d1ba6db-1cae-461e-9716-5d5b36350218" containerID="eb1bb5464e5938959380380e0adb9e0444116a9729409966f1310218a5de0a2a" exitCode=0 Feb 27 08:08:29 crc kubenswrapper[4612]: I0227 08:08:29.122681 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"6d1ba6db-1cae-461e-9716-5d5b36350218","Type":"ContainerDied","Data":"eb1bb5464e5938959380380e0adb9e0444116a9729409966f1310218a5de0a2a"} Feb 27 08:08:29 crc kubenswrapper[4612]: I0227 08:08:29.132437 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8554648995-md72g" Feb 27 08:08:29 crc kubenswrapper[4612]: I0227 08:08:29.134065 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-khpvr" Feb 27 08:08:29 crc kubenswrapper[4612]: I0227 08:08:29.134331 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-khpvr" event={"ID":"998cf80a-2cd6-4869-8967-5b0261ad2dd4","Type":"ContainerDied","Data":"1ef76b21a8112b8cae3acb6dd5bbb2b928acbe3d1eac16855557d284f1dd611e"} Feb 27 08:08:29 crc kubenswrapper[4612]: I0227 08:08:29.134371 4612 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1ef76b21a8112b8cae3acb6dd5bbb2b928acbe3d1eac16855557d284f1dd611e" Feb 27 08:08:29 crc kubenswrapper[4612]: I0227 08:08:29.139491 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8554648995-md72g" Feb 27 08:08:29 crc kubenswrapper[4612]: I0227 08:08:29.139863 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-md72g" event={"ID":"8fc4951a-451c-4194-8776-5e18a3a82479","Type":"ContainerDied","Data":"68d164eaa6c2c77917d82649888e5d6e88e9fdad783de0455d9e4c61961c8e35"} Feb 27 08:08:29 crc kubenswrapper[4612]: I0227 08:08:29.139888 4612 scope.go:117] "RemoveContainer" containerID="026e77480afb4661e5bdd221bfb4cb9ff8c2fd02e88b87d3293ec2748322f273" Feb 27 08:08:29 crc kubenswrapper[4612]: I0227 08:08:29.147562 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-gp8mt" event={"ID":"76db835b-ff4b-4d34-9578-38c5e3a605ef","Type":"ContainerDied","Data":"a661d2d8b6154865ea1cc647ecbcb11463e54468b72800b04ed32606de00511f"} Feb 27 08:08:29 crc kubenswrapper[4612]: I0227 08:08:29.147585 4612 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a661d2d8b6154865ea1cc647ecbcb11463e54468b72800b04ed32606de00511f" Feb 27 08:08:29 crc kubenswrapper[4612]: I0227 08:08:29.147604 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-gp8mt" Feb 27 08:08:29 crc kubenswrapper[4612]: I0227 08:08:29.181137 4612 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/76db835b-ff4b-4d34-9578-38c5e3a605ef-scripts\") on node \"crc\" DevicePath \"\"" Feb 27 08:08:29 crc kubenswrapper[4612]: I0227 08:08:29.184876 4612 generic.go:334] "Generic (PLEG): container finished" podID="0da436b1-803d-41bb-94ab-9cfd6fa769eb" containerID="839c0ba56d6f1ff42f16427d18510198097a735f0b3fdf5b535bc49979e9110b" exitCode=0 Feb 27 08:08:29 crc kubenswrapper[4612]: I0227 08:08:29.184945 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"0da436b1-803d-41bb-94ab-9cfd6fa769eb","Type":"ContainerDied","Data":"839c0ba56d6f1ff42f16427d18510198097a735f0b3fdf5b535bc49979e9110b"} Feb 27 08:08:29 crc kubenswrapper[4612]: I0227 08:08:29.245884 4612 scope.go:117] "RemoveContainer" containerID="6520b230948a858cee9959aa97849aff4bac6864f6ea3c48969b03829394403c" Feb 27 08:08:29 crc kubenswrapper[4612]: I0227 08:08:29.282103 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8fc4951a-451c-4194-8776-5e18a3a82479-dns-svc\") pod \"8fc4951a-451c-4194-8776-5e18a3a82479\" (UID: \"8fc4951a-451c-4194-8776-5e18a3a82479\") " Feb 27 08:08:29 crc kubenswrapper[4612]: I0227 08:08:29.282157 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8fc4951a-451c-4194-8776-5e18a3a82479-ovsdbserver-nb\") pod \"8fc4951a-451c-4194-8776-5e18a3a82479\" (UID: \"8fc4951a-451c-4194-8776-5e18a3a82479\") " Feb 27 08:08:29 crc kubenswrapper[4612]: I0227 08:08:29.282195 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8fc4951a-451c-4194-8776-5e18a3a82479-config\") pod \"8fc4951a-451c-4194-8776-5e18a3a82479\" (UID: \"8fc4951a-451c-4194-8776-5e18a3a82479\") " Feb 27 08:08:29 crc kubenswrapper[4612]: I0227 08:08:29.282326 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p7s4d\" (UniqueName: \"kubernetes.io/projected/8fc4951a-451c-4194-8776-5e18a3a82479-kube-api-access-p7s4d\") pod \"8fc4951a-451c-4194-8776-5e18a3a82479\" (UID: \"8fc4951a-451c-4194-8776-5e18a3a82479\") " Feb 27 08:08:29 crc kubenswrapper[4612]: I0227 08:08:29.282351 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8fc4951a-451c-4194-8776-5e18a3a82479-ovsdbserver-sb\") pod \"8fc4951a-451c-4194-8776-5e18a3a82479\" (UID: \"8fc4951a-451c-4194-8776-5e18a3a82479\") " Feb 27 08:08:29 crc kubenswrapper[4612]: I0227 08:08:29.305391 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8fc4951a-451c-4194-8776-5e18a3a82479-kube-api-access-p7s4d" (OuterVolumeSpecName: "kube-api-access-p7s4d") pod "8fc4951a-451c-4194-8776-5e18a3a82479" (UID: "8fc4951a-451c-4194-8776-5e18a3a82479"). InnerVolumeSpecName "kube-api-access-p7s4d". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 08:08:29 crc kubenswrapper[4612]: I0227 08:08:29.384333 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8fc4951a-451c-4194-8776-5e18a3a82479-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "8fc4951a-451c-4194-8776-5e18a3a82479" (UID: "8fc4951a-451c-4194-8776-5e18a3a82479"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 08:08:29 crc kubenswrapper[4612]: I0227 08:08:29.390349 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p7s4d\" (UniqueName: \"kubernetes.io/projected/8fc4951a-451c-4194-8776-5e18a3a82479-kube-api-access-p7s4d\") on node \"crc\" DevicePath \"\"" Feb 27 08:08:29 crc kubenswrapper[4612]: I0227 08:08:29.390379 4612 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8fc4951a-451c-4194-8776-5e18a3a82479-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 27 08:08:29 crc kubenswrapper[4612]: I0227 08:08:29.399860 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-z7f9r-config-sfpd2"] Feb 27 08:08:29 crc kubenswrapper[4612]: I0227 08:08:29.413481 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8fc4951a-451c-4194-8776-5e18a3a82479-config" (OuterVolumeSpecName: "config") pod "8fc4951a-451c-4194-8776-5e18a3a82479" (UID: "8fc4951a-451c-4194-8776-5e18a3a82479"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 08:08:29 crc kubenswrapper[4612]: I0227 08:08:29.421505 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8fc4951a-451c-4194-8776-5e18a3a82479-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "8fc4951a-451c-4194-8776-5e18a3a82479" (UID: "8fc4951a-451c-4194-8776-5e18a3a82479"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 08:08:29 crc kubenswrapper[4612]: I0227 08:08:29.425384 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8fc4951a-451c-4194-8776-5e18a3a82479-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "8fc4951a-451c-4194-8776-5e18a3a82479" (UID: "8fc4951a-451c-4194-8776-5e18a3a82479"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 08:08:29 crc kubenswrapper[4612]: I0227 08:08:29.435159 4612 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-z7f9r" podUID="1ef66a51-46e7-4b76-a70b-b475d47822fd" containerName="ovn-controller" probeResult="failure" output=< Feb 27 08:08:29 crc kubenswrapper[4612]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Feb 27 08:08:29 crc kubenswrapper[4612]: > Feb 27 08:08:29 crc kubenswrapper[4612]: I0227 08:08:29.479596 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8554648995-md72g"] Feb 27 08:08:29 crc kubenswrapper[4612]: I0227 08:08:29.492082 4612 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8fc4951a-451c-4194-8776-5e18a3a82479-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 27 08:08:29 crc kubenswrapper[4612]: I0227 08:08:29.492117 4612 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8fc4951a-451c-4194-8776-5e18a3a82479-config\") on node \"crc\" DevicePath \"\"" Feb 27 08:08:29 crc kubenswrapper[4612]: I0227 08:08:29.492129 4612 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8fc4951a-451c-4194-8776-5e18a3a82479-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 27 08:08:29 crc kubenswrapper[4612]: I0227 08:08:29.494537 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-8554648995-md72g"] Feb 27 08:08:29 crc kubenswrapper[4612]: I0227 08:08:29.696430 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Feb 27 08:08:30 crc kubenswrapper[4612]: I0227 08:08:30.199430 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"0da436b1-803d-41bb-94ab-9cfd6fa769eb","Type":"ContainerStarted","Data":"edd43cdd132a714448da88e522b53d539c8bad75e13f40e8d8ad5319c1936c48"} Feb 27 08:08:30 crc kubenswrapper[4612]: I0227 08:08:30.199931 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Feb 27 08:08:30 crc kubenswrapper[4612]: I0227 08:08:30.214281 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"6d1ba6db-1cae-461e-9716-5d5b36350218","Type":"ContainerStarted","Data":"d4bbac865331908abcaa583aff895c0dfe51bae301a0ff6fa4b2062f92b46c45"} Feb 27 08:08:30 crc kubenswrapper[4612]: I0227 08:08:30.214570 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Feb 27 08:08:30 crc kubenswrapper[4612]: I0227 08:08:30.216234 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-z7f9r-config-sfpd2" event={"ID":"5a1b6420-ccc5-4302-bfd7-be40d868e632","Type":"ContainerStarted","Data":"69d83402d6290299d70278d5befc9ca664f942eed422c9ff4db495ca44b8976c"} Feb 27 08:08:30 crc kubenswrapper[4612]: I0227 08:08:30.216265 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-z7f9r-config-sfpd2" event={"ID":"5a1b6420-ccc5-4302-bfd7-be40d868e632","Type":"ContainerStarted","Data":"8bb8a1ce851680249ce942a6fb236f1e2326555ead8b72bea9f87a14ddf46f99"} Feb 27 08:08:30 crc kubenswrapper[4612]: I0227 08:08:30.218768 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-4ndtf" event={"ID":"ca30e354-dd13-4fa2-96a9-bbd6c605b1f8","Type":"ContainerStarted","Data":"584b11abbdbbc425cc040fab20d9160a8e7f632e67f8118a75cea1518c3225d7"} Feb 27 08:08:30 crc kubenswrapper[4612]: I0227 08:08:30.220912 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"77288860-cf3c-47e6-a752-5b41aa1967fa","Type":"ContainerStarted","Data":"c9298ab372561a85d182910830eeca323ab35cbe59feb3f37e73442f6dc61071"} Feb 27 08:08:30 crc kubenswrapper[4612]: I0227 08:08:30.290915 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=-9223371959.563946 podStartE2EDuration="1m17.290830466s" podCreationTimestamp="2026-02-27 08:07:13 +0000 UTC" firstStartedPulling="2026-02-27 08:07:15.782209871 +0000 UTC m=+1093.636139869" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 08:08:30.246900387 +0000 UTC m=+1168.100830395" watchObservedRunningTime="2026-02-27 08:08:30.290830466 +0000 UTC m=+1168.144760464" Feb 27 08:08:30 crc kubenswrapper[4612]: I0227 08:08:30.301431 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-z7f9r-config-sfpd2" podStartSLOduration=6.301409589 podStartE2EDuration="6.301409589s" podCreationTimestamp="2026-02-27 08:08:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 08:08:30.279970105 +0000 UTC m=+1168.133900103" watchObservedRunningTime="2026-02-27 08:08:30.301409589 +0000 UTC m=+1168.155339587" Feb 27 08:08:30 crc kubenswrapper[4612]: I0227 08:08:30.304994 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-4ndtf" podStartSLOduration=3.267723529 podStartE2EDuration="22.304981112s" podCreationTimestamp="2026-02-27 08:08:08 +0000 UTC" firstStartedPulling="2026-02-27 08:08:09.814020864 +0000 UTC m=+1147.667950862" lastFinishedPulling="2026-02-27 08:08:28.851278447 +0000 UTC m=+1166.705208445" observedRunningTime="2026-02-27 08:08:30.291938828 +0000 UTC m=+1168.145868826" watchObservedRunningTime="2026-02-27 08:08:30.304981112 +0000 UTC m=+1168.158911110" Feb 27 08:08:30 crc kubenswrapper[4612]: I0227 08:08:30.327790 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=37.832461548 podStartE2EDuration="1m17.327772075s" podCreationTimestamp="2026-02-27 08:07:13 +0000 UTC" firstStartedPulling="2026-02-27 08:07:15.927415261 +0000 UTC m=+1093.781345259" lastFinishedPulling="2026-02-27 08:07:55.422725768 +0000 UTC m=+1133.276655786" observedRunningTime="2026-02-27 08:08:30.321949538 +0000 UTC m=+1168.175879526" watchObservedRunningTime="2026-02-27 08:08:30.327772075 +0000 UTC m=+1168.181702073" Feb 27 08:08:30 crc kubenswrapper[4612]: I0227 08:08:30.871130 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8fc4951a-451c-4194-8776-5e18a3a82479" path="/var/lib/kubelet/pods/8fc4951a-451c-4194-8776-5e18a3a82479/volumes" Feb 27 08:08:31 crc kubenswrapper[4612]: I0227 08:08:31.231465 4612 generic.go:334] "Generic (PLEG): container finished" podID="5a1b6420-ccc5-4302-bfd7-be40d868e632" containerID="69d83402d6290299d70278d5befc9ca664f942eed422c9ff4db495ca44b8976c" exitCode=0 Feb 27 08:08:31 crc kubenswrapper[4612]: I0227 08:08:31.231539 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-z7f9r-config-sfpd2" event={"ID":"5a1b6420-ccc5-4302-bfd7-be40d868e632","Type":"ContainerDied","Data":"69d83402d6290299d70278d5befc9ca664f942eed422c9ff4db495ca44b8976c"} Feb 27 08:08:31 crc kubenswrapper[4612]: I0227 08:08:31.234447 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"77288860-cf3c-47e6-a752-5b41aa1967fa","Type":"ContainerStarted","Data":"23cf030d79020072416d8c998839233a3dca4fac89babe1adc670adf3b6531fc"} Feb 27 08:08:31 crc kubenswrapper[4612]: I0227 08:08:31.234491 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"77288860-cf3c-47e6-a752-5b41aa1967fa","Type":"ContainerStarted","Data":"bc447233cc2ec5c9cd952f08b6314742837aa44e18367c4d1f9b5813f7434396"} Feb 27 08:08:32 crc kubenswrapper[4612]: I0227 08:08:32.241779 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"77288860-cf3c-47e6-a752-5b41aa1967fa","Type":"ContainerStarted","Data":"511d9614d3c5bd6296c918c12a8b462f1f946f3b9d383d1adf3f93e1311040ba"} Feb 27 08:08:32 crc kubenswrapper[4612]: I0227 08:08:32.242214 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"77288860-cf3c-47e6-a752-5b41aa1967fa","Type":"ContainerStarted","Data":"b29c11586df5e3f1598093b2e4bf6967ca6fc4b95df0ec63488b6c3c049f92d9"} Feb 27 08:08:32 crc kubenswrapper[4612]: I0227 08:08:32.658615 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-z7f9r-config-sfpd2" Feb 27 08:08:32 crc kubenswrapper[4612]: I0227 08:08:32.757087 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/5a1b6420-ccc5-4302-bfd7-be40d868e632-additional-scripts\") pod \"5a1b6420-ccc5-4302-bfd7-be40d868e632\" (UID: \"5a1b6420-ccc5-4302-bfd7-be40d868e632\") " Feb 27 08:08:32 crc kubenswrapper[4612]: I0227 08:08:32.757146 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vvptn\" (UniqueName: \"kubernetes.io/projected/5a1b6420-ccc5-4302-bfd7-be40d868e632-kube-api-access-vvptn\") pod \"5a1b6420-ccc5-4302-bfd7-be40d868e632\" (UID: \"5a1b6420-ccc5-4302-bfd7-be40d868e632\") " Feb 27 08:08:32 crc kubenswrapper[4612]: I0227 08:08:32.757261 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5a1b6420-ccc5-4302-bfd7-be40d868e632-scripts\") pod \"5a1b6420-ccc5-4302-bfd7-be40d868e632\" (UID: \"5a1b6420-ccc5-4302-bfd7-be40d868e632\") " Feb 27 08:08:32 crc kubenswrapper[4612]: I0227 08:08:32.757303 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/5a1b6420-ccc5-4302-bfd7-be40d868e632-var-log-ovn\") pod \"5a1b6420-ccc5-4302-bfd7-be40d868e632\" (UID: \"5a1b6420-ccc5-4302-bfd7-be40d868e632\") " Feb 27 08:08:32 crc kubenswrapper[4612]: I0227 08:08:32.757413 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/5a1b6420-ccc5-4302-bfd7-be40d868e632-var-run-ovn\") pod \"5a1b6420-ccc5-4302-bfd7-be40d868e632\" (UID: \"5a1b6420-ccc5-4302-bfd7-be40d868e632\") " Feb 27 08:08:32 crc kubenswrapper[4612]: I0227 08:08:32.757508 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/5a1b6420-ccc5-4302-bfd7-be40d868e632-var-run\") pod \"5a1b6420-ccc5-4302-bfd7-be40d868e632\" (UID: \"5a1b6420-ccc5-4302-bfd7-be40d868e632\") " Feb 27 08:08:32 crc kubenswrapper[4612]: I0227 08:08:32.757791 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5a1b6420-ccc5-4302-bfd7-be40d868e632-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "5a1b6420-ccc5-4302-bfd7-be40d868e632" (UID: "5a1b6420-ccc5-4302-bfd7-be40d868e632"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 08:08:32 crc kubenswrapper[4612]: I0227 08:08:32.757826 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5a1b6420-ccc5-4302-bfd7-be40d868e632-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "5a1b6420-ccc5-4302-bfd7-be40d868e632" (UID: "5a1b6420-ccc5-4302-bfd7-be40d868e632"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 27 08:08:32 crc kubenswrapper[4612]: I0227 08:08:32.758084 4612 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/5a1b6420-ccc5-4302-bfd7-be40d868e632-additional-scripts\") on node \"crc\" DevicePath \"\"" Feb 27 08:08:32 crc kubenswrapper[4612]: I0227 08:08:32.758409 4612 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/5a1b6420-ccc5-4302-bfd7-be40d868e632-var-log-ovn\") on node \"crc\" DevicePath \"\"" Feb 27 08:08:32 crc kubenswrapper[4612]: I0227 08:08:32.758308 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5a1b6420-ccc5-4302-bfd7-be40d868e632-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "5a1b6420-ccc5-4302-bfd7-be40d868e632" (UID: "5a1b6420-ccc5-4302-bfd7-be40d868e632"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 27 08:08:32 crc kubenswrapper[4612]: I0227 08:08:32.758332 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5a1b6420-ccc5-4302-bfd7-be40d868e632-var-run" (OuterVolumeSpecName: "var-run") pod "5a1b6420-ccc5-4302-bfd7-be40d868e632" (UID: "5a1b6420-ccc5-4302-bfd7-be40d868e632"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 27 08:08:32 crc kubenswrapper[4612]: I0227 08:08:32.758407 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5a1b6420-ccc5-4302-bfd7-be40d868e632-scripts" (OuterVolumeSpecName: "scripts") pod "5a1b6420-ccc5-4302-bfd7-be40d868e632" (UID: "5a1b6420-ccc5-4302-bfd7-be40d868e632"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 08:08:32 crc kubenswrapper[4612]: I0227 08:08:32.761684 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5a1b6420-ccc5-4302-bfd7-be40d868e632-kube-api-access-vvptn" (OuterVolumeSpecName: "kube-api-access-vvptn") pod "5a1b6420-ccc5-4302-bfd7-be40d868e632" (UID: "5a1b6420-ccc5-4302-bfd7-be40d868e632"). InnerVolumeSpecName "kube-api-access-vvptn". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 08:08:32 crc kubenswrapper[4612]: I0227 08:08:32.866724 4612 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5a1b6420-ccc5-4302-bfd7-be40d868e632-scripts\") on node \"crc\" DevicePath \"\"" Feb 27 08:08:32 crc kubenswrapper[4612]: I0227 08:08:32.866768 4612 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/5a1b6420-ccc5-4302-bfd7-be40d868e632-var-run-ovn\") on node \"crc\" DevicePath \"\"" Feb 27 08:08:32 crc kubenswrapper[4612]: I0227 08:08:32.866782 4612 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/5a1b6420-ccc5-4302-bfd7-be40d868e632-var-run\") on node \"crc\" DevicePath \"\"" Feb 27 08:08:32 crc kubenswrapper[4612]: I0227 08:08:32.866795 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vvptn\" (UniqueName: \"kubernetes.io/projected/5a1b6420-ccc5-4302-bfd7-be40d868e632-kube-api-access-vvptn\") on node \"crc\" DevicePath \"\"" Feb 27 08:08:33 crc kubenswrapper[4612]: I0227 08:08:33.251209 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-z7f9r-config-sfpd2" event={"ID":"5a1b6420-ccc5-4302-bfd7-be40d868e632","Type":"ContainerDied","Data":"8bb8a1ce851680249ce942a6fb236f1e2326555ead8b72bea9f87a14ddf46f99"} Feb 27 08:08:33 crc kubenswrapper[4612]: I0227 08:08:33.251245 4612 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8bb8a1ce851680249ce942a6fb236f1e2326555ead8b72bea9f87a14ddf46f99" Feb 27 08:08:33 crc kubenswrapper[4612]: I0227 08:08:33.251276 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-z7f9r-config-sfpd2" Feb 27 08:08:33 crc kubenswrapper[4612]: I0227 08:08:33.806577 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-z7f9r-config-sfpd2"] Feb 27 08:08:33 crc kubenswrapper[4612]: I0227 08:08:33.815894 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-z7f9r-config-sfpd2"] Feb 27 08:08:34 crc kubenswrapper[4612]: I0227 08:08:34.261507 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"77288860-cf3c-47e6-a752-5b41aa1967fa","Type":"ContainerStarted","Data":"8d0ce5c27cb4d584b29b71a8705a1d85b312dd6cedc2527928b130a2c5239806"} Feb 27 08:08:34 crc kubenswrapper[4612]: I0227 08:08:34.262559 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"77288860-cf3c-47e6-a752-5b41aa1967fa","Type":"ContainerStarted","Data":"54726572b44539b5a423b40af0ec11808ee6c36e81c03c2fc42a3283860769bb"} Feb 27 08:08:34 crc kubenswrapper[4612]: I0227 08:08:34.433079 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-z7f9r" Feb 27 08:08:34 crc kubenswrapper[4612]: I0227 08:08:34.862919 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5a1b6420-ccc5-4302-bfd7-be40d868e632" path="/var/lib/kubelet/pods/5a1b6420-ccc5-4302-bfd7-be40d868e632/volumes" Feb 27 08:08:35 crc kubenswrapper[4612]: I0227 08:08:35.291004 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"77288860-cf3c-47e6-a752-5b41aa1967fa","Type":"ContainerStarted","Data":"621dfb059c4bf203d0986cc7da19dd9408369bc0f2c1eb022f9f13fb7d8f5c7c"} Feb 27 08:08:35 crc kubenswrapper[4612]: I0227 08:08:35.291074 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"77288860-cf3c-47e6-a752-5b41aa1967fa","Type":"ContainerStarted","Data":"ee80ba78752772d31c97be966837e7338ce6afb476d96ac33e676964732dd4f8"} Feb 27 08:08:37 crc kubenswrapper[4612]: I0227 08:08:37.354140 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"77288860-cf3c-47e6-a752-5b41aa1967fa","Type":"ContainerStarted","Data":"e4395eebf74cb2675dff6721e33f6a2db9ff099d6499acd2e266f0c6841bdce4"} Feb 27 08:08:37 crc kubenswrapper[4612]: I0227 08:08:37.354761 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"77288860-cf3c-47e6-a752-5b41aa1967fa","Type":"ContainerStarted","Data":"75fb51421c0f030b11e45daf988e6aaf792259a12bdf47bf8a67ca19edea7255"} Feb 27 08:08:37 crc kubenswrapper[4612]: I0227 08:08:37.354773 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"77288860-cf3c-47e6-a752-5b41aa1967fa","Type":"ContainerStarted","Data":"dc2162195171d189a5261ddf384c391a2d7e55d536321e71909f0e1d832d0b2a"} Feb 27 08:08:37 crc kubenswrapper[4612]: I0227 08:08:37.354784 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"77288860-cf3c-47e6-a752-5b41aa1967fa","Type":"ContainerStarted","Data":"3d559bbcb17964eeb142bb7fdcf0e58a6466d5f14aac74ff0d8e02ed326e2a31"} Feb 27 08:08:37 crc kubenswrapper[4612]: I0227 08:08:37.354793 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"77288860-cf3c-47e6-a752-5b41aa1967fa","Type":"ContainerStarted","Data":"7b0db109ee28784b1afb777f461ea5ad89dd8f0717c7037e075757365da624f1"} Feb 27 08:08:37 crc kubenswrapper[4612]: I0227 08:08:37.354817 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"77288860-cf3c-47e6-a752-5b41aa1967fa","Type":"ContainerStarted","Data":"cac866751e5be79734a229ea5756dadff2b026a1aaf0b389cd2a0d1664eebd13"} Feb 27 08:08:38 crc kubenswrapper[4612]: I0227 08:08:38.368286 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"77288860-cf3c-47e6-a752-5b41aa1967fa","Type":"ContainerStarted","Data":"8533188e52d43625d6db662f30075e701487fab7f0a06377e886d844410d9585"} Feb 27 08:08:38 crc kubenswrapper[4612]: I0227 08:08:38.415265 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-storage-0" podStartSLOduration=21.910107274 podStartE2EDuration="28.415242642s" podCreationTimestamp="2026-02-27 08:08:10 +0000 UTC" firstStartedPulling="2026-02-27 08:08:29.7528337 +0000 UTC m=+1167.606763698" lastFinishedPulling="2026-02-27 08:08:36.257969068 +0000 UTC m=+1174.111899066" observedRunningTime="2026-02-27 08:08:38.410551068 +0000 UTC m=+1176.264481076" watchObservedRunningTime="2026-02-27 08:08:38.415242642 +0000 UTC m=+1176.269172650" Feb 27 08:08:38 crc kubenswrapper[4612]: I0227 08:08:38.693288 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5c79d794d7-rq6t4"] Feb 27 08:08:38 crc kubenswrapper[4612]: E0227 08:08:38.693662 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8fc4951a-451c-4194-8776-5e18a3a82479" containerName="dnsmasq-dns" Feb 27 08:08:38 crc kubenswrapper[4612]: I0227 08:08:38.693682 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="8fc4951a-451c-4194-8776-5e18a3a82479" containerName="dnsmasq-dns" Feb 27 08:08:38 crc kubenswrapper[4612]: E0227 08:08:38.693717 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="998cf80a-2cd6-4869-8967-5b0261ad2dd4" containerName="mariadb-account-create-update" Feb 27 08:08:38 crc kubenswrapper[4612]: I0227 08:08:38.693727 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="998cf80a-2cd6-4869-8967-5b0261ad2dd4" containerName="mariadb-account-create-update" Feb 27 08:08:38 crc kubenswrapper[4612]: E0227 08:08:38.693753 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="76db835b-ff4b-4d34-9578-38c5e3a605ef" containerName="swift-ring-rebalance" Feb 27 08:08:38 crc kubenswrapper[4612]: I0227 08:08:38.693763 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="76db835b-ff4b-4d34-9578-38c5e3a605ef" containerName="swift-ring-rebalance" Feb 27 08:08:38 crc kubenswrapper[4612]: E0227 08:08:38.693777 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5a1b6420-ccc5-4302-bfd7-be40d868e632" containerName="ovn-config" Feb 27 08:08:38 crc kubenswrapper[4612]: I0227 08:08:38.693784 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="5a1b6420-ccc5-4302-bfd7-be40d868e632" containerName="ovn-config" Feb 27 08:08:38 crc kubenswrapper[4612]: E0227 08:08:38.693803 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8fc4951a-451c-4194-8776-5e18a3a82479" containerName="init" Feb 27 08:08:38 crc kubenswrapper[4612]: I0227 08:08:38.693811 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="8fc4951a-451c-4194-8776-5e18a3a82479" containerName="init" Feb 27 08:08:38 crc kubenswrapper[4612]: I0227 08:08:38.693985 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="8fc4951a-451c-4194-8776-5e18a3a82479" containerName="dnsmasq-dns" Feb 27 08:08:38 crc kubenswrapper[4612]: I0227 08:08:38.694007 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="998cf80a-2cd6-4869-8967-5b0261ad2dd4" containerName="mariadb-account-create-update" Feb 27 08:08:38 crc kubenswrapper[4612]: I0227 08:08:38.694028 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="5a1b6420-ccc5-4302-bfd7-be40d868e632" containerName="ovn-config" Feb 27 08:08:38 crc kubenswrapper[4612]: I0227 08:08:38.694043 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="76db835b-ff4b-4d34-9578-38c5e3a605ef" containerName="swift-ring-rebalance" Feb 27 08:08:38 crc kubenswrapper[4612]: I0227 08:08:38.694949 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c79d794d7-rq6t4" Feb 27 08:08:38 crc kubenswrapper[4612]: I0227 08:08:38.699012 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-swift-storage-0" Feb 27 08:08:38 crc kubenswrapper[4612]: I0227 08:08:38.723443 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c79d794d7-rq6t4"] Feb 27 08:08:38 crc kubenswrapper[4612]: I0227 08:08:38.856878 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pbkgm\" (UniqueName: \"kubernetes.io/projected/20c71169-30e9-4861-906e-26395f534c6a-kube-api-access-pbkgm\") pod \"dnsmasq-dns-5c79d794d7-rq6t4\" (UID: \"20c71169-30e9-4861-906e-26395f534c6a\") " pod="openstack/dnsmasq-dns-5c79d794d7-rq6t4" Feb 27 08:08:38 crc kubenswrapper[4612]: I0227 08:08:38.856927 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/20c71169-30e9-4861-906e-26395f534c6a-ovsdbserver-sb\") pod \"dnsmasq-dns-5c79d794d7-rq6t4\" (UID: \"20c71169-30e9-4861-906e-26395f534c6a\") " pod="openstack/dnsmasq-dns-5c79d794d7-rq6t4" Feb 27 08:08:38 crc kubenswrapper[4612]: I0227 08:08:38.856951 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/20c71169-30e9-4861-906e-26395f534c6a-dns-swift-storage-0\") pod \"dnsmasq-dns-5c79d794d7-rq6t4\" (UID: \"20c71169-30e9-4861-906e-26395f534c6a\") " pod="openstack/dnsmasq-dns-5c79d794d7-rq6t4" Feb 27 08:08:38 crc kubenswrapper[4612]: I0227 08:08:38.857048 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/20c71169-30e9-4861-906e-26395f534c6a-config\") pod \"dnsmasq-dns-5c79d794d7-rq6t4\" (UID: \"20c71169-30e9-4861-906e-26395f534c6a\") " pod="openstack/dnsmasq-dns-5c79d794d7-rq6t4" Feb 27 08:08:38 crc kubenswrapper[4612]: I0227 08:08:38.857078 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/20c71169-30e9-4861-906e-26395f534c6a-ovsdbserver-nb\") pod \"dnsmasq-dns-5c79d794d7-rq6t4\" (UID: \"20c71169-30e9-4861-906e-26395f534c6a\") " pod="openstack/dnsmasq-dns-5c79d794d7-rq6t4" Feb 27 08:08:38 crc kubenswrapper[4612]: I0227 08:08:38.857136 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/20c71169-30e9-4861-906e-26395f534c6a-dns-svc\") pod \"dnsmasq-dns-5c79d794d7-rq6t4\" (UID: \"20c71169-30e9-4861-906e-26395f534c6a\") " pod="openstack/dnsmasq-dns-5c79d794d7-rq6t4" Feb 27 08:08:38 crc kubenswrapper[4612]: I0227 08:08:38.959066 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pbkgm\" (UniqueName: \"kubernetes.io/projected/20c71169-30e9-4861-906e-26395f534c6a-kube-api-access-pbkgm\") pod \"dnsmasq-dns-5c79d794d7-rq6t4\" (UID: \"20c71169-30e9-4861-906e-26395f534c6a\") " pod="openstack/dnsmasq-dns-5c79d794d7-rq6t4" Feb 27 08:08:38 crc kubenswrapper[4612]: I0227 08:08:38.959200 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/20c71169-30e9-4861-906e-26395f534c6a-ovsdbserver-sb\") pod \"dnsmasq-dns-5c79d794d7-rq6t4\" (UID: \"20c71169-30e9-4861-906e-26395f534c6a\") " pod="openstack/dnsmasq-dns-5c79d794d7-rq6t4" Feb 27 08:08:38 crc kubenswrapper[4612]: I0227 08:08:38.959241 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/20c71169-30e9-4861-906e-26395f534c6a-dns-swift-storage-0\") pod \"dnsmasq-dns-5c79d794d7-rq6t4\" (UID: \"20c71169-30e9-4861-906e-26395f534c6a\") " pod="openstack/dnsmasq-dns-5c79d794d7-rq6t4" Feb 27 08:08:38 crc kubenswrapper[4612]: I0227 08:08:38.959333 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/20c71169-30e9-4861-906e-26395f534c6a-config\") pod \"dnsmasq-dns-5c79d794d7-rq6t4\" (UID: \"20c71169-30e9-4861-906e-26395f534c6a\") " pod="openstack/dnsmasq-dns-5c79d794d7-rq6t4" Feb 27 08:08:38 crc kubenswrapper[4612]: I0227 08:08:38.959348 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/20c71169-30e9-4861-906e-26395f534c6a-ovsdbserver-nb\") pod \"dnsmasq-dns-5c79d794d7-rq6t4\" (UID: \"20c71169-30e9-4861-906e-26395f534c6a\") " pod="openstack/dnsmasq-dns-5c79d794d7-rq6t4" Feb 27 08:08:38 crc kubenswrapper[4612]: I0227 08:08:38.959408 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/20c71169-30e9-4861-906e-26395f534c6a-dns-svc\") pod \"dnsmasq-dns-5c79d794d7-rq6t4\" (UID: \"20c71169-30e9-4861-906e-26395f534c6a\") " pod="openstack/dnsmasq-dns-5c79d794d7-rq6t4" Feb 27 08:08:38 crc kubenswrapper[4612]: I0227 08:08:38.960414 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/20c71169-30e9-4861-906e-26395f534c6a-dns-svc\") pod \"dnsmasq-dns-5c79d794d7-rq6t4\" (UID: \"20c71169-30e9-4861-906e-26395f534c6a\") " pod="openstack/dnsmasq-dns-5c79d794d7-rq6t4" Feb 27 08:08:38 crc kubenswrapper[4612]: I0227 08:08:38.961655 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/20c71169-30e9-4861-906e-26395f534c6a-config\") pod \"dnsmasq-dns-5c79d794d7-rq6t4\" (UID: \"20c71169-30e9-4861-906e-26395f534c6a\") " pod="openstack/dnsmasq-dns-5c79d794d7-rq6t4" Feb 27 08:08:38 crc kubenswrapper[4612]: I0227 08:08:38.961660 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/20c71169-30e9-4861-906e-26395f534c6a-ovsdbserver-sb\") pod \"dnsmasq-dns-5c79d794d7-rq6t4\" (UID: \"20c71169-30e9-4861-906e-26395f534c6a\") " pod="openstack/dnsmasq-dns-5c79d794d7-rq6t4" Feb 27 08:08:38 crc kubenswrapper[4612]: I0227 08:08:38.962494 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/20c71169-30e9-4861-906e-26395f534c6a-ovsdbserver-nb\") pod \"dnsmasq-dns-5c79d794d7-rq6t4\" (UID: \"20c71169-30e9-4861-906e-26395f534c6a\") " pod="openstack/dnsmasq-dns-5c79d794d7-rq6t4" Feb 27 08:08:38 crc kubenswrapper[4612]: I0227 08:08:38.962548 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/20c71169-30e9-4861-906e-26395f534c6a-dns-swift-storage-0\") pod \"dnsmasq-dns-5c79d794d7-rq6t4\" (UID: \"20c71169-30e9-4861-906e-26395f534c6a\") " pod="openstack/dnsmasq-dns-5c79d794d7-rq6t4" Feb 27 08:08:38 crc kubenswrapper[4612]: I0227 08:08:38.983566 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pbkgm\" (UniqueName: \"kubernetes.io/projected/20c71169-30e9-4861-906e-26395f534c6a-kube-api-access-pbkgm\") pod \"dnsmasq-dns-5c79d794d7-rq6t4\" (UID: \"20c71169-30e9-4861-906e-26395f534c6a\") " pod="openstack/dnsmasq-dns-5c79d794d7-rq6t4" Feb 27 08:08:39 crc kubenswrapper[4612]: I0227 08:08:39.013511 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c79d794d7-rq6t4" Feb 27 08:08:39 crc kubenswrapper[4612]: I0227 08:08:39.311062 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c79d794d7-rq6t4"] Feb 27 08:08:39 crc kubenswrapper[4612]: I0227 08:08:39.378910 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c79d794d7-rq6t4" event={"ID":"20c71169-30e9-4861-906e-26395f534c6a","Type":"ContainerStarted","Data":"a2f9845705aaccb95e0c282721b85b575a3f0bd767fccc54062c804bcf2899d4"} Feb 27 08:08:40 crc kubenswrapper[4612]: E0227 08:08:40.341141 4612 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podca30e354_dd13_4fa2_96a9_bbd6c605b1f8.slice/crio-conmon-584b11abbdbbc425cc040fab20d9160a8e7f632e67f8118a75cea1518c3225d7.scope\": RecentStats: unable to find data in memory cache]" Feb 27 08:08:40 crc kubenswrapper[4612]: I0227 08:08:40.385111 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-4ndtf" event={"ID":"ca30e354-dd13-4fa2-96a9-bbd6c605b1f8","Type":"ContainerDied","Data":"584b11abbdbbc425cc040fab20d9160a8e7f632e67f8118a75cea1518c3225d7"} Feb 27 08:08:40 crc kubenswrapper[4612]: I0227 08:08:40.385115 4612 generic.go:334] "Generic (PLEG): container finished" podID="ca30e354-dd13-4fa2-96a9-bbd6c605b1f8" containerID="584b11abbdbbc425cc040fab20d9160a8e7f632e67f8118a75cea1518c3225d7" exitCode=0 Feb 27 08:08:40 crc kubenswrapper[4612]: I0227 08:08:40.387605 4612 generic.go:334] "Generic (PLEG): container finished" podID="20c71169-30e9-4861-906e-26395f534c6a" containerID="20e3d2fe4c6b3d454a1c8d0d67a999a7a1a166d1d32a5baf50e01a082203a1b0" exitCode=0 Feb 27 08:08:40 crc kubenswrapper[4612]: I0227 08:08:40.387664 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c79d794d7-rq6t4" event={"ID":"20c71169-30e9-4861-906e-26395f534c6a","Type":"ContainerDied","Data":"20e3d2fe4c6b3d454a1c8d0d67a999a7a1a166d1d32a5baf50e01a082203a1b0"} Feb 27 08:08:41 crc kubenswrapper[4612]: I0227 08:08:41.397377 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c79d794d7-rq6t4" event={"ID":"20c71169-30e9-4861-906e-26395f534c6a","Type":"ContainerStarted","Data":"7ca912d4d21cbf15a81d63c2ea95a5f7aec61209dd2e9ec0dbb1ed15c203b85a"} Feb 27 08:08:41 crc kubenswrapper[4612]: I0227 08:08:41.397745 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5c79d794d7-rq6t4" Feb 27 08:08:41 crc kubenswrapper[4612]: I0227 08:08:41.427209 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5c79d794d7-rq6t4" podStartSLOduration=3.4271861169999998 podStartE2EDuration="3.427186117s" podCreationTimestamp="2026-02-27 08:08:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 08:08:41.418678243 +0000 UTC m=+1179.272608251" watchObservedRunningTime="2026-02-27 08:08:41.427186117 +0000 UTC m=+1179.281116125" Feb 27 08:08:41 crc kubenswrapper[4612]: I0227 08:08:41.781891 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-4ndtf" Feb 27 08:08:41 crc kubenswrapper[4612]: I0227 08:08:41.906951 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/ca30e354-dd13-4fa2-96a9-bbd6c605b1f8-db-sync-config-data\") pod \"ca30e354-dd13-4fa2-96a9-bbd6c605b1f8\" (UID: \"ca30e354-dd13-4fa2-96a9-bbd6c605b1f8\") " Feb 27 08:08:41 crc kubenswrapper[4612]: I0227 08:08:41.906997 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ca30e354-dd13-4fa2-96a9-bbd6c605b1f8-config-data\") pod \"ca30e354-dd13-4fa2-96a9-bbd6c605b1f8\" (UID: \"ca30e354-dd13-4fa2-96a9-bbd6c605b1f8\") " Feb 27 08:08:41 crc kubenswrapper[4612]: I0227 08:08:41.907084 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca30e354-dd13-4fa2-96a9-bbd6c605b1f8-combined-ca-bundle\") pod \"ca30e354-dd13-4fa2-96a9-bbd6c605b1f8\" (UID: \"ca30e354-dd13-4fa2-96a9-bbd6c605b1f8\") " Feb 27 08:08:41 crc kubenswrapper[4612]: I0227 08:08:41.907161 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v7fp5\" (UniqueName: \"kubernetes.io/projected/ca30e354-dd13-4fa2-96a9-bbd6c605b1f8-kube-api-access-v7fp5\") pod \"ca30e354-dd13-4fa2-96a9-bbd6c605b1f8\" (UID: \"ca30e354-dd13-4fa2-96a9-bbd6c605b1f8\") " Feb 27 08:08:41 crc kubenswrapper[4612]: I0227 08:08:41.912364 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ca30e354-dd13-4fa2-96a9-bbd6c605b1f8-kube-api-access-v7fp5" (OuterVolumeSpecName: "kube-api-access-v7fp5") pod "ca30e354-dd13-4fa2-96a9-bbd6c605b1f8" (UID: "ca30e354-dd13-4fa2-96a9-bbd6c605b1f8"). InnerVolumeSpecName "kube-api-access-v7fp5". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 08:08:41 crc kubenswrapper[4612]: I0227 08:08:41.919570 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ca30e354-dd13-4fa2-96a9-bbd6c605b1f8-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "ca30e354-dd13-4fa2-96a9-bbd6c605b1f8" (UID: "ca30e354-dd13-4fa2-96a9-bbd6c605b1f8"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:08:41 crc kubenswrapper[4612]: I0227 08:08:41.932571 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ca30e354-dd13-4fa2-96a9-bbd6c605b1f8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ca30e354-dd13-4fa2-96a9-bbd6c605b1f8" (UID: "ca30e354-dd13-4fa2-96a9-bbd6c605b1f8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:08:41 crc kubenswrapper[4612]: I0227 08:08:41.961815 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ca30e354-dd13-4fa2-96a9-bbd6c605b1f8-config-data" (OuterVolumeSpecName: "config-data") pod "ca30e354-dd13-4fa2-96a9-bbd6c605b1f8" (UID: "ca30e354-dd13-4fa2-96a9-bbd6c605b1f8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:08:42 crc kubenswrapper[4612]: I0227 08:08:42.009309 4612 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca30e354-dd13-4fa2-96a9-bbd6c605b1f8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 08:08:42 crc kubenswrapper[4612]: I0227 08:08:42.009344 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v7fp5\" (UniqueName: \"kubernetes.io/projected/ca30e354-dd13-4fa2-96a9-bbd6c605b1f8-kube-api-access-v7fp5\") on node \"crc\" DevicePath \"\"" Feb 27 08:08:42 crc kubenswrapper[4612]: I0227 08:08:42.009355 4612 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/ca30e354-dd13-4fa2-96a9-bbd6c605b1f8-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Feb 27 08:08:42 crc kubenswrapper[4612]: I0227 08:08:42.009365 4612 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ca30e354-dd13-4fa2-96a9-bbd6c605b1f8-config-data\") on node \"crc\" DevicePath \"\"" Feb 27 08:08:42 crc kubenswrapper[4612]: I0227 08:08:42.406408 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-4ndtf" Feb 27 08:08:42 crc kubenswrapper[4612]: I0227 08:08:42.406453 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-4ndtf" event={"ID":"ca30e354-dd13-4fa2-96a9-bbd6c605b1f8","Type":"ContainerDied","Data":"f2ed3c9fe0c68d2c39512e5c74abd7b79f70e6b1c32ce725ee6d8f9735bae323"} Feb 27 08:08:42 crc kubenswrapper[4612]: I0227 08:08:42.407516 4612 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f2ed3c9fe0c68d2c39512e5c74abd7b79f70e6b1c32ce725ee6d8f9735bae323" Feb 27 08:08:42 crc kubenswrapper[4612]: I0227 08:08:42.889042 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c79d794d7-rq6t4"] Feb 27 08:08:42 crc kubenswrapper[4612]: I0227 08:08:42.918013 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5f59b8f679-j8tjt"] Feb 27 08:08:42 crc kubenswrapper[4612]: E0227 08:08:42.918344 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ca30e354-dd13-4fa2-96a9-bbd6c605b1f8" containerName="glance-db-sync" Feb 27 08:08:42 crc kubenswrapper[4612]: I0227 08:08:42.918360 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="ca30e354-dd13-4fa2-96a9-bbd6c605b1f8" containerName="glance-db-sync" Feb 27 08:08:42 crc kubenswrapper[4612]: I0227 08:08:42.918514 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="ca30e354-dd13-4fa2-96a9-bbd6c605b1f8" containerName="glance-db-sync" Feb 27 08:08:42 crc kubenswrapper[4612]: I0227 08:08:42.919523 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5f59b8f679-j8tjt" Feb 27 08:08:42 crc kubenswrapper[4612]: I0227 08:08:42.924233 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/df886311-677c-4060-8bc7-90198481530b-dns-swift-storage-0\") pod \"dnsmasq-dns-5f59b8f679-j8tjt\" (UID: \"df886311-677c-4060-8bc7-90198481530b\") " pod="openstack/dnsmasq-dns-5f59b8f679-j8tjt" Feb 27 08:08:42 crc kubenswrapper[4612]: I0227 08:08:42.924288 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q2kt8\" (UniqueName: \"kubernetes.io/projected/df886311-677c-4060-8bc7-90198481530b-kube-api-access-q2kt8\") pod \"dnsmasq-dns-5f59b8f679-j8tjt\" (UID: \"df886311-677c-4060-8bc7-90198481530b\") " pod="openstack/dnsmasq-dns-5f59b8f679-j8tjt" Feb 27 08:08:42 crc kubenswrapper[4612]: I0227 08:08:42.924347 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/df886311-677c-4060-8bc7-90198481530b-ovsdbserver-nb\") pod \"dnsmasq-dns-5f59b8f679-j8tjt\" (UID: \"df886311-677c-4060-8bc7-90198481530b\") " pod="openstack/dnsmasq-dns-5f59b8f679-j8tjt" Feb 27 08:08:42 crc kubenswrapper[4612]: I0227 08:08:42.924380 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/df886311-677c-4060-8bc7-90198481530b-dns-svc\") pod \"dnsmasq-dns-5f59b8f679-j8tjt\" (UID: \"df886311-677c-4060-8bc7-90198481530b\") " pod="openstack/dnsmasq-dns-5f59b8f679-j8tjt" Feb 27 08:08:42 crc kubenswrapper[4612]: I0227 08:08:42.924418 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/df886311-677c-4060-8bc7-90198481530b-config\") pod \"dnsmasq-dns-5f59b8f679-j8tjt\" (UID: \"df886311-677c-4060-8bc7-90198481530b\") " pod="openstack/dnsmasq-dns-5f59b8f679-j8tjt" Feb 27 08:08:42 crc kubenswrapper[4612]: I0227 08:08:42.924539 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/df886311-677c-4060-8bc7-90198481530b-ovsdbserver-sb\") pod \"dnsmasq-dns-5f59b8f679-j8tjt\" (UID: \"df886311-677c-4060-8bc7-90198481530b\") " pod="openstack/dnsmasq-dns-5f59b8f679-j8tjt" Feb 27 08:08:43 crc kubenswrapper[4612]: I0227 08:08:43.008099 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5f59b8f679-j8tjt"] Feb 27 08:08:43 crc kubenswrapper[4612]: I0227 08:08:43.025514 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/df886311-677c-4060-8bc7-90198481530b-ovsdbserver-nb\") pod \"dnsmasq-dns-5f59b8f679-j8tjt\" (UID: \"df886311-677c-4060-8bc7-90198481530b\") " pod="openstack/dnsmasq-dns-5f59b8f679-j8tjt" Feb 27 08:08:43 crc kubenswrapper[4612]: I0227 08:08:43.025572 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/df886311-677c-4060-8bc7-90198481530b-dns-svc\") pod \"dnsmasq-dns-5f59b8f679-j8tjt\" (UID: \"df886311-677c-4060-8bc7-90198481530b\") " pod="openstack/dnsmasq-dns-5f59b8f679-j8tjt" Feb 27 08:08:43 crc kubenswrapper[4612]: I0227 08:08:43.025612 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/df886311-677c-4060-8bc7-90198481530b-config\") pod \"dnsmasq-dns-5f59b8f679-j8tjt\" (UID: \"df886311-677c-4060-8bc7-90198481530b\") " pod="openstack/dnsmasq-dns-5f59b8f679-j8tjt" Feb 27 08:08:43 crc kubenswrapper[4612]: I0227 08:08:43.025666 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/df886311-677c-4060-8bc7-90198481530b-ovsdbserver-sb\") pod \"dnsmasq-dns-5f59b8f679-j8tjt\" (UID: \"df886311-677c-4060-8bc7-90198481530b\") " pod="openstack/dnsmasq-dns-5f59b8f679-j8tjt" Feb 27 08:08:43 crc kubenswrapper[4612]: I0227 08:08:43.025729 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/df886311-677c-4060-8bc7-90198481530b-dns-swift-storage-0\") pod \"dnsmasq-dns-5f59b8f679-j8tjt\" (UID: \"df886311-677c-4060-8bc7-90198481530b\") " pod="openstack/dnsmasq-dns-5f59b8f679-j8tjt" Feb 27 08:08:43 crc kubenswrapper[4612]: I0227 08:08:43.025748 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q2kt8\" (UniqueName: \"kubernetes.io/projected/df886311-677c-4060-8bc7-90198481530b-kube-api-access-q2kt8\") pod \"dnsmasq-dns-5f59b8f679-j8tjt\" (UID: \"df886311-677c-4060-8bc7-90198481530b\") " pod="openstack/dnsmasq-dns-5f59b8f679-j8tjt" Feb 27 08:08:43 crc kubenswrapper[4612]: I0227 08:08:43.027518 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/df886311-677c-4060-8bc7-90198481530b-ovsdbserver-nb\") pod \"dnsmasq-dns-5f59b8f679-j8tjt\" (UID: \"df886311-677c-4060-8bc7-90198481530b\") " pod="openstack/dnsmasq-dns-5f59b8f679-j8tjt" Feb 27 08:08:43 crc kubenswrapper[4612]: I0227 08:08:43.027635 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/df886311-677c-4060-8bc7-90198481530b-dns-svc\") pod \"dnsmasq-dns-5f59b8f679-j8tjt\" (UID: \"df886311-677c-4060-8bc7-90198481530b\") " pod="openstack/dnsmasq-dns-5f59b8f679-j8tjt" Feb 27 08:08:43 crc kubenswrapper[4612]: I0227 08:08:43.028192 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/df886311-677c-4060-8bc7-90198481530b-dns-swift-storage-0\") pod \"dnsmasq-dns-5f59b8f679-j8tjt\" (UID: \"df886311-677c-4060-8bc7-90198481530b\") " pod="openstack/dnsmasq-dns-5f59b8f679-j8tjt" Feb 27 08:08:43 crc kubenswrapper[4612]: I0227 08:08:43.028278 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/df886311-677c-4060-8bc7-90198481530b-ovsdbserver-sb\") pod \"dnsmasq-dns-5f59b8f679-j8tjt\" (UID: \"df886311-677c-4060-8bc7-90198481530b\") " pod="openstack/dnsmasq-dns-5f59b8f679-j8tjt" Feb 27 08:08:43 crc kubenswrapper[4612]: I0227 08:08:43.028739 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/df886311-677c-4060-8bc7-90198481530b-config\") pod \"dnsmasq-dns-5f59b8f679-j8tjt\" (UID: \"df886311-677c-4060-8bc7-90198481530b\") " pod="openstack/dnsmasq-dns-5f59b8f679-j8tjt" Feb 27 08:08:43 crc kubenswrapper[4612]: I0227 08:08:43.052585 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q2kt8\" (UniqueName: \"kubernetes.io/projected/df886311-677c-4060-8bc7-90198481530b-kube-api-access-q2kt8\") pod \"dnsmasq-dns-5f59b8f679-j8tjt\" (UID: \"df886311-677c-4060-8bc7-90198481530b\") " pod="openstack/dnsmasq-dns-5f59b8f679-j8tjt" Feb 27 08:08:43 crc kubenswrapper[4612]: I0227 08:08:43.237064 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5f59b8f679-j8tjt" Feb 27 08:08:43 crc kubenswrapper[4612]: I0227 08:08:43.413675 4612 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5c79d794d7-rq6t4" podUID="20c71169-30e9-4861-906e-26395f534c6a" containerName="dnsmasq-dns" containerID="cri-o://7ca912d4d21cbf15a81d63c2ea95a5f7aec61209dd2e9ec0dbb1ed15c203b85a" gracePeriod=10 Feb 27 08:08:43 crc kubenswrapper[4612]: I0227 08:08:43.705197 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5f59b8f679-j8tjt"] Feb 27 08:08:43 crc kubenswrapper[4612]: I0227 08:08:43.789977 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c79d794d7-rq6t4" Feb 27 08:08:43 crc kubenswrapper[4612]: I0227 08:08:43.949277 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pbkgm\" (UniqueName: \"kubernetes.io/projected/20c71169-30e9-4861-906e-26395f534c6a-kube-api-access-pbkgm\") pod \"20c71169-30e9-4861-906e-26395f534c6a\" (UID: \"20c71169-30e9-4861-906e-26395f534c6a\") " Feb 27 08:08:43 crc kubenswrapper[4612]: I0227 08:08:43.949348 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/20c71169-30e9-4861-906e-26395f534c6a-ovsdbserver-nb\") pod \"20c71169-30e9-4861-906e-26395f534c6a\" (UID: \"20c71169-30e9-4861-906e-26395f534c6a\") " Feb 27 08:08:43 crc kubenswrapper[4612]: I0227 08:08:43.949388 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/20c71169-30e9-4861-906e-26395f534c6a-ovsdbserver-sb\") pod \"20c71169-30e9-4861-906e-26395f534c6a\" (UID: \"20c71169-30e9-4861-906e-26395f534c6a\") " Feb 27 08:08:43 crc kubenswrapper[4612]: I0227 08:08:43.949429 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/20c71169-30e9-4861-906e-26395f534c6a-dns-svc\") pod \"20c71169-30e9-4861-906e-26395f534c6a\" (UID: \"20c71169-30e9-4861-906e-26395f534c6a\") " Feb 27 08:08:43 crc kubenswrapper[4612]: I0227 08:08:43.949466 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/20c71169-30e9-4861-906e-26395f534c6a-config\") pod \"20c71169-30e9-4861-906e-26395f534c6a\" (UID: \"20c71169-30e9-4861-906e-26395f534c6a\") " Feb 27 08:08:43 crc kubenswrapper[4612]: I0227 08:08:43.949502 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/20c71169-30e9-4861-906e-26395f534c6a-dns-swift-storage-0\") pod \"20c71169-30e9-4861-906e-26395f534c6a\" (UID: \"20c71169-30e9-4861-906e-26395f534c6a\") " Feb 27 08:08:43 crc kubenswrapper[4612]: I0227 08:08:43.954258 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20c71169-30e9-4861-906e-26395f534c6a-kube-api-access-pbkgm" (OuterVolumeSpecName: "kube-api-access-pbkgm") pod "20c71169-30e9-4861-906e-26395f534c6a" (UID: "20c71169-30e9-4861-906e-26395f534c6a"). InnerVolumeSpecName "kube-api-access-pbkgm". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 08:08:43 crc kubenswrapper[4612]: I0227 08:08:43.989548 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/20c71169-30e9-4861-906e-26395f534c6a-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "20c71169-30e9-4861-906e-26395f534c6a" (UID: "20c71169-30e9-4861-906e-26395f534c6a"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 08:08:43 crc kubenswrapper[4612]: I0227 08:08:43.990210 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/20c71169-30e9-4861-906e-26395f534c6a-config" (OuterVolumeSpecName: "config") pod "20c71169-30e9-4861-906e-26395f534c6a" (UID: "20c71169-30e9-4861-906e-26395f534c6a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 08:08:43 crc kubenswrapper[4612]: I0227 08:08:43.991050 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/20c71169-30e9-4861-906e-26395f534c6a-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "20c71169-30e9-4861-906e-26395f534c6a" (UID: "20c71169-30e9-4861-906e-26395f534c6a"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 08:08:43 crc kubenswrapper[4612]: I0227 08:08:43.991685 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/20c71169-30e9-4861-906e-26395f534c6a-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "20c71169-30e9-4861-906e-26395f534c6a" (UID: "20c71169-30e9-4861-906e-26395f534c6a"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 08:08:44 crc kubenswrapper[4612]: I0227 08:08:44.002782 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/20c71169-30e9-4861-906e-26395f534c6a-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "20c71169-30e9-4861-906e-26395f534c6a" (UID: "20c71169-30e9-4861-906e-26395f534c6a"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 08:08:44 crc kubenswrapper[4612]: I0227 08:08:44.051644 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pbkgm\" (UniqueName: \"kubernetes.io/projected/20c71169-30e9-4861-906e-26395f534c6a-kube-api-access-pbkgm\") on node \"crc\" DevicePath \"\"" Feb 27 08:08:44 crc kubenswrapper[4612]: I0227 08:08:44.051672 4612 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/20c71169-30e9-4861-906e-26395f534c6a-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 27 08:08:44 crc kubenswrapper[4612]: I0227 08:08:44.051681 4612 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/20c71169-30e9-4861-906e-26395f534c6a-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 27 08:08:44 crc kubenswrapper[4612]: I0227 08:08:44.051717 4612 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/20c71169-30e9-4861-906e-26395f534c6a-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 27 08:08:44 crc kubenswrapper[4612]: I0227 08:08:44.051728 4612 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/20c71169-30e9-4861-906e-26395f534c6a-config\") on node \"crc\" DevicePath \"\"" Feb 27 08:08:44 crc kubenswrapper[4612]: I0227 08:08:44.051735 4612 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/20c71169-30e9-4861-906e-26395f534c6a-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Feb 27 08:08:44 crc kubenswrapper[4612]: I0227 08:08:44.422498 4612 generic.go:334] "Generic (PLEG): container finished" podID="df886311-677c-4060-8bc7-90198481530b" containerID="1036f4eb8af1f65a4cb1647bbdf944639cd97efc7f52a091258964de332daa16" exitCode=0 Feb 27 08:08:44 crc kubenswrapper[4612]: I0227 08:08:44.422588 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5f59b8f679-j8tjt" event={"ID":"df886311-677c-4060-8bc7-90198481530b","Type":"ContainerDied","Data":"1036f4eb8af1f65a4cb1647bbdf944639cd97efc7f52a091258964de332daa16"} Feb 27 08:08:44 crc kubenswrapper[4612]: I0227 08:08:44.423971 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5f59b8f679-j8tjt" event={"ID":"df886311-677c-4060-8bc7-90198481530b","Type":"ContainerStarted","Data":"a290ea972eddf4b4c16051587218a705c9b71b91918480048ea775ae359439fa"} Feb 27 08:08:44 crc kubenswrapper[4612]: I0227 08:08:44.426484 4612 generic.go:334] "Generic (PLEG): container finished" podID="20c71169-30e9-4861-906e-26395f534c6a" containerID="7ca912d4d21cbf15a81d63c2ea95a5f7aec61209dd2e9ec0dbb1ed15c203b85a" exitCode=0 Feb 27 08:08:44 crc kubenswrapper[4612]: I0227 08:08:44.426521 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c79d794d7-rq6t4" event={"ID":"20c71169-30e9-4861-906e-26395f534c6a","Type":"ContainerDied","Data":"7ca912d4d21cbf15a81d63c2ea95a5f7aec61209dd2e9ec0dbb1ed15c203b85a"} Feb 27 08:08:44 crc kubenswrapper[4612]: I0227 08:08:44.426544 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c79d794d7-rq6t4" event={"ID":"20c71169-30e9-4861-906e-26395f534c6a","Type":"ContainerDied","Data":"a2f9845705aaccb95e0c282721b85b575a3f0bd767fccc54062c804bcf2899d4"} Feb 27 08:08:44 crc kubenswrapper[4612]: I0227 08:08:44.426562 4612 scope.go:117] "RemoveContainer" containerID="7ca912d4d21cbf15a81d63c2ea95a5f7aec61209dd2e9ec0dbb1ed15c203b85a" Feb 27 08:08:44 crc kubenswrapper[4612]: I0227 08:08:44.426666 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c79d794d7-rq6t4" Feb 27 08:08:44 crc kubenswrapper[4612]: I0227 08:08:44.581020 4612 scope.go:117] "RemoveContainer" containerID="20e3d2fe4c6b3d454a1c8d0d67a999a7a1a166d1d32a5baf50e01a082203a1b0" Feb 27 08:08:44 crc kubenswrapper[4612]: I0227 08:08:44.602504 4612 scope.go:117] "RemoveContainer" containerID="7ca912d4d21cbf15a81d63c2ea95a5f7aec61209dd2e9ec0dbb1ed15c203b85a" Feb 27 08:08:44 crc kubenswrapper[4612]: E0227 08:08:44.603071 4612 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7ca912d4d21cbf15a81d63c2ea95a5f7aec61209dd2e9ec0dbb1ed15c203b85a\": container with ID starting with 7ca912d4d21cbf15a81d63c2ea95a5f7aec61209dd2e9ec0dbb1ed15c203b85a not found: ID does not exist" containerID="7ca912d4d21cbf15a81d63c2ea95a5f7aec61209dd2e9ec0dbb1ed15c203b85a" Feb 27 08:08:44 crc kubenswrapper[4612]: I0227 08:08:44.603109 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7ca912d4d21cbf15a81d63c2ea95a5f7aec61209dd2e9ec0dbb1ed15c203b85a"} err="failed to get container status \"7ca912d4d21cbf15a81d63c2ea95a5f7aec61209dd2e9ec0dbb1ed15c203b85a\": rpc error: code = NotFound desc = could not find container \"7ca912d4d21cbf15a81d63c2ea95a5f7aec61209dd2e9ec0dbb1ed15c203b85a\": container with ID starting with 7ca912d4d21cbf15a81d63c2ea95a5f7aec61209dd2e9ec0dbb1ed15c203b85a not found: ID does not exist" Feb 27 08:08:44 crc kubenswrapper[4612]: I0227 08:08:44.603136 4612 scope.go:117] "RemoveContainer" containerID="20e3d2fe4c6b3d454a1c8d0d67a999a7a1a166d1d32a5baf50e01a082203a1b0" Feb 27 08:08:44 crc kubenswrapper[4612]: E0227 08:08:44.603385 4612 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"20e3d2fe4c6b3d454a1c8d0d67a999a7a1a166d1d32a5baf50e01a082203a1b0\": container with ID starting with 20e3d2fe4c6b3d454a1c8d0d67a999a7a1a166d1d32a5baf50e01a082203a1b0 not found: ID does not exist" containerID="20e3d2fe4c6b3d454a1c8d0d67a999a7a1a166d1d32a5baf50e01a082203a1b0" Feb 27 08:08:44 crc kubenswrapper[4612]: I0227 08:08:44.603410 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"20e3d2fe4c6b3d454a1c8d0d67a999a7a1a166d1d32a5baf50e01a082203a1b0"} err="failed to get container status \"20e3d2fe4c6b3d454a1c8d0d67a999a7a1a166d1d32a5baf50e01a082203a1b0\": rpc error: code = NotFound desc = could not find container \"20e3d2fe4c6b3d454a1c8d0d67a999a7a1a166d1d32a5baf50e01a082203a1b0\": container with ID starting with 20e3d2fe4c6b3d454a1c8d0d67a999a7a1a166d1d32a5baf50e01a082203a1b0 not found: ID does not exist" Feb 27 08:08:44 crc kubenswrapper[4612]: I0227 08:08:44.607040 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c79d794d7-rq6t4"] Feb 27 08:08:44 crc kubenswrapper[4612]: I0227 08:08:44.615604 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5c79d794d7-rq6t4"] Feb 27 08:08:44 crc kubenswrapper[4612]: I0227 08:08:44.861309 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20c71169-30e9-4861-906e-26395f534c6a" path="/var/lib/kubelet/pods/20c71169-30e9-4861-906e-26395f534c6a/volumes" Feb 27 08:08:45 crc kubenswrapper[4612]: I0227 08:08:45.192084 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Feb 27 08:08:45 crc kubenswrapper[4612]: I0227 08:08:45.382781 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Feb 27 08:08:45 crc kubenswrapper[4612]: I0227 08:08:45.456711 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5f59b8f679-j8tjt" event={"ID":"df886311-677c-4060-8bc7-90198481530b","Type":"ContainerStarted","Data":"97b0c335965700a7bc21108de21fdffbd2d00631135a6c4d577a8c6f1acbd803"} Feb 27 08:08:45 crc kubenswrapper[4612]: I0227 08:08:45.457575 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5f59b8f679-j8tjt" Feb 27 08:08:45 crc kubenswrapper[4612]: I0227 08:08:45.498250 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5f59b8f679-j8tjt" podStartSLOduration=3.498227138 podStartE2EDuration="3.498227138s" podCreationTimestamp="2026-02-27 08:08:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 08:08:45.480374767 +0000 UTC m=+1183.334304785" watchObservedRunningTime="2026-02-27 08:08:45.498227138 +0000 UTC m=+1183.352157136" Feb 27 08:08:45 crc kubenswrapper[4612]: I0227 08:08:45.553732 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-qg67v"] Feb 27 08:08:45 crc kubenswrapper[4612]: E0227 08:08:45.554075 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="20c71169-30e9-4861-906e-26395f534c6a" containerName="dnsmasq-dns" Feb 27 08:08:45 crc kubenswrapper[4612]: I0227 08:08:45.554090 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="20c71169-30e9-4861-906e-26395f534c6a" containerName="dnsmasq-dns" Feb 27 08:08:45 crc kubenswrapper[4612]: E0227 08:08:45.554124 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="20c71169-30e9-4861-906e-26395f534c6a" containerName="init" Feb 27 08:08:45 crc kubenswrapper[4612]: I0227 08:08:45.554131 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="20c71169-30e9-4861-906e-26395f534c6a" containerName="init" Feb 27 08:08:45 crc kubenswrapper[4612]: I0227 08:08:45.554270 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="20c71169-30e9-4861-906e-26395f534c6a" containerName="dnsmasq-dns" Feb 27 08:08:45 crc kubenswrapper[4612]: I0227 08:08:45.554849 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-qg67v" Feb 27 08:08:45 crc kubenswrapper[4612]: I0227 08:08:45.619401 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-qg67v"] Feb 27 08:08:45 crc kubenswrapper[4612]: I0227 08:08:45.638850 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-3e28-account-create-update-m8rzl"] Feb 27 08:08:45 crc kubenswrapper[4612]: I0227 08:08:45.639810 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-3e28-account-create-update-m8rzl" Feb 27 08:08:45 crc kubenswrapper[4612]: I0227 08:08:45.641891 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Feb 27 08:08:45 crc kubenswrapper[4612]: I0227 08:08:45.654210 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-3e28-account-create-update-m8rzl"] Feb 27 08:08:45 crc kubenswrapper[4612]: I0227 08:08:45.689959 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4e0957b6-31f8-4664-96d9-f75bba20f23f-operator-scripts\") pod \"cinder-db-create-qg67v\" (UID: \"4e0957b6-31f8-4664-96d9-f75bba20f23f\") " pod="openstack/cinder-db-create-qg67v" Feb 27 08:08:45 crc kubenswrapper[4612]: I0227 08:08:45.690039 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jvz92\" (UniqueName: \"kubernetes.io/projected/4e0957b6-31f8-4664-96d9-f75bba20f23f-kube-api-access-jvz92\") pod \"cinder-db-create-qg67v\" (UID: \"4e0957b6-31f8-4664-96d9-f75bba20f23f\") " pod="openstack/cinder-db-create-qg67v" Feb 27 08:08:45 crc kubenswrapper[4612]: I0227 08:08:45.759978 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-845d9"] Feb 27 08:08:45 crc kubenswrapper[4612]: I0227 08:08:45.764090 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-845d9" Feb 27 08:08:45 crc kubenswrapper[4612]: I0227 08:08:45.788123 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-845d9"] Feb 27 08:08:45 crc kubenswrapper[4612]: I0227 08:08:45.791184 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4e0957b6-31f8-4664-96d9-f75bba20f23f-operator-scripts\") pod \"cinder-db-create-qg67v\" (UID: \"4e0957b6-31f8-4664-96d9-f75bba20f23f\") " pod="openstack/cinder-db-create-qg67v" Feb 27 08:08:45 crc kubenswrapper[4612]: I0227 08:08:45.791267 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jvz92\" (UniqueName: \"kubernetes.io/projected/4e0957b6-31f8-4664-96d9-f75bba20f23f-kube-api-access-jvz92\") pod \"cinder-db-create-qg67v\" (UID: \"4e0957b6-31f8-4664-96d9-f75bba20f23f\") " pod="openstack/cinder-db-create-qg67v" Feb 27 08:08:45 crc kubenswrapper[4612]: I0227 08:08:45.791331 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f2cf4\" (UniqueName: \"kubernetes.io/projected/cc36db3f-167a-47e6-807e-11234c86957a-kube-api-access-f2cf4\") pod \"cinder-3e28-account-create-update-m8rzl\" (UID: \"cc36db3f-167a-47e6-807e-11234c86957a\") " pod="openstack/cinder-3e28-account-create-update-m8rzl" Feb 27 08:08:45 crc kubenswrapper[4612]: I0227 08:08:45.791413 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cc36db3f-167a-47e6-807e-11234c86957a-operator-scripts\") pod \"cinder-3e28-account-create-update-m8rzl\" (UID: \"cc36db3f-167a-47e6-807e-11234c86957a\") " pod="openstack/cinder-3e28-account-create-update-m8rzl" Feb 27 08:08:45 crc kubenswrapper[4612]: I0227 08:08:45.791919 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4e0957b6-31f8-4664-96d9-f75bba20f23f-operator-scripts\") pod \"cinder-db-create-qg67v\" (UID: \"4e0957b6-31f8-4664-96d9-f75bba20f23f\") " pod="openstack/cinder-db-create-qg67v" Feb 27 08:08:45 crc kubenswrapper[4612]: I0227 08:08:45.834267 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jvz92\" (UniqueName: \"kubernetes.io/projected/4e0957b6-31f8-4664-96d9-f75bba20f23f-kube-api-access-jvz92\") pod \"cinder-db-create-qg67v\" (UID: \"4e0957b6-31f8-4664-96d9-f75bba20f23f\") " pod="openstack/cinder-db-create-qg67v" Feb 27 08:08:45 crc kubenswrapper[4612]: I0227 08:08:45.878729 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-qg67v" Feb 27 08:08:45 crc kubenswrapper[4612]: I0227 08:08:45.894417 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fb8bfa80-78a7-46e9-acf5-d99fec1cc79c-operator-scripts\") pod \"barbican-db-create-845d9\" (UID: \"fb8bfa80-78a7-46e9-acf5-d99fec1cc79c\") " pod="openstack/barbican-db-create-845d9" Feb 27 08:08:45 crc kubenswrapper[4612]: I0227 08:08:45.894741 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f2cf4\" (UniqueName: \"kubernetes.io/projected/cc36db3f-167a-47e6-807e-11234c86957a-kube-api-access-f2cf4\") pod \"cinder-3e28-account-create-update-m8rzl\" (UID: \"cc36db3f-167a-47e6-807e-11234c86957a\") " pod="openstack/cinder-3e28-account-create-update-m8rzl" Feb 27 08:08:45 crc kubenswrapper[4612]: I0227 08:08:45.894821 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8lhtp\" (UniqueName: \"kubernetes.io/projected/fb8bfa80-78a7-46e9-acf5-d99fec1cc79c-kube-api-access-8lhtp\") pod \"barbican-db-create-845d9\" (UID: \"fb8bfa80-78a7-46e9-acf5-d99fec1cc79c\") " pod="openstack/barbican-db-create-845d9" Feb 27 08:08:45 crc kubenswrapper[4612]: I0227 08:08:45.894845 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cc36db3f-167a-47e6-807e-11234c86957a-operator-scripts\") pod \"cinder-3e28-account-create-update-m8rzl\" (UID: \"cc36db3f-167a-47e6-807e-11234c86957a\") " pod="openstack/cinder-3e28-account-create-update-m8rzl" Feb 27 08:08:45 crc kubenswrapper[4612]: I0227 08:08:45.895561 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cc36db3f-167a-47e6-807e-11234c86957a-operator-scripts\") pod \"cinder-3e28-account-create-update-m8rzl\" (UID: \"cc36db3f-167a-47e6-807e-11234c86957a\") " pod="openstack/cinder-3e28-account-create-update-m8rzl" Feb 27 08:08:45 crc kubenswrapper[4612]: I0227 08:08:45.973795 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-bdfb-account-create-update-8rtvc"] Feb 27 08:08:45 crc kubenswrapper[4612]: I0227 08:08:45.975444 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f2cf4\" (UniqueName: \"kubernetes.io/projected/cc36db3f-167a-47e6-807e-11234c86957a-kube-api-access-f2cf4\") pod \"cinder-3e28-account-create-update-m8rzl\" (UID: \"cc36db3f-167a-47e6-807e-11234c86957a\") " pod="openstack/cinder-3e28-account-create-update-m8rzl" Feb 27 08:08:45 crc kubenswrapper[4612]: I0227 08:08:45.975631 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-bdfb-account-create-update-8rtvc" Feb 27 08:08:45 crc kubenswrapper[4612]: I0227 08:08:45.978090 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Feb 27 08:08:45 crc kubenswrapper[4612]: I0227 08:08:45.985639 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-k4hmr"] Feb 27 08:08:45 crc kubenswrapper[4612]: I0227 08:08:45.986618 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-k4hmr" Feb 27 08:08:46 crc kubenswrapper[4612]: I0227 08:08:45.999574 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8lhtp\" (UniqueName: \"kubernetes.io/projected/fb8bfa80-78a7-46e9-acf5-d99fec1cc79c-kube-api-access-8lhtp\") pod \"barbican-db-create-845d9\" (UID: \"fb8bfa80-78a7-46e9-acf5-d99fec1cc79c\") " pod="openstack/barbican-db-create-845d9" Feb 27 08:08:46 crc kubenswrapper[4612]: I0227 08:08:45.999702 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fb8bfa80-78a7-46e9-acf5-d99fec1cc79c-operator-scripts\") pod \"barbican-db-create-845d9\" (UID: \"fb8bfa80-78a7-46e9-acf5-d99fec1cc79c\") " pod="openstack/barbican-db-create-845d9" Feb 27 08:08:46 crc kubenswrapper[4612]: I0227 08:08:45.999919 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-bdfb-account-create-update-8rtvc"] Feb 27 08:08:46 crc kubenswrapper[4612]: I0227 08:08:46.000433 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fb8bfa80-78a7-46e9-acf5-d99fec1cc79c-operator-scripts\") pod \"barbican-db-create-845d9\" (UID: \"fb8bfa80-78a7-46e9-acf5-d99fec1cc79c\") " pod="openstack/barbican-db-create-845d9" Feb 27 08:08:46 crc kubenswrapper[4612]: I0227 08:08:46.024151 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-k4hmr"] Feb 27 08:08:46 crc kubenswrapper[4612]: I0227 08:08:46.026719 4612 patch_prober.go:28] interesting pod/machine-config-daemon-924vb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 27 08:08:46 crc kubenswrapper[4612]: I0227 08:08:46.026764 4612 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 27 08:08:46 crc kubenswrapper[4612]: I0227 08:08:46.026812 4612 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-924vb" Feb 27 08:08:46 crc kubenswrapper[4612]: I0227 08:08:46.027499 4612 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"8e7aef6fa8ec4b3d8f700070d994aef18691e45f133524f54d3ffb63aa703a66"} pod="openshift-machine-config-operator/machine-config-daemon-924vb" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 27 08:08:46 crc kubenswrapper[4612]: I0227 08:08:46.027567 4612 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" containerName="machine-config-daemon" containerID="cri-o://8e7aef6fa8ec4b3d8f700070d994aef18691e45f133524f54d3ffb63aa703a66" gracePeriod=600 Feb 27 08:08:46 crc kubenswrapper[4612]: I0227 08:08:46.055125 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8lhtp\" (UniqueName: \"kubernetes.io/projected/fb8bfa80-78a7-46e9-acf5-d99fec1cc79c-kube-api-access-8lhtp\") pod \"barbican-db-create-845d9\" (UID: \"fb8bfa80-78a7-46e9-acf5-d99fec1cc79c\") " pod="openstack/barbican-db-create-845d9" Feb 27 08:08:46 crc kubenswrapper[4612]: I0227 08:08:46.088016 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-845d9" Feb 27 08:08:46 crc kubenswrapper[4612]: I0227 08:08:46.109173 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-thhlx\" (UniqueName: \"kubernetes.io/projected/9708a624-a4ef-4250-b3b1-2afb6b56b5c7-kube-api-access-thhlx\") pod \"barbican-bdfb-account-create-update-8rtvc\" (UID: \"9708a624-a4ef-4250-b3b1-2afb6b56b5c7\") " pod="openstack/barbican-bdfb-account-create-update-8rtvc" Feb 27 08:08:46 crc kubenswrapper[4612]: I0227 08:08:46.109248 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9708a624-a4ef-4250-b3b1-2afb6b56b5c7-operator-scripts\") pod \"barbican-bdfb-account-create-update-8rtvc\" (UID: \"9708a624-a4ef-4250-b3b1-2afb6b56b5c7\") " pod="openstack/barbican-bdfb-account-create-update-8rtvc" Feb 27 08:08:46 crc kubenswrapper[4612]: I0227 08:08:46.109276 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c01bd6e3-e867-4506-ab09-75433d751574-operator-scripts\") pod \"neutron-db-create-k4hmr\" (UID: \"c01bd6e3-e867-4506-ab09-75433d751574\") " pod="openstack/neutron-db-create-k4hmr" Feb 27 08:08:46 crc kubenswrapper[4612]: I0227 08:08:46.109300 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xfmkf\" (UniqueName: \"kubernetes.io/projected/c01bd6e3-e867-4506-ab09-75433d751574-kube-api-access-xfmkf\") pod \"neutron-db-create-k4hmr\" (UID: \"c01bd6e3-e867-4506-ab09-75433d751574\") " pod="openstack/neutron-db-create-k4hmr" Feb 27 08:08:46 crc kubenswrapper[4612]: I0227 08:08:46.118424 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-75fb5"] Feb 27 08:08:46 crc kubenswrapper[4612]: I0227 08:08:46.119750 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-75fb5" Feb 27 08:08:46 crc kubenswrapper[4612]: I0227 08:08:46.126217 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Feb 27 08:08:46 crc kubenswrapper[4612]: I0227 08:08:46.126490 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-zdkwc" Feb 27 08:08:46 crc kubenswrapper[4612]: I0227 08:08:46.126619 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Feb 27 08:08:46 crc kubenswrapper[4612]: I0227 08:08:46.126819 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Feb 27 08:08:46 crc kubenswrapper[4612]: I0227 08:08:46.137938 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-75fb5"] Feb 27 08:08:46 crc kubenswrapper[4612]: I0227 08:08:46.210478 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c1e74a83-8d08-428a-b4d7-9a0c60ca4b5e-config-data\") pod \"keystone-db-sync-75fb5\" (UID: \"c1e74a83-8d08-428a-b4d7-9a0c60ca4b5e\") " pod="openstack/keystone-db-sync-75fb5" Feb 27 08:08:46 crc kubenswrapper[4612]: I0227 08:08:46.210558 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-thhlx\" (UniqueName: \"kubernetes.io/projected/9708a624-a4ef-4250-b3b1-2afb6b56b5c7-kube-api-access-thhlx\") pod \"barbican-bdfb-account-create-update-8rtvc\" (UID: \"9708a624-a4ef-4250-b3b1-2afb6b56b5c7\") " pod="openstack/barbican-bdfb-account-create-update-8rtvc" Feb 27 08:08:46 crc kubenswrapper[4612]: I0227 08:08:46.210582 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d6njk\" (UniqueName: \"kubernetes.io/projected/c1e74a83-8d08-428a-b4d7-9a0c60ca4b5e-kube-api-access-d6njk\") pod \"keystone-db-sync-75fb5\" (UID: \"c1e74a83-8d08-428a-b4d7-9a0c60ca4b5e\") " pod="openstack/keystone-db-sync-75fb5" Feb 27 08:08:46 crc kubenswrapper[4612]: I0227 08:08:46.210625 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9708a624-a4ef-4250-b3b1-2afb6b56b5c7-operator-scripts\") pod \"barbican-bdfb-account-create-update-8rtvc\" (UID: \"9708a624-a4ef-4250-b3b1-2afb6b56b5c7\") " pod="openstack/barbican-bdfb-account-create-update-8rtvc" Feb 27 08:08:46 crc kubenswrapper[4612]: I0227 08:08:46.210648 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c1e74a83-8d08-428a-b4d7-9a0c60ca4b5e-combined-ca-bundle\") pod \"keystone-db-sync-75fb5\" (UID: \"c1e74a83-8d08-428a-b4d7-9a0c60ca4b5e\") " pod="openstack/keystone-db-sync-75fb5" Feb 27 08:08:46 crc kubenswrapper[4612]: I0227 08:08:46.210675 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c01bd6e3-e867-4506-ab09-75433d751574-operator-scripts\") pod \"neutron-db-create-k4hmr\" (UID: \"c01bd6e3-e867-4506-ab09-75433d751574\") " pod="openstack/neutron-db-create-k4hmr" Feb 27 08:08:46 crc kubenswrapper[4612]: I0227 08:08:46.210745 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xfmkf\" (UniqueName: \"kubernetes.io/projected/c01bd6e3-e867-4506-ab09-75433d751574-kube-api-access-xfmkf\") pod \"neutron-db-create-k4hmr\" (UID: \"c01bd6e3-e867-4506-ab09-75433d751574\") " pod="openstack/neutron-db-create-k4hmr" Feb 27 08:08:46 crc kubenswrapper[4612]: I0227 08:08:46.211759 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9708a624-a4ef-4250-b3b1-2afb6b56b5c7-operator-scripts\") pod \"barbican-bdfb-account-create-update-8rtvc\" (UID: \"9708a624-a4ef-4250-b3b1-2afb6b56b5c7\") " pod="openstack/barbican-bdfb-account-create-update-8rtvc" Feb 27 08:08:46 crc kubenswrapper[4612]: I0227 08:08:46.211850 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c01bd6e3-e867-4506-ab09-75433d751574-operator-scripts\") pod \"neutron-db-create-k4hmr\" (UID: \"c01bd6e3-e867-4506-ab09-75433d751574\") " pod="openstack/neutron-db-create-k4hmr" Feb 27 08:08:46 crc kubenswrapper[4612]: I0227 08:08:46.238311 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xfmkf\" (UniqueName: \"kubernetes.io/projected/c01bd6e3-e867-4506-ab09-75433d751574-kube-api-access-xfmkf\") pod \"neutron-db-create-k4hmr\" (UID: \"c01bd6e3-e867-4506-ab09-75433d751574\") " pod="openstack/neutron-db-create-k4hmr" Feb 27 08:08:46 crc kubenswrapper[4612]: I0227 08:08:46.259659 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-thhlx\" (UniqueName: \"kubernetes.io/projected/9708a624-a4ef-4250-b3b1-2afb6b56b5c7-kube-api-access-thhlx\") pod \"barbican-bdfb-account-create-update-8rtvc\" (UID: \"9708a624-a4ef-4250-b3b1-2afb6b56b5c7\") " pod="openstack/barbican-bdfb-account-create-update-8rtvc" Feb 27 08:08:46 crc kubenswrapper[4612]: I0227 08:08:46.275883 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-3e28-account-create-update-m8rzl" Feb 27 08:08:46 crc kubenswrapper[4612]: I0227 08:08:46.291841 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-a95b-account-create-update-wb47g"] Feb 27 08:08:46 crc kubenswrapper[4612]: I0227 08:08:46.293161 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-a95b-account-create-update-wb47g" Feb 27 08:08:46 crc kubenswrapper[4612]: I0227 08:08:46.298314 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Feb 27 08:08:46 crc kubenswrapper[4612]: I0227 08:08:46.308151 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-a95b-account-create-update-wb47g"] Feb 27 08:08:46 crc kubenswrapper[4612]: I0227 08:08:46.318765 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c1e74a83-8d08-428a-b4d7-9a0c60ca4b5e-config-data\") pod \"keystone-db-sync-75fb5\" (UID: \"c1e74a83-8d08-428a-b4d7-9a0c60ca4b5e\") " pod="openstack/keystone-db-sync-75fb5" Feb 27 08:08:46 crc kubenswrapper[4612]: I0227 08:08:46.318883 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d6njk\" (UniqueName: \"kubernetes.io/projected/c1e74a83-8d08-428a-b4d7-9a0c60ca4b5e-kube-api-access-d6njk\") pod \"keystone-db-sync-75fb5\" (UID: \"c1e74a83-8d08-428a-b4d7-9a0c60ca4b5e\") " pod="openstack/keystone-db-sync-75fb5" Feb 27 08:08:46 crc kubenswrapper[4612]: I0227 08:08:46.318957 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c1e74a83-8d08-428a-b4d7-9a0c60ca4b5e-combined-ca-bundle\") pod \"keystone-db-sync-75fb5\" (UID: \"c1e74a83-8d08-428a-b4d7-9a0c60ca4b5e\") " pod="openstack/keystone-db-sync-75fb5" Feb 27 08:08:46 crc kubenswrapper[4612]: I0227 08:08:46.330055 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-bdfb-account-create-update-8rtvc" Feb 27 08:08:46 crc kubenswrapper[4612]: I0227 08:08:46.339678 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c1e74a83-8d08-428a-b4d7-9a0c60ca4b5e-config-data\") pod \"keystone-db-sync-75fb5\" (UID: \"c1e74a83-8d08-428a-b4d7-9a0c60ca4b5e\") " pod="openstack/keystone-db-sync-75fb5" Feb 27 08:08:46 crc kubenswrapper[4612]: I0227 08:08:46.353096 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c1e74a83-8d08-428a-b4d7-9a0c60ca4b5e-combined-ca-bundle\") pod \"keystone-db-sync-75fb5\" (UID: \"c1e74a83-8d08-428a-b4d7-9a0c60ca4b5e\") " pod="openstack/keystone-db-sync-75fb5" Feb 27 08:08:46 crc kubenswrapper[4612]: I0227 08:08:46.367478 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d6njk\" (UniqueName: \"kubernetes.io/projected/c1e74a83-8d08-428a-b4d7-9a0c60ca4b5e-kube-api-access-d6njk\") pod \"keystone-db-sync-75fb5\" (UID: \"c1e74a83-8d08-428a-b4d7-9a0c60ca4b5e\") " pod="openstack/keystone-db-sync-75fb5" Feb 27 08:08:46 crc kubenswrapper[4612]: I0227 08:08:46.409889 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-k4hmr" Feb 27 08:08:46 crc kubenswrapper[4612]: I0227 08:08:46.435935 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/04e60e3e-bfd4-4936-be0a-d6688312aeae-operator-scripts\") pod \"neutron-a95b-account-create-update-wb47g\" (UID: \"04e60e3e-bfd4-4936-be0a-d6688312aeae\") " pod="openstack/neutron-a95b-account-create-update-wb47g" Feb 27 08:08:46 crc kubenswrapper[4612]: I0227 08:08:46.436045 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dkw4c\" (UniqueName: \"kubernetes.io/projected/04e60e3e-bfd4-4936-be0a-d6688312aeae-kube-api-access-dkw4c\") pod \"neutron-a95b-account-create-update-wb47g\" (UID: \"04e60e3e-bfd4-4936-be0a-d6688312aeae\") " pod="openstack/neutron-a95b-account-create-update-wb47g" Feb 27 08:08:46 crc kubenswrapper[4612]: I0227 08:08:46.471522 4612 generic.go:334] "Generic (PLEG): container finished" podID="24599373-7adc-4a1b-8bb4-797bf726f43d" containerID="8e7aef6fa8ec4b3d8f700070d994aef18691e45f133524f54d3ffb63aa703a66" exitCode=0 Feb 27 08:08:46 crc kubenswrapper[4612]: I0227 08:08:46.473332 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-924vb" event={"ID":"24599373-7adc-4a1b-8bb4-797bf726f43d","Type":"ContainerDied","Data":"8e7aef6fa8ec4b3d8f700070d994aef18691e45f133524f54d3ffb63aa703a66"} Feb 27 08:08:46 crc kubenswrapper[4612]: I0227 08:08:46.473445 4612 scope.go:117] "RemoveContainer" containerID="3ee7350e6fe9f892458f31227d045e092d0ee2f4312ea6cc953bb5da78fe0fa7" Feb 27 08:08:46 crc kubenswrapper[4612]: I0227 08:08:46.513266 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-75fb5" Feb 27 08:08:46 crc kubenswrapper[4612]: I0227 08:08:46.537038 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dkw4c\" (UniqueName: \"kubernetes.io/projected/04e60e3e-bfd4-4936-be0a-d6688312aeae-kube-api-access-dkw4c\") pod \"neutron-a95b-account-create-update-wb47g\" (UID: \"04e60e3e-bfd4-4936-be0a-d6688312aeae\") " pod="openstack/neutron-a95b-account-create-update-wb47g" Feb 27 08:08:46 crc kubenswrapper[4612]: I0227 08:08:46.537114 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/04e60e3e-bfd4-4936-be0a-d6688312aeae-operator-scripts\") pod \"neutron-a95b-account-create-update-wb47g\" (UID: \"04e60e3e-bfd4-4936-be0a-d6688312aeae\") " pod="openstack/neutron-a95b-account-create-update-wb47g" Feb 27 08:08:46 crc kubenswrapper[4612]: I0227 08:08:46.537850 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/04e60e3e-bfd4-4936-be0a-d6688312aeae-operator-scripts\") pod \"neutron-a95b-account-create-update-wb47g\" (UID: \"04e60e3e-bfd4-4936-be0a-d6688312aeae\") " pod="openstack/neutron-a95b-account-create-update-wb47g" Feb 27 08:08:46 crc kubenswrapper[4612]: I0227 08:08:46.584228 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dkw4c\" (UniqueName: \"kubernetes.io/projected/04e60e3e-bfd4-4936-be0a-d6688312aeae-kube-api-access-dkw4c\") pod \"neutron-a95b-account-create-update-wb47g\" (UID: \"04e60e3e-bfd4-4936-be0a-d6688312aeae\") " pod="openstack/neutron-a95b-account-create-update-wb47g" Feb 27 08:08:46 crc kubenswrapper[4612]: I0227 08:08:46.667742 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-a95b-account-create-update-wb47g" Feb 27 08:08:46 crc kubenswrapper[4612]: I0227 08:08:46.711926 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-qg67v"] Feb 27 08:08:46 crc kubenswrapper[4612]: I0227 08:08:46.885924 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-845d9"] Feb 27 08:08:47 crc kubenswrapper[4612]: I0227 08:08:47.117025 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-3e28-account-create-update-m8rzl"] Feb 27 08:08:47 crc kubenswrapper[4612]: I0227 08:08:47.150760 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-k4hmr"] Feb 27 08:08:47 crc kubenswrapper[4612]: I0227 08:08:47.295492 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-bdfb-account-create-update-8rtvc"] Feb 27 08:08:47 crc kubenswrapper[4612]: W0227 08:08:47.305019 4612 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9708a624_a4ef_4250_b3b1_2afb6b56b5c7.slice/crio-e196ca6b141d543e965095dc69435f395a88143f2c3edc4e9ca830fdbfbf52f5 WatchSource:0}: Error finding container e196ca6b141d543e965095dc69435f395a88143f2c3edc4e9ca830fdbfbf52f5: Status 404 returned error can't find the container with id e196ca6b141d543e965095dc69435f395a88143f2c3edc4e9ca830fdbfbf52f5 Feb 27 08:08:47 crc kubenswrapper[4612]: I0227 08:08:47.387992 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-a95b-account-create-update-wb47g"] Feb 27 08:08:47 crc kubenswrapper[4612]: W0227 08:08:47.414608 4612 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod04e60e3e_bfd4_4936_be0a_d6688312aeae.slice/crio-5556492ee6b7f1b558b99c2b2636d4a60bd98f9daf5c966a40ee9c7e24f83abe WatchSource:0}: Error finding container 5556492ee6b7f1b558b99c2b2636d4a60bd98f9daf5c966a40ee9c7e24f83abe: Status 404 returned error can't find the container with id 5556492ee6b7f1b558b99c2b2636d4a60bd98f9daf5c966a40ee9c7e24f83abe Feb 27 08:08:47 crc kubenswrapper[4612]: I0227 08:08:47.415740 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-75fb5"] Feb 27 08:08:47 crc kubenswrapper[4612]: W0227 08:08:47.437302 4612 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc1e74a83_8d08_428a_b4d7_9a0c60ca4b5e.slice/crio-6698df905dde354fb0d7c48f03b8351a06b8f716f97cd62a3d743779f105facf WatchSource:0}: Error finding container 6698df905dde354fb0d7c48f03b8351a06b8f716f97cd62a3d743779f105facf: Status 404 returned error can't find the container with id 6698df905dde354fb0d7c48f03b8351a06b8f716f97cd62a3d743779f105facf Feb 27 08:08:47 crc kubenswrapper[4612]: I0227 08:08:47.486781 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-924vb" event={"ID":"24599373-7adc-4a1b-8bb4-797bf726f43d","Type":"ContainerStarted","Data":"650e97edc5e74c75f461feb02d9fe0c4cbdce4ff0887e7ec2d2f50e6d2e7c100"} Feb 27 08:08:47 crc kubenswrapper[4612]: I0227 08:08:47.493043 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-qg67v" event={"ID":"4e0957b6-31f8-4664-96d9-f75bba20f23f","Type":"ContainerStarted","Data":"da0aba518068f84f43b535ddebc9b0c49194bc74f0acf0222d57286ffda800fe"} Feb 27 08:08:47 crc kubenswrapper[4612]: I0227 08:08:47.493088 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-qg67v" event={"ID":"4e0957b6-31f8-4664-96d9-f75bba20f23f","Type":"ContainerStarted","Data":"b5b7f33d9639d054f2d74a58fcb1f044cc3b5fcbeb45221f69dfc51d1fd171b9"} Feb 27 08:08:47 crc kubenswrapper[4612]: I0227 08:08:47.496179 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-75fb5" event={"ID":"c1e74a83-8d08-428a-b4d7-9a0c60ca4b5e","Type":"ContainerStarted","Data":"6698df905dde354fb0d7c48f03b8351a06b8f716f97cd62a3d743779f105facf"} Feb 27 08:08:47 crc kubenswrapper[4612]: I0227 08:08:47.500105 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-bdfb-account-create-update-8rtvc" event={"ID":"9708a624-a4ef-4250-b3b1-2afb6b56b5c7","Type":"ContainerStarted","Data":"e196ca6b141d543e965095dc69435f395a88143f2c3edc4e9ca830fdbfbf52f5"} Feb 27 08:08:47 crc kubenswrapper[4612]: I0227 08:08:47.504008 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-3e28-account-create-update-m8rzl" event={"ID":"cc36db3f-167a-47e6-807e-11234c86957a","Type":"ContainerStarted","Data":"974d225208638855885b81b914b3a8e98860170dad308709cd7b3d2dcddb23c0"} Feb 27 08:08:47 crc kubenswrapper[4612]: I0227 08:08:47.517637 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-845d9" event={"ID":"fb8bfa80-78a7-46e9-acf5-d99fec1cc79c","Type":"ContainerStarted","Data":"2b820b796e8e94cbe25ea8ec156015b0693cd025f7c7aef8eeb8a0a9e12ebfea"} Feb 27 08:08:47 crc kubenswrapper[4612]: I0227 08:08:47.517720 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-845d9" event={"ID":"fb8bfa80-78a7-46e9-acf5-d99fec1cc79c","Type":"ContainerStarted","Data":"b2d16b83b3f792cb646335a98c0e348bbb0c6a37485766d1f9f8ea5d889c8be9"} Feb 27 08:08:47 crc kubenswrapper[4612]: I0227 08:08:47.523717 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-k4hmr" event={"ID":"c01bd6e3-e867-4506-ab09-75433d751574","Type":"ContainerStarted","Data":"46119a49dc8274ac1d37c9d5c8969cc8ba3aa2899a62edae07e07efc66f97063"} Feb 27 08:08:47 crc kubenswrapper[4612]: I0227 08:08:47.529581 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-create-qg67v" podStartSLOduration=2.529541073 podStartE2EDuration="2.529541073s" podCreationTimestamp="2026-02-27 08:08:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 08:08:47.524576991 +0000 UTC m=+1185.378506989" watchObservedRunningTime="2026-02-27 08:08:47.529541073 +0000 UTC m=+1185.383471071" Feb 27 08:08:47 crc kubenswrapper[4612]: I0227 08:08:47.536332 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-a95b-account-create-update-wb47g" event={"ID":"04e60e3e-bfd4-4936-be0a-d6688312aeae","Type":"ContainerStarted","Data":"5556492ee6b7f1b558b99c2b2636d4a60bd98f9daf5c966a40ee9c7e24f83abe"} Feb 27 08:08:47 crc kubenswrapper[4612]: I0227 08:08:47.544561 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-create-845d9" podStartSLOduration=2.544546693 podStartE2EDuration="2.544546693s" podCreationTimestamp="2026-02-27 08:08:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 08:08:47.543204845 +0000 UTC m=+1185.397134843" watchObservedRunningTime="2026-02-27 08:08:47.544546693 +0000 UTC m=+1185.398476691" Feb 27 08:08:47 crc kubenswrapper[4612]: I0227 08:08:47.559589 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-create-k4hmr" podStartSLOduration=2.559573684 podStartE2EDuration="2.559573684s" podCreationTimestamp="2026-02-27 08:08:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 08:08:47.554612131 +0000 UTC m=+1185.408542129" watchObservedRunningTime="2026-02-27 08:08:47.559573684 +0000 UTC m=+1185.413503682" Feb 27 08:08:48 crc kubenswrapper[4612]: I0227 08:08:48.548140 4612 generic.go:334] "Generic (PLEG): container finished" podID="c01bd6e3-e867-4506-ab09-75433d751574" containerID="a5a94c230f575a9cf9c7daa80ed04a668e262a0a6dd0b39f379d608f8b63fe86" exitCode=0 Feb 27 08:08:48 crc kubenswrapper[4612]: I0227 08:08:48.548591 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-k4hmr" event={"ID":"c01bd6e3-e867-4506-ab09-75433d751574","Type":"ContainerDied","Data":"a5a94c230f575a9cf9c7daa80ed04a668e262a0a6dd0b39f379d608f8b63fe86"} Feb 27 08:08:48 crc kubenswrapper[4612]: I0227 08:08:48.553675 4612 generic.go:334] "Generic (PLEG): container finished" podID="04e60e3e-bfd4-4936-be0a-d6688312aeae" containerID="8ea58c74730729970fcdaa1932750137c2afe61af32804267541e8e86631b0b8" exitCode=0 Feb 27 08:08:48 crc kubenswrapper[4612]: I0227 08:08:48.553722 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-a95b-account-create-update-wb47g" event={"ID":"04e60e3e-bfd4-4936-be0a-d6688312aeae","Type":"ContainerDied","Data":"8ea58c74730729970fcdaa1932750137c2afe61af32804267541e8e86631b0b8"} Feb 27 08:08:48 crc kubenswrapper[4612]: I0227 08:08:48.556154 4612 generic.go:334] "Generic (PLEG): container finished" podID="4e0957b6-31f8-4664-96d9-f75bba20f23f" containerID="da0aba518068f84f43b535ddebc9b0c49194bc74f0acf0222d57286ffda800fe" exitCode=0 Feb 27 08:08:48 crc kubenswrapper[4612]: I0227 08:08:48.556206 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-qg67v" event={"ID":"4e0957b6-31f8-4664-96d9-f75bba20f23f","Type":"ContainerDied","Data":"da0aba518068f84f43b535ddebc9b0c49194bc74f0acf0222d57286ffda800fe"} Feb 27 08:08:48 crc kubenswrapper[4612]: I0227 08:08:48.565066 4612 generic.go:334] "Generic (PLEG): container finished" podID="9708a624-a4ef-4250-b3b1-2afb6b56b5c7" containerID="ff707c5ca0ead88846d488396fedd3b9b7a434450cb61ed96e59f642dbef8562" exitCode=0 Feb 27 08:08:48 crc kubenswrapper[4612]: I0227 08:08:48.565141 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-bdfb-account-create-update-8rtvc" event={"ID":"9708a624-a4ef-4250-b3b1-2afb6b56b5c7","Type":"ContainerDied","Data":"ff707c5ca0ead88846d488396fedd3b9b7a434450cb61ed96e59f642dbef8562"} Feb 27 08:08:48 crc kubenswrapper[4612]: I0227 08:08:48.569068 4612 generic.go:334] "Generic (PLEG): container finished" podID="cc36db3f-167a-47e6-807e-11234c86957a" containerID="ff6163e0bc97f75223abc25f0842bbda086dc65a63133377d47d41c542503c2f" exitCode=0 Feb 27 08:08:48 crc kubenswrapper[4612]: I0227 08:08:48.569137 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-3e28-account-create-update-m8rzl" event={"ID":"cc36db3f-167a-47e6-807e-11234c86957a","Type":"ContainerDied","Data":"ff6163e0bc97f75223abc25f0842bbda086dc65a63133377d47d41c542503c2f"} Feb 27 08:08:48 crc kubenswrapper[4612]: I0227 08:08:48.570407 4612 generic.go:334] "Generic (PLEG): container finished" podID="fb8bfa80-78a7-46e9-acf5-d99fec1cc79c" containerID="2b820b796e8e94cbe25ea8ec156015b0693cd025f7c7aef8eeb8a0a9e12ebfea" exitCode=0 Feb 27 08:08:48 crc kubenswrapper[4612]: I0227 08:08:48.571297 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-845d9" event={"ID":"fb8bfa80-78a7-46e9-acf5-d99fec1cc79c","Type":"ContainerDied","Data":"2b820b796e8e94cbe25ea8ec156015b0693cd025f7c7aef8eeb8a0a9e12ebfea"} Feb 27 08:08:52 crc kubenswrapper[4612]: I0227 08:08:52.504570 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-3e28-account-create-update-m8rzl" Feb 27 08:08:52 crc kubenswrapper[4612]: I0227 08:08:52.513234 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-a95b-account-create-update-wb47g" Feb 27 08:08:52 crc kubenswrapper[4612]: I0227 08:08:52.548014 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-845d9" Feb 27 08:08:52 crc kubenswrapper[4612]: I0227 08:08:52.569583 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-k4hmr" Feb 27 08:08:52 crc kubenswrapper[4612]: I0227 08:08:52.575479 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-bdfb-account-create-update-8rtvc" Feb 27 08:08:52 crc kubenswrapper[4612]: I0227 08:08:52.589541 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-qg67v" Feb 27 08:08:52 crc kubenswrapper[4612]: I0227 08:08:52.619262 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-k4hmr" event={"ID":"c01bd6e3-e867-4506-ab09-75433d751574","Type":"ContainerDied","Data":"46119a49dc8274ac1d37c9d5c8969cc8ba3aa2899a62edae07e07efc66f97063"} Feb 27 08:08:52 crc kubenswrapper[4612]: I0227 08:08:52.619450 4612 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="46119a49dc8274ac1d37c9d5c8969cc8ba3aa2899a62edae07e07efc66f97063" Feb 27 08:08:52 crc kubenswrapper[4612]: I0227 08:08:52.619553 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-k4hmr" Feb 27 08:08:52 crc kubenswrapper[4612]: I0227 08:08:52.621258 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-a95b-account-create-update-wb47g" event={"ID":"04e60e3e-bfd4-4936-be0a-d6688312aeae","Type":"ContainerDied","Data":"5556492ee6b7f1b558b99c2b2636d4a60bd98f9daf5c966a40ee9c7e24f83abe"} Feb 27 08:08:52 crc kubenswrapper[4612]: I0227 08:08:52.621349 4612 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5556492ee6b7f1b558b99c2b2636d4a60bd98f9daf5c966a40ee9c7e24f83abe" Feb 27 08:08:52 crc kubenswrapper[4612]: I0227 08:08:52.621317 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-a95b-account-create-update-wb47g" Feb 27 08:08:52 crc kubenswrapper[4612]: I0227 08:08:52.622610 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-qg67v" event={"ID":"4e0957b6-31f8-4664-96d9-f75bba20f23f","Type":"ContainerDied","Data":"b5b7f33d9639d054f2d74a58fcb1f044cc3b5fcbeb45221f69dfc51d1fd171b9"} Feb 27 08:08:52 crc kubenswrapper[4612]: I0227 08:08:52.622713 4612 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b5b7f33d9639d054f2d74a58fcb1f044cc3b5fcbeb45221f69dfc51d1fd171b9" Feb 27 08:08:52 crc kubenswrapper[4612]: I0227 08:08:52.622845 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-qg67v" Feb 27 08:08:52 crc kubenswrapper[4612]: I0227 08:08:52.628990 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-bdfb-account-create-update-8rtvc" event={"ID":"9708a624-a4ef-4250-b3b1-2afb6b56b5c7","Type":"ContainerDied","Data":"e196ca6b141d543e965095dc69435f395a88143f2c3edc4e9ca830fdbfbf52f5"} Feb 27 08:08:52 crc kubenswrapper[4612]: I0227 08:08:52.629023 4612 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e196ca6b141d543e965095dc69435f395a88143f2c3edc4e9ca830fdbfbf52f5" Feb 27 08:08:52 crc kubenswrapper[4612]: I0227 08:08:52.629077 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-bdfb-account-create-update-8rtvc" Feb 27 08:08:52 crc kubenswrapper[4612]: I0227 08:08:52.630907 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-3e28-account-create-update-m8rzl" event={"ID":"cc36db3f-167a-47e6-807e-11234c86957a","Type":"ContainerDied","Data":"974d225208638855885b81b914b3a8e98860170dad308709cd7b3d2dcddb23c0"} Feb 27 08:08:52 crc kubenswrapper[4612]: I0227 08:08:52.630929 4612 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="974d225208638855885b81b914b3a8e98860170dad308709cd7b3d2dcddb23c0" Feb 27 08:08:52 crc kubenswrapper[4612]: I0227 08:08:52.630968 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-3e28-account-create-update-m8rzl" Feb 27 08:08:52 crc kubenswrapper[4612]: I0227 08:08:52.632865 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-845d9" event={"ID":"fb8bfa80-78a7-46e9-acf5-d99fec1cc79c","Type":"ContainerDied","Data":"b2d16b83b3f792cb646335a98c0e348bbb0c6a37485766d1f9f8ea5d889c8be9"} Feb 27 08:08:52 crc kubenswrapper[4612]: I0227 08:08:52.632915 4612 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b2d16b83b3f792cb646335a98c0e348bbb0c6a37485766d1f9f8ea5d889c8be9" Feb 27 08:08:52 crc kubenswrapper[4612]: I0227 08:08:52.632996 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-845d9" Feb 27 08:08:52 crc kubenswrapper[4612]: I0227 08:08:52.653508 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cc36db3f-167a-47e6-807e-11234c86957a-operator-scripts\") pod \"cc36db3f-167a-47e6-807e-11234c86957a\" (UID: \"cc36db3f-167a-47e6-807e-11234c86957a\") " Feb 27 08:08:52 crc kubenswrapper[4612]: I0227 08:08:52.653557 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c01bd6e3-e867-4506-ab09-75433d751574-operator-scripts\") pod \"c01bd6e3-e867-4506-ab09-75433d751574\" (UID: \"c01bd6e3-e867-4506-ab09-75433d751574\") " Feb 27 08:08:52 crc kubenswrapper[4612]: I0227 08:08:52.653640 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4e0957b6-31f8-4664-96d9-f75bba20f23f-operator-scripts\") pod \"4e0957b6-31f8-4664-96d9-f75bba20f23f\" (UID: \"4e0957b6-31f8-4664-96d9-f75bba20f23f\") " Feb 27 08:08:52 crc kubenswrapper[4612]: I0227 08:08:52.653718 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dkw4c\" (UniqueName: \"kubernetes.io/projected/04e60e3e-bfd4-4936-be0a-d6688312aeae-kube-api-access-dkw4c\") pod \"04e60e3e-bfd4-4936-be0a-d6688312aeae\" (UID: \"04e60e3e-bfd4-4936-be0a-d6688312aeae\") " Feb 27 08:08:52 crc kubenswrapper[4612]: I0227 08:08:52.653756 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xfmkf\" (UniqueName: \"kubernetes.io/projected/c01bd6e3-e867-4506-ab09-75433d751574-kube-api-access-xfmkf\") pod \"c01bd6e3-e867-4506-ab09-75433d751574\" (UID: \"c01bd6e3-e867-4506-ab09-75433d751574\") " Feb 27 08:08:52 crc kubenswrapper[4612]: I0227 08:08:52.653788 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fb8bfa80-78a7-46e9-acf5-d99fec1cc79c-operator-scripts\") pod \"fb8bfa80-78a7-46e9-acf5-d99fec1cc79c\" (UID: \"fb8bfa80-78a7-46e9-acf5-d99fec1cc79c\") " Feb 27 08:08:52 crc kubenswrapper[4612]: I0227 08:08:52.653808 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9708a624-a4ef-4250-b3b1-2afb6b56b5c7-operator-scripts\") pod \"9708a624-a4ef-4250-b3b1-2afb6b56b5c7\" (UID: \"9708a624-a4ef-4250-b3b1-2afb6b56b5c7\") " Feb 27 08:08:52 crc kubenswrapper[4612]: I0227 08:08:52.653849 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f2cf4\" (UniqueName: \"kubernetes.io/projected/cc36db3f-167a-47e6-807e-11234c86957a-kube-api-access-f2cf4\") pod \"cc36db3f-167a-47e6-807e-11234c86957a\" (UID: \"cc36db3f-167a-47e6-807e-11234c86957a\") " Feb 27 08:08:52 crc kubenswrapper[4612]: I0227 08:08:52.653894 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-thhlx\" (UniqueName: \"kubernetes.io/projected/9708a624-a4ef-4250-b3b1-2afb6b56b5c7-kube-api-access-thhlx\") pod \"9708a624-a4ef-4250-b3b1-2afb6b56b5c7\" (UID: \"9708a624-a4ef-4250-b3b1-2afb6b56b5c7\") " Feb 27 08:08:52 crc kubenswrapper[4612]: I0227 08:08:52.653923 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/04e60e3e-bfd4-4936-be0a-d6688312aeae-operator-scripts\") pod \"04e60e3e-bfd4-4936-be0a-d6688312aeae\" (UID: \"04e60e3e-bfd4-4936-be0a-d6688312aeae\") " Feb 27 08:08:52 crc kubenswrapper[4612]: I0227 08:08:52.653944 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jvz92\" (UniqueName: \"kubernetes.io/projected/4e0957b6-31f8-4664-96d9-f75bba20f23f-kube-api-access-jvz92\") pod \"4e0957b6-31f8-4664-96d9-f75bba20f23f\" (UID: \"4e0957b6-31f8-4664-96d9-f75bba20f23f\") " Feb 27 08:08:52 crc kubenswrapper[4612]: I0227 08:08:52.653960 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8lhtp\" (UniqueName: \"kubernetes.io/projected/fb8bfa80-78a7-46e9-acf5-d99fec1cc79c-kube-api-access-8lhtp\") pod \"fb8bfa80-78a7-46e9-acf5-d99fec1cc79c\" (UID: \"fb8bfa80-78a7-46e9-acf5-d99fec1cc79c\") " Feb 27 08:08:52 crc kubenswrapper[4612]: I0227 08:08:52.654871 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4e0957b6-31f8-4664-96d9-f75bba20f23f-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "4e0957b6-31f8-4664-96d9-f75bba20f23f" (UID: "4e0957b6-31f8-4664-96d9-f75bba20f23f"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 08:08:52 crc kubenswrapper[4612]: I0227 08:08:52.655226 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cc36db3f-167a-47e6-807e-11234c86957a-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "cc36db3f-167a-47e6-807e-11234c86957a" (UID: "cc36db3f-167a-47e6-807e-11234c86957a"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 08:08:52 crc kubenswrapper[4612]: I0227 08:08:52.655428 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c01bd6e3-e867-4506-ab09-75433d751574-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "c01bd6e3-e867-4506-ab09-75433d751574" (UID: "c01bd6e3-e867-4506-ab09-75433d751574"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 08:08:52 crc kubenswrapper[4612]: I0227 08:08:52.655861 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fb8bfa80-78a7-46e9-acf5-d99fec1cc79c-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "fb8bfa80-78a7-46e9-acf5-d99fec1cc79c" (UID: "fb8bfa80-78a7-46e9-acf5-d99fec1cc79c"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 08:08:52 crc kubenswrapper[4612]: I0227 08:08:52.656077 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/04e60e3e-bfd4-4936-be0a-d6688312aeae-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "04e60e3e-bfd4-4936-be0a-d6688312aeae" (UID: "04e60e3e-bfd4-4936-be0a-d6688312aeae"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 08:08:52 crc kubenswrapper[4612]: I0227 08:08:52.656551 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9708a624-a4ef-4250-b3b1-2afb6b56b5c7-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "9708a624-a4ef-4250-b3b1-2afb6b56b5c7" (UID: "9708a624-a4ef-4250-b3b1-2afb6b56b5c7"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 08:08:52 crc kubenswrapper[4612]: I0227 08:08:52.659272 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4e0957b6-31f8-4664-96d9-f75bba20f23f-kube-api-access-jvz92" (OuterVolumeSpecName: "kube-api-access-jvz92") pod "4e0957b6-31f8-4664-96d9-f75bba20f23f" (UID: "4e0957b6-31f8-4664-96d9-f75bba20f23f"). InnerVolumeSpecName "kube-api-access-jvz92". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 08:08:52 crc kubenswrapper[4612]: I0227 08:08:52.660081 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fb8bfa80-78a7-46e9-acf5-d99fec1cc79c-kube-api-access-8lhtp" (OuterVolumeSpecName: "kube-api-access-8lhtp") pod "fb8bfa80-78a7-46e9-acf5-d99fec1cc79c" (UID: "fb8bfa80-78a7-46e9-acf5-d99fec1cc79c"). InnerVolumeSpecName "kube-api-access-8lhtp". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 08:08:52 crc kubenswrapper[4612]: I0227 08:08:52.661482 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/04e60e3e-bfd4-4936-be0a-d6688312aeae-kube-api-access-dkw4c" (OuterVolumeSpecName: "kube-api-access-dkw4c") pod "04e60e3e-bfd4-4936-be0a-d6688312aeae" (UID: "04e60e3e-bfd4-4936-be0a-d6688312aeae"). InnerVolumeSpecName "kube-api-access-dkw4c". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 08:08:52 crc kubenswrapper[4612]: I0227 08:08:52.662557 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9708a624-a4ef-4250-b3b1-2afb6b56b5c7-kube-api-access-thhlx" (OuterVolumeSpecName: "kube-api-access-thhlx") pod "9708a624-a4ef-4250-b3b1-2afb6b56b5c7" (UID: "9708a624-a4ef-4250-b3b1-2afb6b56b5c7"). InnerVolumeSpecName "kube-api-access-thhlx". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 08:08:52 crc kubenswrapper[4612]: I0227 08:08:52.662816 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cc36db3f-167a-47e6-807e-11234c86957a-kube-api-access-f2cf4" (OuterVolumeSpecName: "kube-api-access-f2cf4") pod "cc36db3f-167a-47e6-807e-11234c86957a" (UID: "cc36db3f-167a-47e6-807e-11234c86957a"). InnerVolumeSpecName "kube-api-access-f2cf4". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 08:08:52 crc kubenswrapper[4612]: I0227 08:08:52.664847 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c01bd6e3-e867-4506-ab09-75433d751574-kube-api-access-xfmkf" (OuterVolumeSpecName: "kube-api-access-xfmkf") pod "c01bd6e3-e867-4506-ab09-75433d751574" (UID: "c01bd6e3-e867-4506-ab09-75433d751574"). InnerVolumeSpecName "kube-api-access-xfmkf". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 08:08:52 crc kubenswrapper[4612]: I0227 08:08:52.755582 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dkw4c\" (UniqueName: \"kubernetes.io/projected/04e60e3e-bfd4-4936-be0a-d6688312aeae-kube-api-access-dkw4c\") on node \"crc\" DevicePath \"\"" Feb 27 08:08:52 crc kubenswrapper[4612]: I0227 08:08:52.755618 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xfmkf\" (UniqueName: \"kubernetes.io/projected/c01bd6e3-e867-4506-ab09-75433d751574-kube-api-access-xfmkf\") on node \"crc\" DevicePath \"\"" Feb 27 08:08:52 crc kubenswrapper[4612]: I0227 08:08:52.755635 4612 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fb8bfa80-78a7-46e9-acf5-d99fec1cc79c-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 27 08:08:52 crc kubenswrapper[4612]: I0227 08:08:52.755648 4612 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9708a624-a4ef-4250-b3b1-2afb6b56b5c7-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 27 08:08:52 crc kubenswrapper[4612]: I0227 08:08:52.755661 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f2cf4\" (UniqueName: \"kubernetes.io/projected/cc36db3f-167a-47e6-807e-11234c86957a-kube-api-access-f2cf4\") on node \"crc\" DevicePath \"\"" Feb 27 08:08:52 crc kubenswrapper[4612]: I0227 08:08:52.755673 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-thhlx\" (UniqueName: \"kubernetes.io/projected/9708a624-a4ef-4250-b3b1-2afb6b56b5c7-kube-api-access-thhlx\") on node \"crc\" DevicePath \"\"" Feb 27 08:08:52 crc kubenswrapper[4612]: I0227 08:08:52.755684 4612 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/04e60e3e-bfd4-4936-be0a-d6688312aeae-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 27 08:08:52 crc kubenswrapper[4612]: I0227 08:08:52.755723 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jvz92\" (UniqueName: \"kubernetes.io/projected/4e0957b6-31f8-4664-96d9-f75bba20f23f-kube-api-access-jvz92\") on node \"crc\" DevicePath \"\"" Feb 27 08:08:52 crc kubenswrapper[4612]: I0227 08:08:52.755738 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8lhtp\" (UniqueName: \"kubernetes.io/projected/fb8bfa80-78a7-46e9-acf5-d99fec1cc79c-kube-api-access-8lhtp\") on node \"crc\" DevicePath \"\"" Feb 27 08:08:52 crc kubenswrapper[4612]: I0227 08:08:52.755750 4612 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cc36db3f-167a-47e6-807e-11234c86957a-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 27 08:08:52 crc kubenswrapper[4612]: I0227 08:08:52.755761 4612 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c01bd6e3-e867-4506-ab09-75433d751574-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 27 08:08:52 crc kubenswrapper[4612]: I0227 08:08:52.755772 4612 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4e0957b6-31f8-4664-96d9-f75bba20f23f-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 27 08:08:53 crc kubenswrapper[4612]: I0227 08:08:53.238942 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5f59b8f679-j8tjt" Feb 27 08:08:53 crc kubenswrapper[4612]: I0227 08:08:53.333831 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-s2x4c"] Feb 27 08:08:53 crc kubenswrapper[4612]: I0227 08:08:53.334366 4612 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-b8fbc5445-s2x4c" podUID="c7a8adaa-cf40-4c58-964e-4bc43934baa1" containerName="dnsmasq-dns" containerID="cri-o://f9dcd8a069d0a8cd4b275be183e3b0b8e7b97cfaa9dc697c783028f9c8d0564b" gracePeriod=10 Feb 27 08:08:53 crc kubenswrapper[4612]: I0227 08:08:53.643495 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-75fb5" event={"ID":"c1e74a83-8d08-428a-b4d7-9a0c60ca4b5e","Type":"ContainerStarted","Data":"377a280bda8254e93f4617a17b0bf8037e8524f573ac151b55eb7e717386282e"} Feb 27 08:08:53 crc kubenswrapper[4612]: I0227 08:08:53.647971 4612 generic.go:334] "Generic (PLEG): container finished" podID="c7a8adaa-cf40-4c58-964e-4bc43934baa1" containerID="f9dcd8a069d0a8cd4b275be183e3b0b8e7b97cfaa9dc697c783028f9c8d0564b" exitCode=0 Feb 27 08:08:53 crc kubenswrapper[4612]: I0227 08:08:53.648325 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-s2x4c" event={"ID":"c7a8adaa-cf40-4c58-964e-4bc43934baa1","Type":"ContainerDied","Data":"f9dcd8a069d0a8cd4b275be183e3b0b8e7b97cfaa9dc697c783028f9c8d0564b"} Feb 27 08:08:53 crc kubenswrapper[4612]: I0227 08:08:53.664888 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-75fb5" podStartSLOduration=2.753126133 podStartE2EDuration="7.664863234s" podCreationTimestamp="2026-02-27 08:08:46 +0000 UTC" firstStartedPulling="2026-02-27 08:08:47.44008404 +0000 UTC m=+1185.294014038" lastFinishedPulling="2026-02-27 08:08:52.351821131 +0000 UTC m=+1190.205751139" observedRunningTime="2026-02-27 08:08:53.663932137 +0000 UTC m=+1191.517862135" watchObservedRunningTime="2026-02-27 08:08:53.664863234 +0000 UTC m=+1191.518793232" Feb 27 08:08:53 crc kubenswrapper[4612]: I0227 08:08:53.853362 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8fbc5445-s2x4c" Feb 27 08:08:54 crc kubenswrapper[4612]: I0227 08:08:54.000718 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c7a8adaa-cf40-4c58-964e-4bc43934baa1-ovsdbserver-nb\") pod \"c7a8adaa-cf40-4c58-964e-4bc43934baa1\" (UID: \"c7a8adaa-cf40-4c58-964e-4bc43934baa1\") " Feb 27 08:08:54 crc kubenswrapper[4612]: I0227 08:08:54.000777 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c7a8adaa-cf40-4c58-964e-4bc43934baa1-dns-svc\") pod \"c7a8adaa-cf40-4c58-964e-4bc43934baa1\" (UID: \"c7a8adaa-cf40-4c58-964e-4bc43934baa1\") " Feb 27 08:08:54 crc kubenswrapper[4612]: I0227 08:08:54.000810 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c7a8adaa-cf40-4c58-964e-4bc43934baa1-ovsdbserver-sb\") pod \"c7a8adaa-cf40-4c58-964e-4bc43934baa1\" (UID: \"c7a8adaa-cf40-4c58-964e-4bc43934baa1\") " Feb 27 08:08:54 crc kubenswrapper[4612]: I0227 08:08:54.000840 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c7a8adaa-cf40-4c58-964e-4bc43934baa1-config\") pod \"c7a8adaa-cf40-4c58-964e-4bc43934baa1\" (UID: \"c7a8adaa-cf40-4c58-964e-4bc43934baa1\") " Feb 27 08:08:54 crc kubenswrapper[4612]: I0227 08:08:54.000919 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5lxnn\" (UniqueName: \"kubernetes.io/projected/c7a8adaa-cf40-4c58-964e-4bc43934baa1-kube-api-access-5lxnn\") pod \"c7a8adaa-cf40-4c58-964e-4bc43934baa1\" (UID: \"c7a8adaa-cf40-4c58-964e-4bc43934baa1\") " Feb 27 08:08:54 crc kubenswrapper[4612]: I0227 08:08:54.013019 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c7a8adaa-cf40-4c58-964e-4bc43934baa1-kube-api-access-5lxnn" (OuterVolumeSpecName: "kube-api-access-5lxnn") pod "c7a8adaa-cf40-4c58-964e-4bc43934baa1" (UID: "c7a8adaa-cf40-4c58-964e-4bc43934baa1"). InnerVolumeSpecName "kube-api-access-5lxnn". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 08:08:54 crc kubenswrapper[4612]: I0227 08:08:54.036327 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c7a8adaa-cf40-4c58-964e-4bc43934baa1-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "c7a8adaa-cf40-4c58-964e-4bc43934baa1" (UID: "c7a8adaa-cf40-4c58-964e-4bc43934baa1"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 08:08:54 crc kubenswrapper[4612]: I0227 08:08:54.057585 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c7a8adaa-cf40-4c58-964e-4bc43934baa1-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "c7a8adaa-cf40-4c58-964e-4bc43934baa1" (UID: "c7a8adaa-cf40-4c58-964e-4bc43934baa1"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 08:08:54 crc kubenswrapper[4612]: I0227 08:08:54.059034 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c7a8adaa-cf40-4c58-964e-4bc43934baa1-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "c7a8adaa-cf40-4c58-964e-4bc43934baa1" (UID: "c7a8adaa-cf40-4c58-964e-4bc43934baa1"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 08:08:54 crc kubenswrapper[4612]: I0227 08:08:54.089322 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c7a8adaa-cf40-4c58-964e-4bc43934baa1-config" (OuterVolumeSpecName: "config") pod "c7a8adaa-cf40-4c58-964e-4bc43934baa1" (UID: "c7a8adaa-cf40-4c58-964e-4bc43934baa1"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 08:08:54 crc kubenswrapper[4612]: I0227 08:08:54.104878 4612 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c7a8adaa-cf40-4c58-964e-4bc43934baa1-config\") on node \"crc\" DevicePath \"\"" Feb 27 08:08:54 crc kubenswrapper[4612]: I0227 08:08:54.104924 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5lxnn\" (UniqueName: \"kubernetes.io/projected/c7a8adaa-cf40-4c58-964e-4bc43934baa1-kube-api-access-5lxnn\") on node \"crc\" DevicePath \"\"" Feb 27 08:08:54 crc kubenswrapper[4612]: I0227 08:08:54.104938 4612 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c7a8adaa-cf40-4c58-964e-4bc43934baa1-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 27 08:08:54 crc kubenswrapper[4612]: I0227 08:08:54.104950 4612 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c7a8adaa-cf40-4c58-964e-4bc43934baa1-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 27 08:08:54 crc kubenswrapper[4612]: I0227 08:08:54.104963 4612 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c7a8adaa-cf40-4c58-964e-4bc43934baa1-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 27 08:08:54 crc kubenswrapper[4612]: I0227 08:08:54.659115 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8fbc5445-s2x4c" Feb 27 08:08:54 crc kubenswrapper[4612]: I0227 08:08:54.659122 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-s2x4c" event={"ID":"c7a8adaa-cf40-4c58-964e-4bc43934baa1","Type":"ContainerDied","Data":"34641ee5ef5076347513a394ae91cb085c75f5f357ab620d54332604c68b8bad"} Feb 27 08:08:54 crc kubenswrapper[4612]: I0227 08:08:54.659982 4612 scope.go:117] "RemoveContainer" containerID="f9dcd8a069d0a8cd4b275be183e3b0b8e7b97cfaa9dc697c783028f9c8d0564b" Feb 27 08:08:54 crc kubenswrapper[4612]: I0227 08:08:54.686117 4612 scope.go:117] "RemoveContainer" containerID="5a3878c9a7846dbfb666822a446df679f70e9cabd148b9bb6c58aa17645290b9" Feb 27 08:08:54 crc kubenswrapper[4612]: I0227 08:08:54.691550 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-s2x4c"] Feb 27 08:08:54 crc kubenswrapper[4612]: I0227 08:08:54.703279 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-s2x4c"] Feb 27 08:08:54 crc kubenswrapper[4612]: I0227 08:08:54.864397 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c7a8adaa-cf40-4c58-964e-4bc43934baa1" path="/var/lib/kubelet/pods/c7a8adaa-cf40-4c58-964e-4bc43934baa1/volumes" Feb 27 08:08:56 crc kubenswrapper[4612]: I0227 08:08:56.680644 4612 generic.go:334] "Generic (PLEG): container finished" podID="c1e74a83-8d08-428a-b4d7-9a0c60ca4b5e" containerID="377a280bda8254e93f4617a17b0bf8037e8524f573ac151b55eb7e717386282e" exitCode=0 Feb 27 08:08:56 crc kubenswrapper[4612]: I0227 08:08:56.680725 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-75fb5" event={"ID":"c1e74a83-8d08-428a-b4d7-9a0c60ca4b5e","Type":"ContainerDied","Data":"377a280bda8254e93f4617a17b0bf8037e8524f573ac151b55eb7e717386282e"} Feb 27 08:08:58 crc kubenswrapper[4612]: I0227 08:08:58.026599 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-75fb5" Feb 27 08:08:58 crc kubenswrapper[4612]: I0227 08:08:58.203443 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c1e74a83-8d08-428a-b4d7-9a0c60ca4b5e-config-data\") pod \"c1e74a83-8d08-428a-b4d7-9a0c60ca4b5e\" (UID: \"c1e74a83-8d08-428a-b4d7-9a0c60ca4b5e\") " Feb 27 08:08:58 crc kubenswrapper[4612]: I0227 08:08:58.203543 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6njk\" (UniqueName: \"kubernetes.io/projected/c1e74a83-8d08-428a-b4d7-9a0c60ca4b5e-kube-api-access-d6njk\") pod \"c1e74a83-8d08-428a-b4d7-9a0c60ca4b5e\" (UID: \"c1e74a83-8d08-428a-b4d7-9a0c60ca4b5e\") " Feb 27 08:08:58 crc kubenswrapper[4612]: I0227 08:08:58.203646 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c1e74a83-8d08-428a-b4d7-9a0c60ca4b5e-combined-ca-bundle\") pod \"c1e74a83-8d08-428a-b4d7-9a0c60ca4b5e\" (UID: \"c1e74a83-8d08-428a-b4d7-9a0c60ca4b5e\") " Feb 27 08:08:58 crc kubenswrapper[4612]: I0227 08:08:58.216758 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c1e74a83-8d08-428a-b4d7-9a0c60ca4b5e-kube-api-access-d6njk" (OuterVolumeSpecName: "kube-api-access-d6njk") pod "c1e74a83-8d08-428a-b4d7-9a0c60ca4b5e" (UID: "c1e74a83-8d08-428a-b4d7-9a0c60ca4b5e"). InnerVolumeSpecName "kube-api-access-d6njk". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 08:08:58 crc kubenswrapper[4612]: I0227 08:08:58.235423 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c1e74a83-8d08-428a-b4d7-9a0c60ca4b5e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c1e74a83-8d08-428a-b4d7-9a0c60ca4b5e" (UID: "c1e74a83-8d08-428a-b4d7-9a0c60ca4b5e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:08:58 crc kubenswrapper[4612]: I0227 08:08:58.240832 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c1e74a83-8d08-428a-b4d7-9a0c60ca4b5e-config-data" (OuterVolumeSpecName: "config-data") pod "c1e74a83-8d08-428a-b4d7-9a0c60ca4b5e" (UID: "c1e74a83-8d08-428a-b4d7-9a0c60ca4b5e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:08:58 crc kubenswrapper[4612]: I0227 08:08:58.306005 4612 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c1e74a83-8d08-428a-b4d7-9a0c60ca4b5e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 08:08:58 crc kubenswrapper[4612]: I0227 08:08:58.306040 4612 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c1e74a83-8d08-428a-b4d7-9a0c60ca4b5e-config-data\") on node \"crc\" DevicePath \"\"" Feb 27 08:08:58 crc kubenswrapper[4612]: I0227 08:08:58.306056 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6njk\" (UniqueName: \"kubernetes.io/projected/c1e74a83-8d08-428a-b4d7-9a0c60ca4b5e-kube-api-access-d6njk\") on node \"crc\" DevicePath \"\"" Feb 27 08:08:58 crc kubenswrapper[4612]: I0227 08:08:58.703889 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-75fb5" event={"ID":"c1e74a83-8d08-428a-b4d7-9a0c60ca4b5e","Type":"ContainerDied","Data":"6698df905dde354fb0d7c48f03b8351a06b8f716f97cd62a3d743779f105facf"} Feb 27 08:08:58 crc kubenswrapper[4612]: I0227 08:08:58.703946 4612 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6698df905dde354fb0d7c48f03b8351a06b8f716f97cd62a3d743779f105facf" Feb 27 08:08:58 crc kubenswrapper[4612]: I0227 08:08:58.703979 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-75fb5" Feb 27 08:08:59 crc kubenswrapper[4612]: I0227 08:08:59.030960 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-bbf5cc879-qrwqr"] Feb 27 08:08:59 crc kubenswrapper[4612]: E0227 08:08:59.031325 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fb8bfa80-78a7-46e9-acf5-d99fec1cc79c" containerName="mariadb-database-create" Feb 27 08:08:59 crc kubenswrapper[4612]: I0227 08:08:59.031339 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="fb8bfa80-78a7-46e9-acf5-d99fec1cc79c" containerName="mariadb-database-create" Feb 27 08:08:59 crc kubenswrapper[4612]: E0227 08:08:59.031350 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c01bd6e3-e867-4506-ab09-75433d751574" containerName="mariadb-database-create" Feb 27 08:08:59 crc kubenswrapper[4612]: I0227 08:08:59.031356 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="c01bd6e3-e867-4506-ab09-75433d751574" containerName="mariadb-database-create" Feb 27 08:08:59 crc kubenswrapper[4612]: E0227 08:08:59.031367 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c7a8adaa-cf40-4c58-964e-4bc43934baa1" containerName="dnsmasq-dns" Feb 27 08:08:59 crc kubenswrapper[4612]: I0227 08:08:59.031375 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="c7a8adaa-cf40-4c58-964e-4bc43934baa1" containerName="dnsmasq-dns" Feb 27 08:08:59 crc kubenswrapper[4612]: E0227 08:08:59.031394 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="04e60e3e-bfd4-4936-be0a-d6688312aeae" containerName="mariadb-account-create-update" Feb 27 08:08:59 crc kubenswrapper[4612]: I0227 08:08:59.031399 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="04e60e3e-bfd4-4936-be0a-d6688312aeae" containerName="mariadb-account-create-update" Feb 27 08:08:59 crc kubenswrapper[4612]: E0227 08:08:59.031409 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c7a8adaa-cf40-4c58-964e-4bc43934baa1" containerName="init" Feb 27 08:08:59 crc kubenswrapper[4612]: I0227 08:08:59.031415 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="c7a8adaa-cf40-4c58-964e-4bc43934baa1" containerName="init" Feb 27 08:08:59 crc kubenswrapper[4612]: E0227 08:08:59.031423 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc36db3f-167a-47e6-807e-11234c86957a" containerName="mariadb-account-create-update" Feb 27 08:08:59 crc kubenswrapper[4612]: I0227 08:08:59.031429 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc36db3f-167a-47e6-807e-11234c86957a" containerName="mariadb-account-create-update" Feb 27 08:08:59 crc kubenswrapper[4612]: E0227 08:08:59.031442 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9708a624-a4ef-4250-b3b1-2afb6b56b5c7" containerName="mariadb-account-create-update" Feb 27 08:08:59 crc kubenswrapper[4612]: I0227 08:08:59.031449 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="9708a624-a4ef-4250-b3b1-2afb6b56b5c7" containerName="mariadb-account-create-update" Feb 27 08:08:59 crc kubenswrapper[4612]: E0227 08:08:59.031460 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c1e74a83-8d08-428a-b4d7-9a0c60ca4b5e" containerName="keystone-db-sync" Feb 27 08:08:59 crc kubenswrapper[4612]: I0227 08:08:59.031467 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="c1e74a83-8d08-428a-b4d7-9a0c60ca4b5e" containerName="keystone-db-sync" Feb 27 08:08:59 crc kubenswrapper[4612]: E0227 08:08:59.031476 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4e0957b6-31f8-4664-96d9-f75bba20f23f" containerName="mariadb-database-create" Feb 27 08:08:59 crc kubenswrapper[4612]: I0227 08:08:59.031481 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="4e0957b6-31f8-4664-96d9-f75bba20f23f" containerName="mariadb-database-create" Feb 27 08:08:59 crc kubenswrapper[4612]: I0227 08:08:59.031615 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="fb8bfa80-78a7-46e9-acf5-d99fec1cc79c" containerName="mariadb-database-create" Feb 27 08:08:59 crc kubenswrapper[4612]: I0227 08:08:59.031627 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="c7a8adaa-cf40-4c58-964e-4bc43934baa1" containerName="dnsmasq-dns" Feb 27 08:08:59 crc kubenswrapper[4612]: I0227 08:08:59.031633 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="04e60e3e-bfd4-4936-be0a-d6688312aeae" containerName="mariadb-account-create-update" Feb 27 08:08:59 crc kubenswrapper[4612]: I0227 08:08:59.031641 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="9708a624-a4ef-4250-b3b1-2afb6b56b5c7" containerName="mariadb-account-create-update" Feb 27 08:08:59 crc kubenswrapper[4612]: I0227 08:08:59.031651 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="c1e74a83-8d08-428a-b4d7-9a0c60ca4b5e" containerName="keystone-db-sync" Feb 27 08:08:59 crc kubenswrapper[4612]: I0227 08:08:59.031659 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="c01bd6e3-e867-4506-ab09-75433d751574" containerName="mariadb-database-create" Feb 27 08:08:59 crc kubenswrapper[4612]: I0227 08:08:59.031666 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="4e0957b6-31f8-4664-96d9-f75bba20f23f" containerName="mariadb-database-create" Feb 27 08:08:59 crc kubenswrapper[4612]: I0227 08:08:59.031683 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="cc36db3f-167a-47e6-807e-11234c86957a" containerName="mariadb-account-create-update" Feb 27 08:08:59 crc kubenswrapper[4612]: I0227 08:08:59.032510 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bbf5cc879-qrwqr" Feb 27 08:08:59 crc kubenswrapper[4612]: I0227 08:08:59.042403 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-4nm8j"] Feb 27 08:08:59 crc kubenswrapper[4612]: I0227 08:08:59.043662 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-4nm8j" Feb 27 08:08:59 crc kubenswrapper[4612]: I0227 08:08:59.047744 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Feb 27 08:08:59 crc kubenswrapper[4612]: I0227 08:08:59.050453 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Feb 27 08:08:59 crc kubenswrapper[4612]: I0227 08:08:59.050611 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-zdkwc" Feb 27 08:08:59 crc kubenswrapper[4612]: I0227 08:08:59.050708 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Feb 27 08:08:59 crc kubenswrapper[4612]: I0227 08:08:59.057560 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Feb 27 08:08:59 crc kubenswrapper[4612]: I0227 08:08:59.075294 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-bbf5cc879-qrwqr"] Feb 27 08:08:59 crc kubenswrapper[4612]: I0227 08:08:59.095163 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-4nm8j"] Feb 27 08:08:59 crc kubenswrapper[4612]: I0227 08:08:59.220352 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/dfebce69-1679-4abd-8b9d-270046157f42-credential-keys\") pod \"keystone-bootstrap-4nm8j\" (UID: \"dfebce69-1679-4abd-8b9d-270046157f42\") " pod="openstack/keystone-bootstrap-4nm8j" Feb 27 08:08:59 crc kubenswrapper[4612]: I0227 08:08:59.220397 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-klrr2\" (UniqueName: \"kubernetes.io/projected/0a980065-12fa-4005-9bc7-cdde0637d18e-kube-api-access-klrr2\") pod \"dnsmasq-dns-bbf5cc879-qrwqr\" (UID: \"0a980065-12fa-4005-9bc7-cdde0637d18e\") " pod="openstack/dnsmasq-dns-bbf5cc879-qrwqr" Feb 27 08:08:59 crc kubenswrapper[4612]: I0227 08:08:59.220467 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-flp5k\" (UniqueName: \"kubernetes.io/projected/dfebce69-1679-4abd-8b9d-270046157f42-kube-api-access-flp5k\") pod \"keystone-bootstrap-4nm8j\" (UID: \"dfebce69-1679-4abd-8b9d-270046157f42\") " pod="openstack/keystone-bootstrap-4nm8j" Feb 27 08:08:59 crc kubenswrapper[4612]: I0227 08:08:59.220516 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0a980065-12fa-4005-9bc7-cdde0637d18e-config\") pod \"dnsmasq-dns-bbf5cc879-qrwqr\" (UID: \"0a980065-12fa-4005-9bc7-cdde0637d18e\") " pod="openstack/dnsmasq-dns-bbf5cc879-qrwqr" Feb 27 08:08:59 crc kubenswrapper[4612]: I0227 08:08:59.220535 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0a980065-12fa-4005-9bc7-cdde0637d18e-ovsdbserver-sb\") pod \"dnsmasq-dns-bbf5cc879-qrwqr\" (UID: \"0a980065-12fa-4005-9bc7-cdde0637d18e\") " pod="openstack/dnsmasq-dns-bbf5cc879-qrwqr" Feb 27 08:08:59 crc kubenswrapper[4612]: I0227 08:08:59.220565 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dfebce69-1679-4abd-8b9d-270046157f42-scripts\") pod \"keystone-bootstrap-4nm8j\" (UID: \"dfebce69-1679-4abd-8b9d-270046157f42\") " pod="openstack/keystone-bootstrap-4nm8j" Feb 27 08:08:59 crc kubenswrapper[4612]: I0227 08:08:59.220616 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dfebce69-1679-4abd-8b9d-270046157f42-combined-ca-bundle\") pod \"keystone-bootstrap-4nm8j\" (UID: \"dfebce69-1679-4abd-8b9d-270046157f42\") " pod="openstack/keystone-bootstrap-4nm8j" Feb 27 08:08:59 crc kubenswrapper[4612]: I0227 08:08:59.220641 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0a980065-12fa-4005-9bc7-cdde0637d18e-ovsdbserver-nb\") pod \"dnsmasq-dns-bbf5cc879-qrwqr\" (UID: \"0a980065-12fa-4005-9bc7-cdde0637d18e\") " pod="openstack/dnsmasq-dns-bbf5cc879-qrwqr" Feb 27 08:08:59 crc kubenswrapper[4612]: I0227 08:08:59.220659 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0a980065-12fa-4005-9bc7-cdde0637d18e-dns-svc\") pod \"dnsmasq-dns-bbf5cc879-qrwqr\" (UID: \"0a980065-12fa-4005-9bc7-cdde0637d18e\") " pod="openstack/dnsmasq-dns-bbf5cc879-qrwqr" Feb 27 08:08:59 crc kubenswrapper[4612]: I0227 08:08:59.220674 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dfebce69-1679-4abd-8b9d-270046157f42-config-data\") pod \"keystone-bootstrap-4nm8j\" (UID: \"dfebce69-1679-4abd-8b9d-270046157f42\") " pod="openstack/keystone-bootstrap-4nm8j" Feb 27 08:08:59 crc kubenswrapper[4612]: I0227 08:08:59.220708 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0a980065-12fa-4005-9bc7-cdde0637d18e-dns-swift-storage-0\") pod \"dnsmasq-dns-bbf5cc879-qrwqr\" (UID: \"0a980065-12fa-4005-9bc7-cdde0637d18e\") " pod="openstack/dnsmasq-dns-bbf5cc879-qrwqr" Feb 27 08:08:59 crc kubenswrapper[4612]: I0227 08:08:59.220727 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/dfebce69-1679-4abd-8b9d-270046157f42-fernet-keys\") pod \"keystone-bootstrap-4nm8j\" (UID: \"dfebce69-1679-4abd-8b9d-270046157f42\") " pod="openstack/keystone-bootstrap-4nm8j" Feb 27 08:08:59 crc kubenswrapper[4612]: I0227 08:08:59.309668 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-l5z5s"] Feb 27 08:08:59 crc kubenswrapper[4612]: I0227 08:08:59.310988 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-l5z5s" Feb 27 08:08:59 crc kubenswrapper[4612]: I0227 08:08:59.321744 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0a980065-12fa-4005-9bc7-cdde0637d18e-dns-swift-storage-0\") pod \"dnsmasq-dns-bbf5cc879-qrwqr\" (UID: \"0a980065-12fa-4005-9bc7-cdde0637d18e\") " pod="openstack/dnsmasq-dns-bbf5cc879-qrwqr" Feb 27 08:08:59 crc kubenswrapper[4612]: I0227 08:08:59.322082 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/dfebce69-1679-4abd-8b9d-270046157f42-fernet-keys\") pod \"keystone-bootstrap-4nm8j\" (UID: \"dfebce69-1679-4abd-8b9d-270046157f42\") " pod="openstack/keystone-bootstrap-4nm8j" Feb 27 08:08:59 crc kubenswrapper[4612]: I0227 08:08:59.322155 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/dfebce69-1679-4abd-8b9d-270046157f42-credential-keys\") pod \"keystone-bootstrap-4nm8j\" (UID: \"dfebce69-1679-4abd-8b9d-270046157f42\") " pod="openstack/keystone-bootstrap-4nm8j" Feb 27 08:08:59 crc kubenswrapper[4612]: I0227 08:08:59.322177 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-klrr2\" (UniqueName: \"kubernetes.io/projected/0a980065-12fa-4005-9bc7-cdde0637d18e-kube-api-access-klrr2\") pod \"dnsmasq-dns-bbf5cc879-qrwqr\" (UID: \"0a980065-12fa-4005-9bc7-cdde0637d18e\") " pod="openstack/dnsmasq-dns-bbf5cc879-qrwqr" Feb 27 08:08:59 crc kubenswrapper[4612]: I0227 08:08:59.322208 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-flp5k\" (UniqueName: \"kubernetes.io/projected/dfebce69-1679-4abd-8b9d-270046157f42-kube-api-access-flp5k\") pod \"keystone-bootstrap-4nm8j\" (UID: \"dfebce69-1679-4abd-8b9d-270046157f42\") " pod="openstack/keystone-bootstrap-4nm8j" Feb 27 08:08:59 crc kubenswrapper[4612]: I0227 08:08:59.322228 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0a980065-12fa-4005-9bc7-cdde0637d18e-config\") pod \"dnsmasq-dns-bbf5cc879-qrwqr\" (UID: \"0a980065-12fa-4005-9bc7-cdde0637d18e\") " pod="openstack/dnsmasq-dns-bbf5cc879-qrwqr" Feb 27 08:08:59 crc kubenswrapper[4612]: I0227 08:08:59.322247 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0a980065-12fa-4005-9bc7-cdde0637d18e-ovsdbserver-sb\") pod \"dnsmasq-dns-bbf5cc879-qrwqr\" (UID: \"0a980065-12fa-4005-9bc7-cdde0637d18e\") " pod="openstack/dnsmasq-dns-bbf5cc879-qrwqr" Feb 27 08:08:59 crc kubenswrapper[4612]: I0227 08:08:59.322276 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dfebce69-1679-4abd-8b9d-270046157f42-scripts\") pod \"keystone-bootstrap-4nm8j\" (UID: \"dfebce69-1679-4abd-8b9d-270046157f42\") " pod="openstack/keystone-bootstrap-4nm8j" Feb 27 08:08:59 crc kubenswrapper[4612]: I0227 08:08:59.322320 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dfebce69-1679-4abd-8b9d-270046157f42-combined-ca-bundle\") pod \"keystone-bootstrap-4nm8j\" (UID: \"dfebce69-1679-4abd-8b9d-270046157f42\") " pod="openstack/keystone-bootstrap-4nm8j" Feb 27 08:08:59 crc kubenswrapper[4612]: I0227 08:08:59.322346 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0a980065-12fa-4005-9bc7-cdde0637d18e-ovsdbserver-nb\") pod \"dnsmasq-dns-bbf5cc879-qrwqr\" (UID: \"0a980065-12fa-4005-9bc7-cdde0637d18e\") " pod="openstack/dnsmasq-dns-bbf5cc879-qrwqr" Feb 27 08:08:59 crc kubenswrapper[4612]: I0227 08:08:59.322365 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0a980065-12fa-4005-9bc7-cdde0637d18e-dns-svc\") pod \"dnsmasq-dns-bbf5cc879-qrwqr\" (UID: \"0a980065-12fa-4005-9bc7-cdde0637d18e\") " pod="openstack/dnsmasq-dns-bbf5cc879-qrwqr" Feb 27 08:08:59 crc kubenswrapper[4612]: I0227 08:08:59.322383 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dfebce69-1679-4abd-8b9d-270046157f42-config-data\") pod \"keystone-bootstrap-4nm8j\" (UID: \"dfebce69-1679-4abd-8b9d-270046157f42\") " pod="openstack/keystone-bootstrap-4nm8j" Feb 27 08:08:59 crc kubenswrapper[4612]: I0227 08:08:59.322919 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0a980065-12fa-4005-9bc7-cdde0637d18e-dns-swift-storage-0\") pod \"dnsmasq-dns-bbf5cc879-qrwqr\" (UID: \"0a980065-12fa-4005-9bc7-cdde0637d18e\") " pod="openstack/dnsmasq-dns-bbf5cc879-qrwqr" Feb 27 08:08:59 crc kubenswrapper[4612]: I0227 08:08:59.323275 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0a980065-12fa-4005-9bc7-cdde0637d18e-config\") pod \"dnsmasq-dns-bbf5cc879-qrwqr\" (UID: \"0a980065-12fa-4005-9bc7-cdde0637d18e\") " pod="openstack/dnsmasq-dns-bbf5cc879-qrwqr" Feb 27 08:08:59 crc kubenswrapper[4612]: I0227 08:08:59.323335 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0a980065-12fa-4005-9bc7-cdde0637d18e-ovsdbserver-sb\") pod \"dnsmasq-dns-bbf5cc879-qrwqr\" (UID: \"0a980065-12fa-4005-9bc7-cdde0637d18e\") " pod="openstack/dnsmasq-dns-bbf5cc879-qrwqr" Feb 27 08:08:59 crc kubenswrapper[4612]: I0227 08:08:59.323566 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0a980065-12fa-4005-9bc7-cdde0637d18e-ovsdbserver-nb\") pod \"dnsmasq-dns-bbf5cc879-qrwqr\" (UID: \"0a980065-12fa-4005-9bc7-cdde0637d18e\") " pod="openstack/dnsmasq-dns-bbf5cc879-qrwqr" Feb 27 08:08:59 crc kubenswrapper[4612]: I0227 08:08:59.323879 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0a980065-12fa-4005-9bc7-cdde0637d18e-dns-svc\") pod \"dnsmasq-dns-bbf5cc879-qrwqr\" (UID: \"0a980065-12fa-4005-9bc7-cdde0637d18e\") " pod="openstack/dnsmasq-dns-bbf5cc879-qrwqr" Feb 27 08:08:59 crc kubenswrapper[4612]: I0227 08:08:59.334117 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dfebce69-1679-4abd-8b9d-270046157f42-config-data\") pod \"keystone-bootstrap-4nm8j\" (UID: \"dfebce69-1679-4abd-8b9d-270046157f42\") " pod="openstack/keystone-bootstrap-4nm8j" Feb 27 08:08:59 crc kubenswrapper[4612]: I0227 08:08:59.334231 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/dfebce69-1679-4abd-8b9d-270046157f42-credential-keys\") pod \"keystone-bootstrap-4nm8j\" (UID: \"dfebce69-1679-4abd-8b9d-270046157f42\") " pod="openstack/keystone-bootstrap-4nm8j" Feb 27 08:08:59 crc kubenswrapper[4612]: I0227 08:08:59.334348 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dfebce69-1679-4abd-8b9d-270046157f42-combined-ca-bundle\") pod \"keystone-bootstrap-4nm8j\" (UID: \"dfebce69-1679-4abd-8b9d-270046157f42\") " pod="openstack/keystone-bootstrap-4nm8j" Feb 27 08:08:59 crc kubenswrapper[4612]: I0227 08:08:59.343143 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Feb 27 08:08:59 crc kubenswrapper[4612]: I0227 08:08:59.343333 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-5v76w" Feb 27 08:08:59 crc kubenswrapper[4612]: I0227 08:08:59.343415 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Feb 27 08:08:59 crc kubenswrapper[4612]: I0227 08:08:59.346054 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dfebce69-1679-4abd-8b9d-270046157f42-scripts\") pod \"keystone-bootstrap-4nm8j\" (UID: \"dfebce69-1679-4abd-8b9d-270046157f42\") " pod="openstack/keystone-bootstrap-4nm8j" Feb 27 08:08:59 crc kubenswrapper[4612]: I0227 08:08:59.347666 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/dfebce69-1679-4abd-8b9d-270046157f42-fernet-keys\") pod \"keystone-bootstrap-4nm8j\" (UID: \"dfebce69-1679-4abd-8b9d-270046157f42\") " pod="openstack/keystone-bootstrap-4nm8j" Feb 27 08:08:59 crc kubenswrapper[4612]: I0227 08:08:59.356948 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-flp5k\" (UniqueName: \"kubernetes.io/projected/dfebce69-1679-4abd-8b9d-270046157f42-kube-api-access-flp5k\") pod \"keystone-bootstrap-4nm8j\" (UID: \"dfebce69-1679-4abd-8b9d-270046157f42\") " pod="openstack/keystone-bootstrap-4nm8j" Feb 27 08:08:59 crc kubenswrapper[4612]: I0227 08:08:59.366302 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-4nm8j" Feb 27 08:08:59 crc kubenswrapper[4612]: I0227 08:08:59.390593 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-4ljbd"] Feb 27 08:08:59 crc kubenswrapper[4612]: I0227 08:08:59.391845 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-4ljbd" Feb 27 08:08:59 crc kubenswrapper[4612]: I0227 08:08:59.409199 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Feb 27 08:08:59 crc kubenswrapper[4612]: I0227 08:08:59.409383 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Feb 27 08:08:59 crc kubenswrapper[4612]: I0227 08:08:59.409495 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-2j2mr" Feb 27 08:08:59 crc kubenswrapper[4612]: I0227 08:08:59.425766 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-4ljbd"] Feb 27 08:08:59 crc kubenswrapper[4612]: I0227 08:08:59.426746 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0c9482e8-9eb1-4c30-95e9-bdaa5ccd239b-scripts\") pod \"cinder-db-sync-l5z5s\" (UID: \"0c9482e8-9eb1-4c30-95e9-bdaa5ccd239b\") " pod="openstack/cinder-db-sync-l5z5s" Feb 27 08:08:59 crc kubenswrapper[4612]: I0227 08:08:59.426799 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0c9482e8-9eb1-4c30-95e9-bdaa5ccd239b-etc-machine-id\") pod \"cinder-db-sync-l5z5s\" (UID: \"0c9482e8-9eb1-4c30-95e9-bdaa5ccd239b\") " pod="openstack/cinder-db-sync-l5z5s" Feb 27 08:08:59 crc kubenswrapper[4612]: I0227 08:08:59.426840 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0c9482e8-9eb1-4c30-95e9-bdaa5ccd239b-config-data\") pod \"cinder-db-sync-l5z5s\" (UID: \"0c9482e8-9eb1-4c30-95e9-bdaa5ccd239b\") " pod="openstack/cinder-db-sync-l5z5s" Feb 27 08:08:59 crc kubenswrapper[4612]: I0227 08:08:59.426900 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f63a4b12-f7e9-43a4-91d2-fd46f65daf92-combined-ca-bundle\") pod \"neutron-db-sync-4ljbd\" (UID: \"f63a4b12-f7e9-43a4-91d2-fd46f65daf92\") " pod="openstack/neutron-db-sync-4ljbd" Feb 27 08:08:59 crc kubenswrapper[4612]: I0227 08:08:59.426922 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c9482e8-9eb1-4c30-95e9-bdaa5ccd239b-combined-ca-bundle\") pod \"cinder-db-sync-l5z5s\" (UID: \"0c9482e8-9eb1-4c30-95e9-bdaa5ccd239b\") " pod="openstack/cinder-db-sync-l5z5s" Feb 27 08:08:59 crc kubenswrapper[4612]: I0227 08:08:59.426992 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fhfxv\" (UniqueName: \"kubernetes.io/projected/0c9482e8-9eb1-4c30-95e9-bdaa5ccd239b-kube-api-access-fhfxv\") pod \"cinder-db-sync-l5z5s\" (UID: \"0c9482e8-9eb1-4c30-95e9-bdaa5ccd239b\") " pod="openstack/cinder-db-sync-l5z5s" Feb 27 08:08:59 crc kubenswrapper[4612]: I0227 08:08:59.427019 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/0c9482e8-9eb1-4c30-95e9-bdaa5ccd239b-db-sync-config-data\") pod \"cinder-db-sync-l5z5s\" (UID: \"0c9482e8-9eb1-4c30-95e9-bdaa5ccd239b\") " pod="openstack/cinder-db-sync-l5z5s" Feb 27 08:08:59 crc kubenswrapper[4612]: I0227 08:08:59.427084 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/f63a4b12-f7e9-43a4-91d2-fd46f65daf92-config\") pod \"neutron-db-sync-4ljbd\" (UID: \"f63a4b12-f7e9-43a4-91d2-fd46f65daf92\") " pod="openstack/neutron-db-sync-4ljbd" Feb 27 08:08:59 crc kubenswrapper[4612]: I0227 08:08:59.427173 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zrxvv\" (UniqueName: \"kubernetes.io/projected/f63a4b12-f7e9-43a4-91d2-fd46f65daf92-kube-api-access-zrxvv\") pod \"neutron-db-sync-4ljbd\" (UID: \"f63a4b12-f7e9-43a4-91d2-fd46f65daf92\") " pod="openstack/neutron-db-sync-4ljbd" Feb 27 08:08:59 crc kubenswrapper[4612]: I0227 08:08:59.453396 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-klrr2\" (UniqueName: \"kubernetes.io/projected/0a980065-12fa-4005-9bc7-cdde0637d18e-kube-api-access-klrr2\") pod \"dnsmasq-dns-bbf5cc879-qrwqr\" (UID: \"0a980065-12fa-4005-9bc7-cdde0637d18e\") " pod="openstack/dnsmasq-dns-bbf5cc879-qrwqr" Feb 27 08:08:59 crc kubenswrapper[4612]: I0227 08:08:59.483024 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-l5z5s"] Feb 27 08:08:59 crc kubenswrapper[4612]: I0227 08:08:59.540408 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zrxvv\" (UniqueName: \"kubernetes.io/projected/f63a4b12-f7e9-43a4-91d2-fd46f65daf92-kube-api-access-zrxvv\") pod \"neutron-db-sync-4ljbd\" (UID: \"f63a4b12-f7e9-43a4-91d2-fd46f65daf92\") " pod="openstack/neutron-db-sync-4ljbd" Feb 27 08:08:59 crc kubenswrapper[4612]: I0227 08:08:59.540461 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0c9482e8-9eb1-4c30-95e9-bdaa5ccd239b-scripts\") pod \"cinder-db-sync-l5z5s\" (UID: \"0c9482e8-9eb1-4c30-95e9-bdaa5ccd239b\") " pod="openstack/cinder-db-sync-l5z5s" Feb 27 08:08:59 crc kubenswrapper[4612]: I0227 08:08:59.540490 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0c9482e8-9eb1-4c30-95e9-bdaa5ccd239b-etc-machine-id\") pod \"cinder-db-sync-l5z5s\" (UID: \"0c9482e8-9eb1-4c30-95e9-bdaa5ccd239b\") " pod="openstack/cinder-db-sync-l5z5s" Feb 27 08:08:59 crc kubenswrapper[4612]: I0227 08:08:59.540512 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0c9482e8-9eb1-4c30-95e9-bdaa5ccd239b-config-data\") pod \"cinder-db-sync-l5z5s\" (UID: \"0c9482e8-9eb1-4c30-95e9-bdaa5ccd239b\") " pod="openstack/cinder-db-sync-l5z5s" Feb 27 08:08:59 crc kubenswrapper[4612]: I0227 08:08:59.540545 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f63a4b12-f7e9-43a4-91d2-fd46f65daf92-combined-ca-bundle\") pod \"neutron-db-sync-4ljbd\" (UID: \"f63a4b12-f7e9-43a4-91d2-fd46f65daf92\") " pod="openstack/neutron-db-sync-4ljbd" Feb 27 08:08:59 crc kubenswrapper[4612]: I0227 08:08:59.540566 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c9482e8-9eb1-4c30-95e9-bdaa5ccd239b-combined-ca-bundle\") pod \"cinder-db-sync-l5z5s\" (UID: \"0c9482e8-9eb1-4c30-95e9-bdaa5ccd239b\") " pod="openstack/cinder-db-sync-l5z5s" Feb 27 08:08:59 crc kubenswrapper[4612]: I0227 08:08:59.540600 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fhfxv\" (UniqueName: \"kubernetes.io/projected/0c9482e8-9eb1-4c30-95e9-bdaa5ccd239b-kube-api-access-fhfxv\") pod \"cinder-db-sync-l5z5s\" (UID: \"0c9482e8-9eb1-4c30-95e9-bdaa5ccd239b\") " pod="openstack/cinder-db-sync-l5z5s" Feb 27 08:08:59 crc kubenswrapper[4612]: I0227 08:08:59.540623 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/0c9482e8-9eb1-4c30-95e9-bdaa5ccd239b-db-sync-config-data\") pod \"cinder-db-sync-l5z5s\" (UID: \"0c9482e8-9eb1-4c30-95e9-bdaa5ccd239b\") " pod="openstack/cinder-db-sync-l5z5s" Feb 27 08:08:59 crc kubenswrapper[4612]: I0227 08:08:59.540661 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/f63a4b12-f7e9-43a4-91d2-fd46f65daf92-config\") pod \"neutron-db-sync-4ljbd\" (UID: \"f63a4b12-f7e9-43a4-91d2-fd46f65daf92\") " pod="openstack/neutron-db-sync-4ljbd" Feb 27 08:08:59 crc kubenswrapper[4612]: I0227 08:08:59.565888 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0c9482e8-9eb1-4c30-95e9-bdaa5ccd239b-etc-machine-id\") pod \"cinder-db-sync-l5z5s\" (UID: \"0c9482e8-9eb1-4c30-95e9-bdaa5ccd239b\") " pod="openstack/cinder-db-sync-l5z5s" Feb 27 08:08:59 crc kubenswrapper[4612]: I0227 08:08:59.566224 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/f63a4b12-f7e9-43a4-91d2-fd46f65daf92-config\") pod \"neutron-db-sync-4ljbd\" (UID: \"f63a4b12-f7e9-43a4-91d2-fd46f65daf92\") " pod="openstack/neutron-db-sync-4ljbd" Feb 27 08:08:59 crc kubenswrapper[4612]: I0227 08:08:59.574983 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0c9482e8-9eb1-4c30-95e9-bdaa5ccd239b-config-data\") pod \"cinder-db-sync-l5z5s\" (UID: \"0c9482e8-9eb1-4c30-95e9-bdaa5ccd239b\") " pod="openstack/cinder-db-sync-l5z5s" Feb 27 08:08:59 crc kubenswrapper[4612]: I0227 08:08:59.581249 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0c9482e8-9eb1-4c30-95e9-bdaa5ccd239b-scripts\") pod \"cinder-db-sync-l5z5s\" (UID: \"0c9482e8-9eb1-4c30-95e9-bdaa5ccd239b\") " pod="openstack/cinder-db-sync-l5z5s" Feb 27 08:08:59 crc kubenswrapper[4612]: I0227 08:08:59.593853 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fhfxv\" (UniqueName: \"kubernetes.io/projected/0c9482e8-9eb1-4c30-95e9-bdaa5ccd239b-kube-api-access-fhfxv\") pod \"cinder-db-sync-l5z5s\" (UID: \"0c9482e8-9eb1-4c30-95e9-bdaa5ccd239b\") " pod="openstack/cinder-db-sync-l5z5s" Feb 27 08:08:59 crc kubenswrapper[4612]: I0227 08:08:59.594350 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f63a4b12-f7e9-43a4-91d2-fd46f65daf92-combined-ca-bundle\") pod \"neutron-db-sync-4ljbd\" (UID: \"f63a4b12-f7e9-43a4-91d2-fd46f65daf92\") " pod="openstack/neutron-db-sync-4ljbd" Feb 27 08:08:59 crc kubenswrapper[4612]: I0227 08:08:59.630220 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-7f9b6cd66c-5lwzw"] Feb 27 08:08:59 crc kubenswrapper[4612]: I0227 08:08:59.634684 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7f9b6cd66c-5lwzw" Feb 27 08:08:59 crc kubenswrapper[4612]: I0227 08:08:59.642589 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zrxvv\" (UniqueName: \"kubernetes.io/projected/f63a4b12-f7e9-43a4-91d2-fd46f65daf92-kube-api-access-zrxvv\") pod \"neutron-db-sync-4ljbd\" (UID: \"f63a4b12-f7e9-43a4-91d2-fd46f65daf92\") " pod="openstack/neutron-db-sync-4ljbd" Feb 27 08:08:59 crc kubenswrapper[4612]: I0227 08:08:59.648854 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c9482e8-9eb1-4c30-95e9-bdaa5ccd239b-combined-ca-bundle\") pod \"cinder-db-sync-l5z5s\" (UID: \"0c9482e8-9eb1-4c30-95e9-bdaa5ccd239b\") " pod="openstack/cinder-db-sync-l5z5s" Feb 27 08:08:59 crc kubenswrapper[4612]: I0227 08:08:59.649411 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bbf5cc879-qrwqr" Feb 27 08:08:59 crc kubenswrapper[4612]: I0227 08:08:59.654866 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon" Feb 27 08:08:59 crc kubenswrapper[4612]: I0227 08:08:59.655132 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon-horizon-dockercfg-plwzh" Feb 27 08:08:59 crc kubenswrapper[4612]: I0227 08:08:59.655274 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-scripts" Feb 27 08:08:59 crc kubenswrapper[4612]: I0227 08:08:59.655480 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-config-data" Feb 27 08:08:59 crc kubenswrapper[4612]: I0227 08:08:59.694792 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/0c9482e8-9eb1-4c30-95e9-bdaa5ccd239b-db-sync-config-data\") pod \"cinder-db-sync-l5z5s\" (UID: \"0c9482e8-9eb1-4c30-95e9-bdaa5ccd239b\") " pod="openstack/cinder-db-sync-l5z5s" Feb 27 08:08:59 crc kubenswrapper[4612]: I0227 08:08:59.716143 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7f9b6cd66c-5lwzw"] Feb 27 08:08:59 crc kubenswrapper[4612]: I0227 08:08:59.742366 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-58b79554cc-f9ms6"] Feb 27 08:08:59 crc kubenswrapper[4612]: I0227 08:08:59.744331 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-58b79554cc-f9ms6" Feb 27 08:08:59 crc kubenswrapper[4612]: I0227 08:08:59.756312 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-bbf5cc879-qrwqr"] Feb 27 08:08:59 crc kubenswrapper[4612]: I0227 08:08:59.765519 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a10f6a85-63bd-4aa7-91a6-b3aec6ecb9d4-logs\") pod \"horizon-7f9b6cd66c-5lwzw\" (UID: \"a10f6a85-63bd-4aa7-91a6-b3aec6ecb9d4\") " pod="openstack/horizon-7f9b6cd66c-5lwzw" Feb 27 08:08:59 crc kubenswrapper[4612]: I0227 08:08:59.765625 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a10f6a85-63bd-4aa7-91a6-b3aec6ecb9d4-scripts\") pod \"horizon-7f9b6cd66c-5lwzw\" (UID: \"a10f6a85-63bd-4aa7-91a6-b3aec6ecb9d4\") " pod="openstack/horizon-7f9b6cd66c-5lwzw" Feb 27 08:08:59 crc kubenswrapper[4612]: I0227 08:08:59.765664 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a10f6a85-63bd-4aa7-91a6-b3aec6ecb9d4-config-data\") pod \"horizon-7f9b6cd66c-5lwzw\" (UID: \"a10f6a85-63bd-4aa7-91a6-b3aec6ecb9d4\") " pod="openstack/horizon-7f9b6cd66c-5lwzw" Feb 27 08:08:59 crc kubenswrapper[4612]: I0227 08:08:59.765686 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9zthz\" (UniqueName: \"kubernetes.io/projected/a10f6a85-63bd-4aa7-91a6-b3aec6ecb9d4-kube-api-access-9zthz\") pod \"horizon-7f9b6cd66c-5lwzw\" (UID: \"a10f6a85-63bd-4aa7-91a6-b3aec6ecb9d4\") " pod="openstack/horizon-7f9b6cd66c-5lwzw" Feb 27 08:08:59 crc kubenswrapper[4612]: I0227 08:08:59.765721 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/a10f6a85-63bd-4aa7-91a6-b3aec6ecb9d4-horizon-secret-key\") pod \"horizon-7f9b6cd66c-5lwzw\" (UID: \"a10f6a85-63bd-4aa7-91a6-b3aec6ecb9d4\") " pod="openstack/horizon-7f9b6cd66c-5lwzw" Feb 27 08:08:59 crc kubenswrapper[4612]: I0227 08:08:59.769507 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-58b79554cc-f9ms6"] Feb 27 08:08:59 crc kubenswrapper[4612]: I0227 08:08:59.876575 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a10f6a85-63bd-4aa7-91a6-b3aec6ecb9d4-scripts\") pod \"horizon-7f9b6cd66c-5lwzw\" (UID: \"a10f6a85-63bd-4aa7-91a6-b3aec6ecb9d4\") " pod="openstack/horizon-7f9b6cd66c-5lwzw" Feb 27 08:08:59 crc kubenswrapper[4612]: I0227 08:08:59.876624 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/53ff41b8-ca85-4358-a8fe-cb490ec40793-config-data\") pod \"horizon-58b79554cc-f9ms6\" (UID: \"53ff41b8-ca85-4358-a8fe-cb490ec40793\") " pod="openstack/horizon-58b79554cc-f9ms6" Feb 27 08:08:59 crc kubenswrapper[4612]: I0227 08:08:59.876648 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/53ff41b8-ca85-4358-a8fe-cb490ec40793-logs\") pod \"horizon-58b79554cc-f9ms6\" (UID: \"53ff41b8-ca85-4358-a8fe-cb490ec40793\") " pod="openstack/horizon-58b79554cc-f9ms6" Feb 27 08:08:59 crc kubenswrapper[4612]: I0227 08:08:59.876666 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/53ff41b8-ca85-4358-a8fe-cb490ec40793-horizon-secret-key\") pod \"horizon-58b79554cc-f9ms6\" (UID: \"53ff41b8-ca85-4358-a8fe-cb490ec40793\") " pod="openstack/horizon-58b79554cc-f9ms6" Feb 27 08:08:59 crc kubenswrapper[4612]: I0227 08:08:59.876702 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a10f6a85-63bd-4aa7-91a6-b3aec6ecb9d4-config-data\") pod \"horizon-7f9b6cd66c-5lwzw\" (UID: \"a10f6a85-63bd-4aa7-91a6-b3aec6ecb9d4\") " pod="openstack/horizon-7f9b6cd66c-5lwzw" Feb 27 08:08:59 crc kubenswrapper[4612]: I0227 08:08:59.876731 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9zthz\" (UniqueName: \"kubernetes.io/projected/a10f6a85-63bd-4aa7-91a6-b3aec6ecb9d4-kube-api-access-9zthz\") pod \"horizon-7f9b6cd66c-5lwzw\" (UID: \"a10f6a85-63bd-4aa7-91a6-b3aec6ecb9d4\") " pod="openstack/horizon-7f9b6cd66c-5lwzw" Feb 27 08:08:59 crc kubenswrapper[4612]: I0227 08:08:59.876755 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/a10f6a85-63bd-4aa7-91a6-b3aec6ecb9d4-horizon-secret-key\") pod \"horizon-7f9b6cd66c-5lwzw\" (UID: \"a10f6a85-63bd-4aa7-91a6-b3aec6ecb9d4\") " pod="openstack/horizon-7f9b6cd66c-5lwzw" Feb 27 08:08:59 crc kubenswrapper[4612]: I0227 08:08:59.876882 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t2rpv\" (UniqueName: \"kubernetes.io/projected/53ff41b8-ca85-4358-a8fe-cb490ec40793-kube-api-access-t2rpv\") pod \"horizon-58b79554cc-f9ms6\" (UID: \"53ff41b8-ca85-4358-a8fe-cb490ec40793\") " pod="openstack/horizon-58b79554cc-f9ms6" Feb 27 08:08:59 crc kubenswrapper[4612]: I0227 08:08:59.876952 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/53ff41b8-ca85-4358-a8fe-cb490ec40793-scripts\") pod \"horizon-58b79554cc-f9ms6\" (UID: \"53ff41b8-ca85-4358-a8fe-cb490ec40793\") " pod="openstack/horizon-58b79554cc-f9ms6" Feb 27 08:08:59 crc kubenswrapper[4612]: I0227 08:08:59.877095 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a10f6a85-63bd-4aa7-91a6-b3aec6ecb9d4-logs\") pod \"horizon-7f9b6cd66c-5lwzw\" (UID: \"a10f6a85-63bd-4aa7-91a6-b3aec6ecb9d4\") " pod="openstack/horizon-7f9b6cd66c-5lwzw" Feb 27 08:08:59 crc kubenswrapper[4612]: I0227 08:08:59.877559 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a10f6a85-63bd-4aa7-91a6-b3aec6ecb9d4-logs\") pod \"horizon-7f9b6cd66c-5lwzw\" (UID: \"a10f6a85-63bd-4aa7-91a6-b3aec6ecb9d4\") " pod="openstack/horizon-7f9b6cd66c-5lwzw" Feb 27 08:08:59 crc kubenswrapper[4612]: I0227 08:08:59.878106 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a10f6a85-63bd-4aa7-91a6-b3aec6ecb9d4-scripts\") pod \"horizon-7f9b6cd66c-5lwzw\" (UID: \"a10f6a85-63bd-4aa7-91a6-b3aec6ecb9d4\") " pod="openstack/horizon-7f9b6cd66c-5lwzw" Feb 27 08:08:59 crc kubenswrapper[4612]: I0227 08:08:59.882206 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-56df8fb6b7-9p274"] Feb 27 08:08:59 crc kubenswrapper[4612]: I0227 08:08:59.883462 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-56df8fb6b7-9p274" Feb 27 08:08:59 crc kubenswrapper[4612]: I0227 08:08:59.884792 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-4ljbd" Feb 27 08:08:59 crc kubenswrapper[4612]: I0227 08:08:59.894219 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a10f6a85-63bd-4aa7-91a6-b3aec6ecb9d4-config-data\") pod \"horizon-7f9b6cd66c-5lwzw\" (UID: \"a10f6a85-63bd-4aa7-91a6-b3aec6ecb9d4\") " pod="openstack/horizon-7f9b6cd66c-5lwzw" Feb 27 08:08:59 crc kubenswrapper[4612]: I0227 08:08:59.894289 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-r8dpp"] Feb 27 08:08:59 crc kubenswrapper[4612]: I0227 08:08:59.904429 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/a10f6a85-63bd-4aa7-91a6-b3aec6ecb9d4-horizon-secret-key\") pod \"horizon-7f9b6cd66c-5lwzw\" (UID: \"a10f6a85-63bd-4aa7-91a6-b3aec6ecb9d4\") " pod="openstack/horizon-7f9b6cd66c-5lwzw" Feb 27 08:08:59 crc kubenswrapper[4612]: I0227 08:08:59.904981 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-r8dpp" Feb 27 08:08:59 crc kubenswrapper[4612]: I0227 08:08:59.912084 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-8zrcl"] Feb 27 08:08:59 crc kubenswrapper[4612]: I0227 08:08:59.913345 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-8zrcl" Feb 27 08:08:59 crc kubenswrapper[4612]: I0227 08:08:59.921759 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Feb 27 08:08:59 crc kubenswrapper[4612]: I0227 08:08:59.921921 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-msh9n" Feb 27 08:08:59 crc kubenswrapper[4612]: I0227 08:08:59.925609 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-l5z5s" Feb 27 08:08:59 crc kubenswrapper[4612]: I0227 08:08:59.940376 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Feb 27 08:08:59 crc kubenswrapper[4612]: I0227 08:08:59.978112 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/fd3bd506-aad1-468a-b469-64ad78ec33fe-ovsdbserver-sb\") pod \"dnsmasq-dns-56df8fb6b7-9p274\" (UID: \"fd3bd506-aad1-468a-b469-64ad78ec33fe\") " pod="openstack/dnsmasq-dns-56df8fb6b7-9p274" Feb 27 08:08:59 crc kubenswrapper[4612]: I0227 08:08:59.978145 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fd3bd506-aad1-468a-b469-64ad78ec33fe-dns-svc\") pod \"dnsmasq-dns-56df8fb6b7-9p274\" (UID: \"fd3bd506-aad1-468a-b469-64ad78ec33fe\") " pod="openstack/dnsmasq-dns-56df8fb6b7-9p274" Feb 27 08:08:59 crc kubenswrapper[4612]: I0227 08:08:59.978166 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fd3bd506-aad1-468a-b469-64ad78ec33fe-config\") pod \"dnsmasq-dns-56df8fb6b7-9p274\" (UID: \"fd3bd506-aad1-468a-b469-64ad78ec33fe\") " pod="openstack/dnsmasq-dns-56df8fb6b7-9p274" Feb 27 08:08:59 crc kubenswrapper[4612]: I0227 08:08:59.978251 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/53ff41b8-ca85-4358-a8fe-cb490ec40793-config-data\") pod \"horizon-58b79554cc-f9ms6\" (UID: \"53ff41b8-ca85-4358-a8fe-cb490ec40793\") " pod="openstack/horizon-58b79554cc-f9ms6" Feb 27 08:08:59 crc kubenswrapper[4612]: I0227 08:08:59.978272 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/53ff41b8-ca85-4358-a8fe-cb490ec40793-logs\") pod \"horizon-58b79554cc-f9ms6\" (UID: \"53ff41b8-ca85-4358-a8fe-cb490ec40793\") " pod="openstack/horizon-58b79554cc-f9ms6" Feb 27 08:08:59 crc kubenswrapper[4612]: I0227 08:08:59.978286 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/53ff41b8-ca85-4358-a8fe-cb490ec40793-horizon-secret-key\") pod \"horizon-58b79554cc-f9ms6\" (UID: \"53ff41b8-ca85-4358-a8fe-cb490ec40793\") " pod="openstack/horizon-58b79554cc-f9ms6" Feb 27 08:08:59 crc kubenswrapper[4612]: I0227 08:08:59.978301 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fd3bd506-aad1-468a-b469-64ad78ec33fe-ovsdbserver-nb\") pod \"dnsmasq-dns-56df8fb6b7-9p274\" (UID: \"fd3bd506-aad1-468a-b469-64ad78ec33fe\") " pod="openstack/dnsmasq-dns-56df8fb6b7-9p274" Feb 27 08:08:59 crc kubenswrapper[4612]: I0227 08:08:59.978339 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rm4pj\" (UniqueName: \"kubernetes.io/projected/fd3bd506-aad1-468a-b469-64ad78ec33fe-kube-api-access-rm4pj\") pod \"dnsmasq-dns-56df8fb6b7-9p274\" (UID: \"fd3bd506-aad1-468a-b469-64ad78ec33fe\") " pod="openstack/dnsmasq-dns-56df8fb6b7-9p274" Feb 27 08:08:59 crc kubenswrapper[4612]: I0227 08:08:59.978371 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t2rpv\" (UniqueName: \"kubernetes.io/projected/53ff41b8-ca85-4358-a8fe-cb490ec40793-kube-api-access-t2rpv\") pod \"horizon-58b79554cc-f9ms6\" (UID: \"53ff41b8-ca85-4358-a8fe-cb490ec40793\") " pod="openstack/horizon-58b79554cc-f9ms6" Feb 27 08:08:59 crc kubenswrapper[4612]: I0227 08:08:59.978389 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/87df2a96-7ae3-47af-92bb-97db038b957c-db-sync-config-data\") pod \"barbican-db-sync-r8dpp\" (UID: \"87df2a96-7ae3-47af-92bb-97db038b957c\") " pod="openstack/barbican-db-sync-r8dpp" Feb 27 08:08:59 crc kubenswrapper[4612]: I0227 08:08:59.978427 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/53ff41b8-ca85-4358-a8fe-cb490ec40793-scripts\") pod \"horizon-58b79554cc-f9ms6\" (UID: \"53ff41b8-ca85-4358-a8fe-cb490ec40793\") " pod="openstack/horizon-58b79554cc-f9ms6" Feb 27 08:08:59 crc kubenswrapper[4612]: I0227 08:08:59.978445 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/fd3bd506-aad1-468a-b469-64ad78ec33fe-dns-swift-storage-0\") pod \"dnsmasq-dns-56df8fb6b7-9p274\" (UID: \"fd3bd506-aad1-468a-b469-64ad78ec33fe\") " pod="openstack/dnsmasq-dns-56df8fb6b7-9p274" Feb 27 08:08:59 crc kubenswrapper[4612]: I0227 08:08:59.978468 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-625qb\" (UniqueName: \"kubernetes.io/projected/87df2a96-7ae3-47af-92bb-97db038b957c-kube-api-access-625qb\") pod \"barbican-db-sync-r8dpp\" (UID: \"87df2a96-7ae3-47af-92bb-97db038b957c\") " pod="openstack/barbican-db-sync-r8dpp" Feb 27 08:08:59 crc kubenswrapper[4612]: I0227 08:08:59.978496 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87df2a96-7ae3-47af-92bb-97db038b957c-combined-ca-bundle\") pod \"barbican-db-sync-r8dpp\" (UID: \"87df2a96-7ae3-47af-92bb-97db038b957c\") " pod="openstack/barbican-db-sync-r8dpp" Feb 27 08:08:59 crc kubenswrapper[4612]: I0227 08:08:59.980120 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/53ff41b8-ca85-4358-a8fe-cb490ec40793-logs\") pod \"horizon-58b79554cc-f9ms6\" (UID: \"53ff41b8-ca85-4358-a8fe-cb490ec40793\") " pod="openstack/horizon-58b79554cc-f9ms6" Feb 27 08:08:59 crc kubenswrapper[4612]: I0227 08:08:59.980205 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/53ff41b8-ca85-4358-a8fe-cb490ec40793-scripts\") pod \"horizon-58b79554cc-f9ms6\" (UID: \"53ff41b8-ca85-4358-a8fe-cb490ec40793\") " pod="openstack/horizon-58b79554cc-f9ms6" Feb 27 08:08:59 crc kubenswrapper[4612]: I0227 08:08:59.980908 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/53ff41b8-ca85-4358-a8fe-cb490ec40793-config-data\") pod \"horizon-58b79554cc-f9ms6\" (UID: \"53ff41b8-ca85-4358-a8fe-cb490ec40793\") " pod="openstack/horizon-58b79554cc-f9ms6" Feb 27 08:08:59 crc kubenswrapper[4612]: I0227 08:08:59.987760 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/53ff41b8-ca85-4358-a8fe-cb490ec40793-horizon-secret-key\") pod \"horizon-58b79554cc-f9ms6\" (UID: \"53ff41b8-ca85-4358-a8fe-cb490ec40793\") " pod="openstack/horizon-58b79554cc-f9ms6" Feb 27 08:09:00 crc kubenswrapper[4612]: I0227 08:08:59.999344 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-n7qd9" Feb 27 08:09:00 crc kubenswrapper[4612]: I0227 08:08:59.999578 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Feb 27 08:09:00 crc kubenswrapper[4612]: I0227 08:09:00.026429 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-r8dpp"] Feb 27 08:09:00 crc kubenswrapper[4612]: I0227 08:09:00.029369 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9zthz\" (UniqueName: \"kubernetes.io/projected/a10f6a85-63bd-4aa7-91a6-b3aec6ecb9d4-kube-api-access-9zthz\") pod \"horizon-7f9b6cd66c-5lwzw\" (UID: \"a10f6a85-63bd-4aa7-91a6-b3aec6ecb9d4\") " pod="openstack/horizon-7f9b6cd66c-5lwzw" Feb 27 08:09:00 crc kubenswrapper[4612]: I0227 08:09:00.030078 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7f9b6cd66c-5lwzw" Feb 27 08:09:00 crc kubenswrapper[4612]: I0227 08:09:00.053760 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t2rpv\" (UniqueName: \"kubernetes.io/projected/53ff41b8-ca85-4358-a8fe-cb490ec40793-kube-api-access-t2rpv\") pod \"horizon-58b79554cc-f9ms6\" (UID: \"53ff41b8-ca85-4358-a8fe-cb490ec40793\") " pod="openstack/horizon-58b79554cc-f9ms6" Feb 27 08:09:00 crc kubenswrapper[4612]: I0227 08:09:00.053839 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-56df8fb6b7-9p274"] Feb 27 08:09:00 crc kubenswrapper[4612]: I0227 08:09:00.080758 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-8zrcl"] Feb 27 08:09:00 crc kubenswrapper[4612]: I0227 08:09:00.081686 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/87df2a96-7ae3-47af-92bb-97db038b957c-db-sync-config-data\") pod \"barbican-db-sync-r8dpp\" (UID: \"87df2a96-7ae3-47af-92bb-97db038b957c\") " pod="openstack/barbican-db-sync-r8dpp" Feb 27 08:09:00 crc kubenswrapper[4612]: I0227 08:09:00.081742 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/fd3bd506-aad1-468a-b469-64ad78ec33fe-dns-swift-storage-0\") pod \"dnsmasq-dns-56df8fb6b7-9p274\" (UID: \"fd3bd506-aad1-468a-b469-64ad78ec33fe\") " pod="openstack/dnsmasq-dns-56df8fb6b7-9p274" Feb 27 08:09:00 crc kubenswrapper[4612]: I0227 08:09:00.081783 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-625qb\" (UniqueName: \"kubernetes.io/projected/87df2a96-7ae3-47af-92bb-97db038b957c-kube-api-access-625qb\") pod \"barbican-db-sync-r8dpp\" (UID: \"87df2a96-7ae3-47af-92bb-97db038b957c\") " pod="openstack/barbican-db-sync-r8dpp" Feb 27 08:09:00 crc kubenswrapper[4612]: I0227 08:09:00.081821 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87df2a96-7ae3-47af-92bb-97db038b957c-combined-ca-bundle\") pod \"barbican-db-sync-r8dpp\" (UID: \"87df2a96-7ae3-47af-92bb-97db038b957c\") " pod="openstack/barbican-db-sync-r8dpp" Feb 27 08:09:00 crc kubenswrapper[4612]: I0227 08:09:00.081907 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/fd3bd506-aad1-468a-b469-64ad78ec33fe-ovsdbserver-sb\") pod \"dnsmasq-dns-56df8fb6b7-9p274\" (UID: \"fd3bd506-aad1-468a-b469-64ad78ec33fe\") " pod="openstack/dnsmasq-dns-56df8fb6b7-9p274" Feb 27 08:09:00 crc kubenswrapper[4612]: I0227 08:09:00.081934 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fd3bd506-aad1-468a-b469-64ad78ec33fe-dns-svc\") pod \"dnsmasq-dns-56df8fb6b7-9p274\" (UID: \"fd3bd506-aad1-468a-b469-64ad78ec33fe\") " pod="openstack/dnsmasq-dns-56df8fb6b7-9p274" Feb 27 08:09:00 crc kubenswrapper[4612]: I0227 08:09:00.081958 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fd3bd506-aad1-468a-b469-64ad78ec33fe-config\") pod \"dnsmasq-dns-56df8fb6b7-9p274\" (UID: \"fd3bd506-aad1-468a-b469-64ad78ec33fe\") " pod="openstack/dnsmasq-dns-56df8fb6b7-9p274" Feb 27 08:09:00 crc kubenswrapper[4612]: I0227 08:09:00.082056 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fd3bd506-aad1-468a-b469-64ad78ec33fe-ovsdbserver-nb\") pod \"dnsmasq-dns-56df8fb6b7-9p274\" (UID: \"fd3bd506-aad1-468a-b469-64ad78ec33fe\") " pod="openstack/dnsmasq-dns-56df8fb6b7-9p274" Feb 27 08:09:00 crc kubenswrapper[4612]: I0227 08:09:00.082091 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rm4pj\" (UniqueName: \"kubernetes.io/projected/fd3bd506-aad1-468a-b469-64ad78ec33fe-kube-api-access-rm4pj\") pod \"dnsmasq-dns-56df8fb6b7-9p274\" (UID: \"fd3bd506-aad1-468a-b469-64ad78ec33fe\") " pod="openstack/dnsmasq-dns-56df8fb6b7-9p274" Feb 27 08:09:00 crc kubenswrapper[4612]: I0227 08:09:00.087514 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/87df2a96-7ae3-47af-92bb-97db038b957c-db-sync-config-data\") pod \"barbican-db-sync-r8dpp\" (UID: \"87df2a96-7ae3-47af-92bb-97db038b957c\") " pod="openstack/barbican-db-sync-r8dpp" Feb 27 08:09:00 crc kubenswrapper[4612]: I0227 08:09:00.088178 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/fd3bd506-aad1-468a-b469-64ad78ec33fe-dns-swift-storage-0\") pod \"dnsmasq-dns-56df8fb6b7-9p274\" (UID: \"fd3bd506-aad1-468a-b469-64ad78ec33fe\") " pod="openstack/dnsmasq-dns-56df8fb6b7-9p274" Feb 27 08:09:00 crc kubenswrapper[4612]: I0227 08:09:00.090592 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fd3bd506-aad1-468a-b469-64ad78ec33fe-dns-svc\") pod \"dnsmasq-dns-56df8fb6b7-9p274\" (UID: \"fd3bd506-aad1-468a-b469-64ad78ec33fe\") " pod="openstack/dnsmasq-dns-56df8fb6b7-9p274" Feb 27 08:09:00 crc kubenswrapper[4612]: I0227 08:09:00.090954 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fd3bd506-aad1-468a-b469-64ad78ec33fe-config\") pod \"dnsmasq-dns-56df8fb6b7-9p274\" (UID: \"fd3bd506-aad1-468a-b469-64ad78ec33fe\") " pod="openstack/dnsmasq-dns-56df8fb6b7-9p274" Feb 27 08:09:00 crc kubenswrapper[4612]: I0227 08:09:00.092082 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/fd3bd506-aad1-468a-b469-64ad78ec33fe-ovsdbserver-sb\") pod \"dnsmasq-dns-56df8fb6b7-9p274\" (UID: \"fd3bd506-aad1-468a-b469-64ad78ec33fe\") " pod="openstack/dnsmasq-dns-56df8fb6b7-9p274" Feb 27 08:09:00 crc kubenswrapper[4612]: I0227 08:09:00.093029 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fd3bd506-aad1-468a-b469-64ad78ec33fe-ovsdbserver-nb\") pod \"dnsmasq-dns-56df8fb6b7-9p274\" (UID: \"fd3bd506-aad1-468a-b469-64ad78ec33fe\") " pod="openstack/dnsmasq-dns-56df8fb6b7-9p274" Feb 27 08:09:00 crc kubenswrapper[4612]: I0227 08:09:00.101361 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87df2a96-7ae3-47af-92bb-97db038b957c-combined-ca-bundle\") pod \"barbican-db-sync-r8dpp\" (UID: \"87df2a96-7ae3-47af-92bb-97db038b957c\") " pod="openstack/barbican-db-sync-r8dpp" Feb 27 08:09:00 crc kubenswrapper[4612]: I0227 08:09:00.122157 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-58b79554cc-f9ms6" Feb 27 08:09:00 crc kubenswrapper[4612]: I0227 08:09:00.128651 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rm4pj\" (UniqueName: \"kubernetes.io/projected/fd3bd506-aad1-468a-b469-64ad78ec33fe-kube-api-access-rm4pj\") pod \"dnsmasq-dns-56df8fb6b7-9p274\" (UID: \"fd3bd506-aad1-468a-b469-64ad78ec33fe\") " pod="openstack/dnsmasq-dns-56df8fb6b7-9p274" Feb 27 08:09:00 crc kubenswrapper[4612]: I0227 08:09:00.129120 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-625qb\" (UniqueName: \"kubernetes.io/projected/87df2a96-7ae3-47af-92bb-97db038b957c-kube-api-access-625qb\") pod \"barbican-db-sync-r8dpp\" (UID: \"87df2a96-7ae3-47af-92bb-97db038b957c\") " pod="openstack/barbican-db-sync-r8dpp" Feb 27 08:09:00 crc kubenswrapper[4612]: I0227 08:09:00.186858 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/36b549a3-c6e9-4f7e-8cff-85216c8e0e3d-config-data\") pod \"placement-db-sync-8zrcl\" (UID: \"36b549a3-c6e9-4f7e-8cff-85216c8e0e3d\") " pod="openstack/placement-db-sync-8zrcl" Feb 27 08:09:00 crc kubenswrapper[4612]: I0227 08:09:00.186948 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36b549a3-c6e9-4f7e-8cff-85216c8e0e3d-combined-ca-bundle\") pod \"placement-db-sync-8zrcl\" (UID: \"36b549a3-c6e9-4f7e-8cff-85216c8e0e3d\") " pod="openstack/placement-db-sync-8zrcl" Feb 27 08:09:00 crc kubenswrapper[4612]: I0227 08:09:00.186983 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qncrt\" (UniqueName: \"kubernetes.io/projected/36b549a3-c6e9-4f7e-8cff-85216c8e0e3d-kube-api-access-qncrt\") pod \"placement-db-sync-8zrcl\" (UID: \"36b549a3-c6e9-4f7e-8cff-85216c8e0e3d\") " pod="openstack/placement-db-sync-8zrcl" Feb 27 08:09:00 crc kubenswrapper[4612]: I0227 08:09:00.187025 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/36b549a3-c6e9-4f7e-8cff-85216c8e0e3d-logs\") pod \"placement-db-sync-8zrcl\" (UID: \"36b549a3-c6e9-4f7e-8cff-85216c8e0e3d\") " pod="openstack/placement-db-sync-8zrcl" Feb 27 08:09:00 crc kubenswrapper[4612]: I0227 08:09:00.187104 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/36b549a3-c6e9-4f7e-8cff-85216c8e0e3d-scripts\") pod \"placement-db-sync-8zrcl\" (UID: \"36b549a3-c6e9-4f7e-8cff-85216c8e0e3d\") " pod="openstack/placement-db-sync-8zrcl" Feb 27 08:09:00 crc kubenswrapper[4612]: I0227 08:09:00.202913 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Feb 27 08:09:00 crc kubenswrapper[4612]: I0227 08:09:00.204216 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 27 08:09:00 crc kubenswrapper[4612]: I0227 08:09:00.215573 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Feb 27 08:09:00 crc kubenswrapper[4612]: I0227 08:09:00.215731 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Feb 27 08:09:00 crc kubenswrapper[4612]: I0227 08:09:00.215965 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Feb 27 08:09:00 crc kubenswrapper[4612]: I0227 08:09:00.216084 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-55plp" Feb 27 08:09:00 crc kubenswrapper[4612]: I0227 08:09:00.245154 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-56df8fb6b7-9p274" Feb 27 08:09:00 crc kubenswrapper[4612]: I0227 08:09:00.273304 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Feb 27 08:09:00 crc kubenswrapper[4612]: I0227 08:09:00.275248 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 27 08:09:00 crc kubenswrapper[4612]: I0227 08:09:00.280260 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Feb 27 08:09:00 crc kubenswrapper[4612]: I0227 08:09:00.280427 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Feb 27 08:09:00 crc kubenswrapper[4612]: I0227 08:09:00.280943 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-r8dpp" Feb 27 08:09:00 crc kubenswrapper[4612]: I0227 08:09:00.290203 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/36b549a3-c6e9-4f7e-8cff-85216c8e0e3d-scripts\") pod \"placement-db-sync-8zrcl\" (UID: \"36b549a3-c6e9-4f7e-8cff-85216c8e0e3d\") " pod="openstack/placement-db-sync-8zrcl" Feb 27 08:09:00 crc kubenswrapper[4612]: I0227 08:09:00.290260 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/36b549a3-c6e9-4f7e-8cff-85216c8e0e3d-config-data\") pod \"placement-db-sync-8zrcl\" (UID: \"36b549a3-c6e9-4f7e-8cff-85216c8e0e3d\") " pod="openstack/placement-db-sync-8zrcl" Feb 27 08:09:00 crc kubenswrapper[4612]: I0227 08:09:00.290304 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36b549a3-c6e9-4f7e-8cff-85216c8e0e3d-combined-ca-bundle\") pod \"placement-db-sync-8zrcl\" (UID: \"36b549a3-c6e9-4f7e-8cff-85216c8e0e3d\") " pod="openstack/placement-db-sync-8zrcl" Feb 27 08:09:00 crc kubenswrapper[4612]: I0227 08:09:00.290339 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qncrt\" (UniqueName: \"kubernetes.io/projected/36b549a3-c6e9-4f7e-8cff-85216c8e0e3d-kube-api-access-qncrt\") pod \"placement-db-sync-8zrcl\" (UID: \"36b549a3-c6e9-4f7e-8cff-85216c8e0e3d\") " pod="openstack/placement-db-sync-8zrcl" Feb 27 08:09:00 crc kubenswrapper[4612]: I0227 08:09:00.290369 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/36b549a3-c6e9-4f7e-8cff-85216c8e0e3d-logs\") pod \"placement-db-sync-8zrcl\" (UID: \"36b549a3-c6e9-4f7e-8cff-85216c8e0e3d\") " pod="openstack/placement-db-sync-8zrcl" Feb 27 08:09:00 crc kubenswrapper[4612]: I0227 08:09:00.290824 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/36b549a3-c6e9-4f7e-8cff-85216c8e0e3d-logs\") pod \"placement-db-sync-8zrcl\" (UID: \"36b549a3-c6e9-4f7e-8cff-85216c8e0e3d\") " pod="openstack/placement-db-sync-8zrcl" Feb 27 08:09:00 crc kubenswrapper[4612]: I0227 08:09:00.298904 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/36b549a3-c6e9-4f7e-8cff-85216c8e0e3d-scripts\") pod \"placement-db-sync-8zrcl\" (UID: \"36b549a3-c6e9-4f7e-8cff-85216c8e0e3d\") " pod="openstack/placement-db-sync-8zrcl" Feb 27 08:09:00 crc kubenswrapper[4612]: I0227 08:09:00.312533 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/36b549a3-c6e9-4f7e-8cff-85216c8e0e3d-config-data\") pod \"placement-db-sync-8zrcl\" (UID: \"36b549a3-c6e9-4f7e-8cff-85216c8e0e3d\") " pod="openstack/placement-db-sync-8zrcl" Feb 27 08:09:00 crc kubenswrapper[4612]: I0227 08:09:00.348089 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qncrt\" (UniqueName: \"kubernetes.io/projected/36b549a3-c6e9-4f7e-8cff-85216c8e0e3d-kube-api-access-qncrt\") pod \"placement-db-sync-8zrcl\" (UID: \"36b549a3-c6e9-4f7e-8cff-85216c8e0e3d\") " pod="openstack/placement-db-sync-8zrcl" Feb 27 08:09:00 crc kubenswrapper[4612]: I0227 08:09:00.356049 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 27 08:09:00 crc kubenswrapper[4612]: I0227 08:09:00.399547 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d5de72cb-312f-4631-b2e2-e0df3f759e90-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"d5de72cb-312f-4631-b2e2-e0df3f759e90\") " pod="openstack/glance-default-external-api-0" Feb 27 08:09:00 crc kubenswrapper[4612]: I0227 08:09:00.399876 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9ab32d5a-a797-47a8-961a-a534f55fdfa7-log-httpd\") pod \"ceilometer-0\" (UID: \"9ab32d5a-a797-47a8-961a-a534f55fdfa7\") " pod="openstack/ceilometer-0" Feb 27 08:09:00 crc kubenswrapper[4612]: I0227 08:09:00.399962 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d5de72cb-312f-4631-b2e2-e0df3f759e90-scripts\") pod \"glance-default-external-api-0\" (UID: \"d5de72cb-312f-4631-b2e2-e0df3f759e90\") " pod="openstack/glance-default-external-api-0" Feb 27 08:09:00 crc kubenswrapper[4612]: I0227 08:09:00.400044 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d5de72cb-312f-4631-b2e2-e0df3f759e90-logs\") pod \"glance-default-external-api-0\" (UID: \"d5de72cb-312f-4631-b2e2-e0df3f759e90\") " pod="openstack/glance-default-external-api-0" Feb 27 08:09:00 crc kubenswrapper[4612]: I0227 08:09:00.400109 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-external-api-0\" (UID: \"d5de72cb-312f-4631-b2e2-e0df3f759e90\") " pod="openstack/glance-default-external-api-0" Feb 27 08:09:00 crc kubenswrapper[4612]: I0227 08:09:00.400181 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9ab32d5a-a797-47a8-961a-a534f55fdfa7-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"9ab32d5a-a797-47a8-961a-a534f55fdfa7\") " pod="openstack/ceilometer-0" Feb 27 08:09:00 crc kubenswrapper[4612]: I0227 08:09:00.400257 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sh9th\" (UniqueName: \"kubernetes.io/projected/9ab32d5a-a797-47a8-961a-a534f55fdfa7-kube-api-access-sh9th\") pod \"ceilometer-0\" (UID: \"9ab32d5a-a797-47a8-961a-a534f55fdfa7\") " pod="openstack/ceilometer-0" Feb 27 08:09:00 crc kubenswrapper[4612]: I0227 08:09:00.400324 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9ab32d5a-a797-47a8-961a-a534f55fdfa7-run-httpd\") pod \"ceilometer-0\" (UID: \"9ab32d5a-a797-47a8-961a-a534f55fdfa7\") " pod="openstack/ceilometer-0" Feb 27 08:09:00 crc kubenswrapper[4612]: I0227 08:09:00.400415 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h7q6m\" (UniqueName: \"kubernetes.io/projected/d5de72cb-312f-4631-b2e2-e0df3f759e90-kube-api-access-h7q6m\") pod \"glance-default-external-api-0\" (UID: \"d5de72cb-312f-4631-b2e2-e0df3f759e90\") " pod="openstack/glance-default-external-api-0" Feb 27 08:09:00 crc kubenswrapper[4612]: I0227 08:09:00.400480 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d5de72cb-312f-4631-b2e2-e0df3f759e90-config-data\") pod \"glance-default-external-api-0\" (UID: \"d5de72cb-312f-4631-b2e2-e0df3f759e90\") " pod="openstack/glance-default-external-api-0" Feb 27 08:09:00 crc kubenswrapper[4612]: I0227 08:09:00.400542 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d5de72cb-312f-4631-b2e2-e0df3f759e90-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"d5de72cb-312f-4631-b2e2-e0df3f759e90\") " pod="openstack/glance-default-external-api-0" Feb 27 08:09:00 crc kubenswrapper[4612]: I0227 08:09:00.400628 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d5de72cb-312f-4631-b2e2-e0df3f759e90-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"d5de72cb-312f-4631-b2e2-e0df3f759e90\") " pod="openstack/glance-default-external-api-0" Feb 27 08:09:00 crc kubenswrapper[4612]: I0227 08:09:00.400718 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9ab32d5a-a797-47a8-961a-a534f55fdfa7-scripts\") pod \"ceilometer-0\" (UID: \"9ab32d5a-a797-47a8-961a-a534f55fdfa7\") " pod="openstack/ceilometer-0" Feb 27 08:09:00 crc kubenswrapper[4612]: I0227 08:09:00.400812 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9ab32d5a-a797-47a8-961a-a534f55fdfa7-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"9ab32d5a-a797-47a8-961a-a534f55fdfa7\") " pod="openstack/ceilometer-0" Feb 27 08:09:00 crc kubenswrapper[4612]: I0227 08:09:00.400882 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9ab32d5a-a797-47a8-961a-a534f55fdfa7-config-data\") pod \"ceilometer-0\" (UID: \"9ab32d5a-a797-47a8-961a-a534f55fdfa7\") " pod="openstack/ceilometer-0" Feb 27 08:09:00 crc kubenswrapper[4612]: I0227 08:09:00.406927 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36b549a3-c6e9-4f7e-8cff-85216c8e0e3d-combined-ca-bundle\") pod \"placement-db-sync-8zrcl\" (UID: \"36b549a3-c6e9-4f7e-8cff-85216c8e0e3d\") " pod="openstack/placement-db-sync-8zrcl" Feb 27 08:09:00 crc kubenswrapper[4612]: I0227 08:09:00.430686 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 27 08:09:00 crc kubenswrapper[4612]: I0227 08:09:00.493117 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 27 08:09:00 crc kubenswrapper[4612]: I0227 08:09:00.495745 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 27 08:09:00 crc kubenswrapper[4612]: I0227 08:09:00.506216 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d5de72cb-312f-4631-b2e2-e0df3f759e90-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"d5de72cb-312f-4631-b2e2-e0df3f759e90\") " pod="openstack/glance-default-external-api-0" Feb 27 08:09:00 crc kubenswrapper[4612]: I0227 08:09:00.506425 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9ab32d5a-a797-47a8-961a-a534f55fdfa7-scripts\") pod \"ceilometer-0\" (UID: \"9ab32d5a-a797-47a8-961a-a534f55fdfa7\") " pod="openstack/ceilometer-0" Feb 27 08:09:00 crc kubenswrapper[4612]: I0227 08:09:00.506525 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9ab32d5a-a797-47a8-961a-a534f55fdfa7-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"9ab32d5a-a797-47a8-961a-a534f55fdfa7\") " pod="openstack/ceilometer-0" Feb 27 08:09:00 crc kubenswrapper[4612]: I0227 08:09:00.506600 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9ab32d5a-a797-47a8-961a-a534f55fdfa7-config-data\") pod \"ceilometer-0\" (UID: \"9ab32d5a-a797-47a8-961a-a534f55fdfa7\") " pod="openstack/ceilometer-0" Feb 27 08:09:00 crc kubenswrapper[4612]: I0227 08:09:00.507678 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d5de72cb-312f-4631-b2e2-e0df3f759e90-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"d5de72cb-312f-4631-b2e2-e0df3f759e90\") " pod="openstack/glance-default-external-api-0" Feb 27 08:09:00 crc kubenswrapper[4612]: I0227 08:09:00.507876 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9ab32d5a-a797-47a8-961a-a534f55fdfa7-log-httpd\") pod \"ceilometer-0\" (UID: \"9ab32d5a-a797-47a8-961a-a534f55fdfa7\") " pod="openstack/ceilometer-0" Feb 27 08:09:00 crc kubenswrapper[4612]: I0227 08:09:00.508194 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d5de72cb-312f-4631-b2e2-e0df3f759e90-scripts\") pod \"glance-default-external-api-0\" (UID: \"d5de72cb-312f-4631-b2e2-e0df3f759e90\") " pod="openstack/glance-default-external-api-0" Feb 27 08:09:00 crc kubenswrapper[4612]: I0227 08:09:00.508352 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d5de72cb-312f-4631-b2e2-e0df3f759e90-logs\") pod \"glance-default-external-api-0\" (UID: \"d5de72cb-312f-4631-b2e2-e0df3f759e90\") " pod="openstack/glance-default-external-api-0" Feb 27 08:09:00 crc kubenswrapper[4612]: I0227 08:09:00.508461 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-external-api-0\" (UID: \"d5de72cb-312f-4631-b2e2-e0df3f759e90\") " pod="openstack/glance-default-external-api-0" Feb 27 08:09:00 crc kubenswrapper[4612]: I0227 08:09:00.508598 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9ab32d5a-a797-47a8-961a-a534f55fdfa7-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"9ab32d5a-a797-47a8-961a-a534f55fdfa7\") " pod="openstack/ceilometer-0" Feb 27 08:09:00 crc kubenswrapper[4612]: I0227 08:09:00.508740 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sh9th\" (UniqueName: \"kubernetes.io/projected/9ab32d5a-a797-47a8-961a-a534f55fdfa7-kube-api-access-sh9th\") pod \"ceilometer-0\" (UID: \"9ab32d5a-a797-47a8-961a-a534f55fdfa7\") " pod="openstack/ceilometer-0" Feb 27 08:09:00 crc kubenswrapper[4612]: I0227 08:09:00.508846 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9ab32d5a-a797-47a8-961a-a534f55fdfa7-run-httpd\") pod \"ceilometer-0\" (UID: \"9ab32d5a-a797-47a8-961a-a534f55fdfa7\") " pod="openstack/ceilometer-0" Feb 27 08:09:00 crc kubenswrapper[4612]: I0227 08:09:00.509058 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h7q6m\" (UniqueName: \"kubernetes.io/projected/d5de72cb-312f-4631-b2e2-e0df3f759e90-kube-api-access-h7q6m\") pod \"glance-default-external-api-0\" (UID: \"d5de72cb-312f-4631-b2e2-e0df3f759e90\") " pod="openstack/glance-default-external-api-0" Feb 27 08:09:00 crc kubenswrapper[4612]: I0227 08:09:00.509152 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d5de72cb-312f-4631-b2e2-e0df3f759e90-config-data\") pod \"glance-default-external-api-0\" (UID: \"d5de72cb-312f-4631-b2e2-e0df3f759e90\") " pod="openstack/glance-default-external-api-0" Feb 27 08:09:00 crc kubenswrapper[4612]: I0227 08:09:00.509251 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d5de72cb-312f-4631-b2e2-e0df3f759e90-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"d5de72cb-312f-4631-b2e2-e0df3f759e90\") " pod="openstack/glance-default-external-api-0" Feb 27 08:09:00 crc kubenswrapper[4612]: I0227 08:09:00.525425 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Feb 27 08:09:00 crc kubenswrapper[4612]: I0227 08:09:00.525601 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Feb 27 08:09:00 crc kubenswrapper[4612]: I0227 08:09:00.539962 4612 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-external-api-0\" (UID: \"d5de72cb-312f-4631-b2e2-e0df3f759e90\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/glance-default-external-api-0" Feb 27 08:09:00 crc kubenswrapper[4612]: I0227 08:09:00.549974 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d5de72cb-312f-4631-b2e2-e0df3f759e90-scripts\") pod \"glance-default-external-api-0\" (UID: \"d5de72cb-312f-4631-b2e2-e0df3f759e90\") " pod="openstack/glance-default-external-api-0" Feb 27 08:09:00 crc kubenswrapper[4612]: I0227 08:09:00.554413 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d5de72cb-312f-4631-b2e2-e0df3f759e90-logs\") pod \"glance-default-external-api-0\" (UID: \"d5de72cb-312f-4631-b2e2-e0df3f759e90\") " pod="openstack/glance-default-external-api-0" Feb 27 08:09:00 crc kubenswrapper[4612]: I0227 08:09:00.555933 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9ab32d5a-a797-47a8-961a-a534f55fdfa7-log-httpd\") pod \"ceilometer-0\" (UID: \"9ab32d5a-a797-47a8-961a-a534f55fdfa7\") " pod="openstack/ceilometer-0" Feb 27 08:09:00 crc kubenswrapper[4612]: I0227 08:09:00.556150 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9ab32d5a-a797-47a8-961a-a534f55fdfa7-run-httpd\") pod \"ceilometer-0\" (UID: \"9ab32d5a-a797-47a8-961a-a534f55fdfa7\") " pod="openstack/ceilometer-0" Feb 27 08:09:00 crc kubenswrapper[4612]: I0227 08:09:00.556738 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9ab32d5a-a797-47a8-961a-a534f55fdfa7-scripts\") pod \"ceilometer-0\" (UID: \"9ab32d5a-a797-47a8-961a-a534f55fdfa7\") " pod="openstack/ceilometer-0" Feb 27 08:09:00 crc kubenswrapper[4612]: I0227 08:09:00.560155 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d5de72cb-312f-4631-b2e2-e0df3f759e90-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"d5de72cb-312f-4631-b2e2-e0df3f759e90\") " pod="openstack/glance-default-external-api-0" Feb 27 08:09:00 crc kubenswrapper[4612]: I0227 08:09:00.560661 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d5de72cb-312f-4631-b2e2-e0df3f759e90-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"d5de72cb-312f-4631-b2e2-e0df3f759e90\") " pod="openstack/glance-default-external-api-0" Feb 27 08:09:00 crc kubenswrapper[4612]: I0227 08:09:00.561548 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9ab32d5a-a797-47a8-961a-a534f55fdfa7-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"9ab32d5a-a797-47a8-961a-a534f55fdfa7\") " pod="openstack/ceilometer-0" Feb 27 08:09:00 crc kubenswrapper[4612]: I0227 08:09:00.564561 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d5de72cb-312f-4631-b2e2-e0df3f759e90-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"d5de72cb-312f-4631-b2e2-e0df3f759e90\") " pod="openstack/glance-default-external-api-0" Feb 27 08:09:00 crc kubenswrapper[4612]: I0227 08:09:00.615626 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-8zrcl" Feb 27 08:09:00 crc kubenswrapper[4612]: I0227 08:09:00.617609 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7fc3d9c6-3a82-4dcf-bfb5-fbd7b8225f90-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"7fc3d9c6-3a82-4dcf-bfb5-fbd7b8225f90\") " pod="openstack/glance-default-internal-api-0" Feb 27 08:09:00 crc kubenswrapper[4612]: I0227 08:09:00.617640 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"7fc3d9c6-3a82-4dcf-bfb5-fbd7b8225f90\") " pod="openstack/glance-default-internal-api-0" Feb 27 08:09:00 crc kubenswrapper[4612]: I0227 08:09:00.617660 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7fc3d9c6-3a82-4dcf-bfb5-fbd7b8225f90-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"7fc3d9c6-3a82-4dcf-bfb5-fbd7b8225f90\") " pod="openstack/glance-default-internal-api-0" Feb 27 08:09:00 crc kubenswrapper[4612]: I0227 08:09:00.617726 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7fc3d9c6-3a82-4dcf-bfb5-fbd7b8225f90-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"7fc3d9c6-3a82-4dcf-bfb5-fbd7b8225f90\") " pod="openstack/glance-default-internal-api-0" Feb 27 08:09:00 crc kubenswrapper[4612]: I0227 08:09:00.617766 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7fc3d9c6-3a82-4dcf-bfb5-fbd7b8225f90-config-data\") pod \"glance-default-internal-api-0\" (UID: \"7fc3d9c6-3a82-4dcf-bfb5-fbd7b8225f90\") " pod="openstack/glance-default-internal-api-0" Feb 27 08:09:00 crc kubenswrapper[4612]: I0227 08:09:00.617787 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h7m9b\" (UniqueName: \"kubernetes.io/projected/7fc3d9c6-3a82-4dcf-bfb5-fbd7b8225f90-kube-api-access-h7m9b\") pod \"glance-default-internal-api-0\" (UID: \"7fc3d9c6-3a82-4dcf-bfb5-fbd7b8225f90\") " pod="openstack/glance-default-internal-api-0" Feb 27 08:09:00 crc kubenswrapper[4612]: I0227 08:09:00.617840 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7fc3d9c6-3a82-4dcf-bfb5-fbd7b8225f90-scripts\") pod \"glance-default-internal-api-0\" (UID: \"7fc3d9c6-3a82-4dcf-bfb5-fbd7b8225f90\") " pod="openstack/glance-default-internal-api-0" Feb 27 08:09:00 crc kubenswrapper[4612]: I0227 08:09:00.617877 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7fc3d9c6-3a82-4dcf-bfb5-fbd7b8225f90-logs\") pod \"glance-default-internal-api-0\" (UID: \"7fc3d9c6-3a82-4dcf-bfb5-fbd7b8225f90\") " pod="openstack/glance-default-internal-api-0" Feb 27 08:09:00 crc kubenswrapper[4612]: I0227 08:09:00.621124 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9ab32d5a-a797-47a8-961a-a534f55fdfa7-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"9ab32d5a-a797-47a8-961a-a534f55fdfa7\") " pod="openstack/ceilometer-0" Feb 27 08:09:00 crc kubenswrapper[4612]: I0227 08:09:00.624780 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9ab32d5a-a797-47a8-961a-a534f55fdfa7-config-data\") pod \"ceilometer-0\" (UID: \"9ab32d5a-a797-47a8-961a-a534f55fdfa7\") " pod="openstack/ceilometer-0" Feb 27 08:09:00 crc kubenswrapper[4612]: I0227 08:09:00.639275 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d5de72cb-312f-4631-b2e2-e0df3f759e90-config-data\") pod \"glance-default-external-api-0\" (UID: \"d5de72cb-312f-4631-b2e2-e0df3f759e90\") " pod="openstack/glance-default-external-api-0" Feb 27 08:09:00 crc kubenswrapper[4612]: I0227 08:09:00.639658 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 27 08:09:00 crc kubenswrapper[4612]: I0227 08:09:00.646964 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-external-api-0\" (UID: \"d5de72cb-312f-4631-b2e2-e0df3f759e90\") " pod="openstack/glance-default-external-api-0" Feb 27 08:09:00 crc kubenswrapper[4612]: I0227 08:09:00.657888 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h7q6m\" (UniqueName: \"kubernetes.io/projected/d5de72cb-312f-4631-b2e2-e0df3f759e90-kube-api-access-h7q6m\") pod \"glance-default-external-api-0\" (UID: \"d5de72cb-312f-4631-b2e2-e0df3f759e90\") " pod="openstack/glance-default-external-api-0" Feb 27 08:09:00 crc kubenswrapper[4612]: I0227 08:09:00.721583 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7fc3d9c6-3a82-4dcf-bfb5-fbd7b8225f90-logs\") pod \"glance-default-internal-api-0\" (UID: \"7fc3d9c6-3a82-4dcf-bfb5-fbd7b8225f90\") " pod="openstack/glance-default-internal-api-0" Feb 27 08:09:00 crc kubenswrapper[4612]: I0227 08:09:00.721677 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7fc3d9c6-3a82-4dcf-bfb5-fbd7b8225f90-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"7fc3d9c6-3a82-4dcf-bfb5-fbd7b8225f90\") " pod="openstack/glance-default-internal-api-0" Feb 27 08:09:00 crc kubenswrapper[4612]: I0227 08:09:00.721742 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"7fc3d9c6-3a82-4dcf-bfb5-fbd7b8225f90\") " pod="openstack/glance-default-internal-api-0" Feb 27 08:09:00 crc kubenswrapper[4612]: I0227 08:09:00.721821 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7fc3d9c6-3a82-4dcf-bfb5-fbd7b8225f90-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"7fc3d9c6-3a82-4dcf-bfb5-fbd7b8225f90\") " pod="openstack/glance-default-internal-api-0" Feb 27 08:09:00 crc kubenswrapper[4612]: I0227 08:09:00.721902 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7fc3d9c6-3a82-4dcf-bfb5-fbd7b8225f90-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"7fc3d9c6-3a82-4dcf-bfb5-fbd7b8225f90\") " pod="openstack/glance-default-internal-api-0" Feb 27 08:09:00 crc kubenswrapper[4612]: I0227 08:09:00.732298 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7fc3d9c6-3a82-4dcf-bfb5-fbd7b8225f90-config-data\") pod \"glance-default-internal-api-0\" (UID: \"7fc3d9c6-3a82-4dcf-bfb5-fbd7b8225f90\") " pod="openstack/glance-default-internal-api-0" Feb 27 08:09:00 crc kubenswrapper[4612]: I0227 08:09:00.732375 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h7m9b\" (UniqueName: \"kubernetes.io/projected/7fc3d9c6-3a82-4dcf-bfb5-fbd7b8225f90-kube-api-access-h7m9b\") pod \"glance-default-internal-api-0\" (UID: \"7fc3d9c6-3a82-4dcf-bfb5-fbd7b8225f90\") " pod="openstack/glance-default-internal-api-0" Feb 27 08:09:00 crc kubenswrapper[4612]: I0227 08:09:00.732545 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7fc3d9c6-3a82-4dcf-bfb5-fbd7b8225f90-scripts\") pod \"glance-default-internal-api-0\" (UID: \"7fc3d9c6-3a82-4dcf-bfb5-fbd7b8225f90\") " pod="openstack/glance-default-internal-api-0" Feb 27 08:09:00 crc kubenswrapper[4612]: I0227 08:09:00.736984 4612 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"7fc3d9c6-3a82-4dcf-bfb5-fbd7b8225f90\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/glance-default-internal-api-0" Feb 27 08:09:00 crc kubenswrapper[4612]: I0227 08:09:00.737471 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sh9th\" (UniqueName: \"kubernetes.io/projected/9ab32d5a-a797-47a8-961a-a534f55fdfa7-kube-api-access-sh9th\") pod \"ceilometer-0\" (UID: \"9ab32d5a-a797-47a8-961a-a534f55fdfa7\") " pod="openstack/ceilometer-0" Feb 27 08:09:00 crc kubenswrapper[4612]: I0227 08:09:00.737571 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7fc3d9c6-3a82-4dcf-bfb5-fbd7b8225f90-scripts\") pod \"glance-default-internal-api-0\" (UID: \"7fc3d9c6-3a82-4dcf-bfb5-fbd7b8225f90\") " pod="openstack/glance-default-internal-api-0" Feb 27 08:09:00 crc kubenswrapper[4612]: I0227 08:09:00.739060 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7fc3d9c6-3a82-4dcf-bfb5-fbd7b8225f90-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"7fc3d9c6-3a82-4dcf-bfb5-fbd7b8225f90\") " pod="openstack/glance-default-internal-api-0" Feb 27 08:09:00 crc kubenswrapper[4612]: I0227 08:09:00.739455 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7fc3d9c6-3a82-4dcf-bfb5-fbd7b8225f90-logs\") pod \"glance-default-internal-api-0\" (UID: \"7fc3d9c6-3a82-4dcf-bfb5-fbd7b8225f90\") " pod="openstack/glance-default-internal-api-0" Feb 27 08:09:00 crc kubenswrapper[4612]: I0227 08:09:00.739583 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7fc3d9c6-3a82-4dcf-bfb5-fbd7b8225f90-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"7fc3d9c6-3a82-4dcf-bfb5-fbd7b8225f90\") " pod="openstack/glance-default-internal-api-0" Feb 27 08:09:00 crc kubenswrapper[4612]: I0227 08:09:00.746684 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7fc3d9c6-3a82-4dcf-bfb5-fbd7b8225f90-config-data\") pod \"glance-default-internal-api-0\" (UID: \"7fc3d9c6-3a82-4dcf-bfb5-fbd7b8225f90\") " pod="openstack/glance-default-internal-api-0" Feb 27 08:09:00 crc kubenswrapper[4612]: I0227 08:09:00.750949 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7fc3d9c6-3a82-4dcf-bfb5-fbd7b8225f90-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"7fc3d9c6-3a82-4dcf-bfb5-fbd7b8225f90\") " pod="openstack/glance-default-internal-api-0" Feb 27 08:09:00 crc kubenswrapper[4612]: I0227 08:09:00.752230 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-4nm8j"] Feb 27 08:09:00 crc kubenswrapper[4612]: I0227 08:09:00.768999 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h7m9b\" (UniqueName: \"kubernetes.io/projected/7fc3d9c6-3a82-4dcf-bfb5-fbd7b8225f90-kube-api-access-h7m9b\") pod \"glance-default-internal-api-0\" (UID: \"7fc3d9c6-3a82-4dcf-bfb5-fbd7b8225f90\") " pod="openstack/glance-default-internal-api-0" Feb 27 08:09:00 crc kubenswrapper[4612]: I0227 08:09:00.832617 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"7fc3d9c6-3a82-4dcf-bfb5-fbd7b8225f90\") " pod="openstack/glance-default-internal-api-0" Feb 27 08:09:00 crc kubenswrapper[4612]: I0227 08:09:00.858484 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 27 08:09:00 crc kubenswrapper[4612]: I0227 08:09:00.874766 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-4nm8j" event={"ID":"dfebce69-1679-4abd-8b9d-270046157f42","Type":"ContainerStarted","Data":"4fdf68d99fd0d497fa371456c87bcc3cc9674d0be1100134047f677eea693c47"} Feb 27 08:09:00 crc kubenswrapper[4612]: I0227 08:09:00.924890 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 27 08:09:00 crc kubenswrapper[4612]: I0227 08:09:00.977109 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 27 08:09:01 crc kubenswrapper[4612]: I0227 08:09:01.062732 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-bbf5cc879-qrwqr"] Feb 27 08:09:01 crc kubenswrapper[4612]: I0227 08:09:01.126113 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-58b79554cc-f9ms6"] Feb 27 08:09:01 crc kubenswrapper[4612]: I0227 08:09:01.171836 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-l5z5s"] Feb 27 08:09:01 crc kubenswrapper[4612]: I0227 08:09:01.305574 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-r8dpp"] Feb 27 08:09:01 crc kubenswrapper[4612]: I0227 08:09:01.413603 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-4ljbd"] Feb 27 08:09:01 crc kubenswrapper[4612]: I0227 08:09:01.429148 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7f9b6cd66c-5lwzw"] Feb 27 08:09:01 crc kubenswrapper[4612]: I0227 08:09:01.515633 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-56df8fb6b7-9p274"] Feb 27 08:09:01 crc kubenswrapper[4612]: I0227 08:09:01.566241 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-8zrcl"] Feb 27 08:09:01 crc kubenswrapper[4612]: I0227 08:09:01.816513 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 27 08:09:01 crc kubenswrapper[4612]: W0227 08:09:01.826973 4612 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd5de72cb_312f_4631_b2e2_e0df3f759e90.slice/crio-151dfb248c456fba8792f445e32b168295c2da2652c0a16d94350cf45392c2f6 WatchSource:0}: Error finding container 151dfb248c456fba8792f445e32b168295c2da2652c0a16d94350cf45392c2f6: Status 404 returned error can't find the container with id 151dfb248c456fba8792f445e32b168295c2da2652c0a16d94350cf45392c2f6 Feb 27 08:09:01 crc kubenswrapper[4612]: I0227 08:09:01.893594 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"d5de72cb-312f-4631-b2e2-e0df3f759e90","Type":"ContainerStarted","Data":"151dfb248c456fba8792f445e32b168295c2da2652c0a16d94350cf45392c2f6"} Feb 27 08:09:01 crc kubenswrapper[4612]: I0227 08:09:01.894606 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7f9b6cd66c-5lwzw" event={"ID":"a10f6a85-63bd-4aa7-91a6-b3aec6ecb9d4","Type":"ContainerStarted","Data":"edb7c7f2a1e1708daa090ff5bc33972634b6a95993156afb29b1a25cf6beaa21"} Feb 27 08:09:01 crc kubenswrapper[4612]: W0227 08:09:01.901861 4612 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9ab32d5a_a797_47a8_961a_a534f55fdfa7.slice/crio-0674026bc8c0b7f46f8592de3f04756fc96cc88411ef5b665e4e65ca68b325b1 WatchSource:0}: Error finding container 0674026bc8c0b7f46f8592de3f04756fc96cc88411ef5b665e4e65ca68b325b1: Status 404 returned error can't find the container with id 0674026bc8c0b7f46f8592de3f04756fc96cc88411ef5b665e4e65ca68b325b1 Feb 27 08:09:01 crc kubenswrapper[4612]: I0227 08:09:01.903020 4612 generic.go:334] "Generic (PLEG): container finished" podID="0a980065-12fa-4005-9bc7-cdde0637d18e" containerID="552346eb6628f28a502393dc2fd40fdb5c913624fd6c002b985c10a0fb8d41a8" exitCode=0 Feb 27 08:09:01 crc kubenswrapper[4612]: I0227 08:09:01.903186 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bbf5cc879-qrwqr" event={"ID":"0a980065-12fa-4005-9bc7-cdde0637d18e","Type":"ContainerDied","Data":"552346eb6628f28a502393dc2fd40fdb5c913624fd6c002b985c10a0fb8d41a8"} Feb 27 08:09:01 crc kubenswrapper[4612]: I0227 08:09:01.903263 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bbf5cc879-qrwqr" event={"ID":"0a980065-12fa-4005-9bc7-cdde0637d18e","Type":"ContainerStarted","Data":"1b5437fb736591b4502a1ae18e859156022c7f2b227ab76c9fdab032d7d6e510"} Feb 27 08:09:01 crc kubenswrapper[4612]: I0227 08:09:01.906653 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-r8dpp" event={"ID":"87df2a96-7ae3-47af-92bb-97db038b957c","Type":"ContainerStarted","Data":"c19fec2225fbc4c95b2593b77eca8d9e14fa64d610e4c9180c77c7d4d60a5b8a"} Feb 27 08:09:01 crc kubenswrapper[4612]: I0227 08:09:01.914045 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-58b79554cc-f9ms6" event={"ID":"53ff41b8-ca85-4358-a8fe-cb490ec40793","Type":"ContainerStarted","Data":"5272d46ea77a4e7e73f94df6aed6387d54219176f3bcd356974de1c682c0856a"} Feb 27 08:09:01 crc kubenswrapper[4612]: I0227 08:09:01.915133 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-4ljbd" event={"ID":"f63a4b12-f7e9-43a4-91d2-fd46f65daf92","Type":"ContainerStarted","Data":"8b548933be8af10568a38da48000e3a51d2fe9ff3b04b577671c700e4196b621"} Feb 27 08:09:01 crc kubenswrapper[4612]: I0227 08:09:01.915156 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-4ljbd" event={"ID":"f63a4b12-f7e9-43a4-91d2-fd46f65daf92","Type":"ContainerStarted","Data":"e12eddf1ed7f59c3a3f7e5ee4f4984107cdd6c22f0eebd0205e1ad560b240c98"} Feb 27 08:09:01 crc kubenswrapper[4612]: I0227 08:09:01.916702 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-8zrcl" event={"ID":"36b549a3-c6e9-4f7e-8cff-85216c8e0e3d","Type":"ContainerStarted","Data":"bc5d4192eb9c649d7f8a8e0e4464574f124851dc9893ac58466888063eb6d9b4"} Feb 27 08:09:01 crc kubenswrapper[4612]: I0227 08:09:01.930428 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-l5z5s" event={"ID":"0c9482e8-9eb1-4c30-95e9-bdaa5ccd239b","Type":"ContainerStarted","Data":"90344fc51bffd173e570a5b1aa79b9bda0567598d62ff08e647ad6b0b87af68a"} Feb 27 08:09:01 crc kubenswrapper[4612]: I0227 08:09:01.940304 4612 generic.go:334] "Generic (PLEG): container finished" podID="fd3bd506-aad1-468a-b469-64ad78ec33fe" containerID="1d2b35afd6c9fee7411ec5debd6b49bc6753e4f253045d15c54fb9f8d61fd65e" exitCode=0 Feb 27 08:09:01 crc kubenswrapper[4612]: I0227 08:09:01.940399 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56df8fb6b7-9p274" event={"ID":"fd3bd506-aad1-468a-b469-64ad78ec33fe","Type":"ContainerDied","Data":"1d2b35afd6c9fee7411ec5debd6b49bc6753e4f253045d15c54fb9f8d61fd65e"} Feb 27 08:09:01 crc kubenswrapper[4612]: I0227 08:09:01.940433 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56df8fb6b7-9p274" event={"ID":"fd3bd506-aad1-468a-b469-64ad78ec33fe","Type":"ContainerStarted","Data":"ed72968c7818e0306d78da7f79912f6d1d2741262a26a0eef8e6cfbaf13ad954"} Feb 27 08:09:01 crc kubenswrapper[4612]: I0227 08:09:01.946275 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 27 08:09:01 crc kubenswrapper[4612]: I0227 08:09:01.958764 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-4nm8j" event={"ID":"dfebce69-1679-4abd-8b9d-270046157f42","Type":"ContainerStarted","Data":"9a40f17b5d0d9218b64aab79491862655274f46636610ed2f3fac9e5e60ebdf4"} Feb 27 08:09:01 crc kubenswrapper[4612]: I0227 08:09:01.977392 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-4ljbd" podStartSLOduration=2.9773760400000002 podStartE2EDuration="2.97737604s" podCreationTimestamp="2026-02-27 08:08:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 08:09:01.96409164 +0000 UTC m=+1199.818021638" watchObservedRunningTime="2026-02-27 08:09:01.97737604 +0000 UTC m=+1199.831306038" Feb 27 08:09:02 crc kubenswrapper[4612]: I0227 08:09:02.028770 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-4nm8j" podStartSLOduration=4.028747562 podStartE2EDuration="4.028747562s" podCreationTimestamp="2026-02-27 08:08:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 08:09:01.998282439 +0000 UTC m=+1199.852212437" watchObservedRunningTime="2026-02-27 08:09:02.028747562 +0000 UTC m=+1199.882677560" Feb 27 08:09:02 crc kubenswrapper[4612]: W0227 08:09:02.059193 4612 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7fc3d9c6_3a82_4dcf_bfb5_fbd7b8225f90.slice/crio-900b007f2caf8828e2d318815e2cbea0bd48685fbec9cf487f1e59bc1ed25dd1 WatchSource:0}: Error finding container 900b007f2caf8828e2d318815e2cbea0bd48685fbec9cf487f1e59bc1ed25dd1: Status 404 returned error can't find the container with id 900b007f2caf8828e2d318815e2cbea0bd48685fbec9cf487f1e59bc1ed25dd1 Feb 27 08:09:02 crc kubenswrapper[4612]: I0227 08:09:02.105680 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 27 08:09:02 crc kubenswrapper[4612]: I0227 08:09:02.174941 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 27 08:09:02 crc kubenswrapper[4612]: I0227 08:09:02.203801 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-58b79554cc-f9ms6"] Feb 27 08:09:02 crc kubenswrapper[4612]: I0227 08:09:02.248148 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-594c9f8cf-q282v"] Feb 27 08:09:02 crc kubenswrapper[4612]: I0227 08:09:02.263673 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-594c9f8cf-q282v" Feb 27 08:09:02 crc kubenswrapper[4612]: I0227 08:09:02.381763 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-594c9f8cf-q282v"] Feb 27 08:09:02 crc kubenswrapper[4612]: I0227 08:09:02.394325 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zpvms\" (UniqueName: \"kubernetes.io/projected/f37f5ce9-74c3-48fc-9080-ae81464dba49-kube-api-access-zpvms\") pod \"horizon-594c9f8cf-q282v\" (UID: \"f37f5ce9-74c3-48fc-9080-ae81464dba49\") " pod="openstack/horizon-594c9f8cf-q282v" Feb 27 08:09:02 crc kubenswrapper[4612]: I0227 08:09:02.487620 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f37f5ce9-74c3-48fc-9080-ae81464dba49-config-data\") pod \"horizon-594c9f8cf-q282v\" (UID: \"f37f5ce9-74c3-48fc-9080-ae81464dba49\") " pod="openstack/horizon-594c9f8cf-q282v" Feb 27 08:09:02 crc kubenswrapper[4612]: I0227 08:09:02.487665 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/f37f5ce9-74c3-48fc-9080-ae81464dba49-horizon-secret-key\") pod \"horizon-594c9f8cf-q282v\" (UID: \"f37f5ce9-74c3-48fc-9080-ae81464dba49\") " pod="openstack/horizon-594c9f8cf-q282v" Feb 27 08:09:02 crc kubenswrapper[4612]: I0227 08:09:02.487725 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f37f5ce9-74c3-48fc-9080-ae81464dba49-logs\") pod \"horizon-594c9f8cf-q282v\" (UID: \"f37f5ce9-74c3-48fc-9080-ae81464dba49\") " pod="openstack/horizon-594c9f8cf-q282v" Feb 27 08:09:02 crc kubenswrapper[4612]: I0227 08:09:02.487822 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f37f5ce9-74c3-48fc-9080-ae81464dba49-scripts\") pod \"horizon-594c9f8cf-q282v\" (UID: \"f37f5ce9-74c3-48fc-9080-ae81464dba49\") " pod="openstack/horizon-594c9f8cf-q282v" Feb 27 08:09:02 crc kubenswrapper[4612]: I0227 08:09:02.453880 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 27 08:09:02 crc kubenswrapper[4612]: I0227 08:09:02.488065 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 27 08:09:02 crc kubenswrapper[4612]: I0227 08:09:02.588518 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zpvms\" (UniqueName: \"kubernetes.io/projected/f37f5ce9-74c3-48fc-9080-ae81464dba49-kube-api-access-zpvms\") pod \"horizon-594c9f8cf-q282v\" (UID: \"f37f5ce9-74c3-48fc-9080-ae81464dba49\") " pod="openstack/horizon-594c9f8cf-q282v" Feb 27 08:09:02 crc kubenswrapper[4612]: I0227 08:09:02.588936 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f37f5ce9-74c3-48fc-9080-ae81464dba49-config-data\") pod \"horizon-594c9f8cf-q282v\" (UID: \"f37f5ce9-74c3-48fc-9080-ae81464dba49\") " pod="openstack/horizon-594c9f8cf-q282v" Feb 27 08:09:02 crc kubenswrapper[4612]: I0227 08:09:02.588967 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/f37f5ce9-74c3-48fc-9080-ae81464dba49-horizon-secret-key\") pod \"horizon-594c9f8cf-q282v\" (UID: \"f37f5ce9-74c3-48fc-9080-ae81464dba49\") " pod="openstack/horizon-594c9f8cf-q282v" Feb 27 08:09:02 crc kubenswrapper[4612]: I0227 08:09:02.588999 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f37f5ce9-74c3-48fc-9080-ae81464dba49-logs\") pod \"horizon-594c9f8cf-q282v\" (UID: \"f37f5ce9-74c3-48fc-9080-ae81464dba49\") " pod="openstack/horizon-594c9f8cf-q282v" Feb 27 08:09:02 crc kubenswrapper[4612]: I0227 08:09:02.589055 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f37f5ce9-74c3-48fc-9080-ae81464dba49-scripts\") pod \"horizon-594c9f8cf-q282v\" (UID: \"f37f5ce9-74c3-48fc-9080-ae81464dba49\") " pod="openstack/horizon-594c9f8cf-q282v" Feb 27 08:09:02 crc kubenswrapper[4612]: I0227 08:09:02.589925 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f37f5ce9-74c3-48fc-9080-ae81464dba49-scripts\") pod \"horizon-594c9f8cf-q282v\" (UID: \"f37f5ce9-74c3-48fc-9080-ae81464dba49\") " pod="openstack/horizon-594c9f8cf-q282v" Feb 27 08:09:02 crc kubenswrapper[4612]: I0227 08:09:02.591930 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f37f5ce9-74c3-48fc-9080-ae81464dba49-config-data\") pod \"horizon-594c9f8cf-q282v\" (UID: \"f37f5ce9-74c3-48fc-9080-ae81464dba49\") " pod="openstack/horizon-594c9f8cf-q282v" Feb 27 08:09:02 crc kubenswrapper[4612]: I0227 08:09:02.599093 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f37f5ce9-74c3-48fc-9080-ae81464dba49-logs\") pod \"horizon-594c9f8cf-q282v\" (UID: \"f37f5ce9-74c3-48fc-9080-ae81464dba49\") " pod="openstack/horizon-594c9f8cf-q282v" Feb 27 08:09:02 crc kubenswrapper[4612]: I0227 08:09:02.604730 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/f37f5ce9-74c3-48fc-9080-ae81464dba49-horizon-secret-key\") pod \"horizon-594c9f8cf-q282v\" (UID: \"f37f5ce9-74c3-48fc-9080-ae81464dba49\") " pod="openstack/horizon-594c9f8cf-q282v" Feb 27 08:09:02 crc kubenswrapper[4612]: I0227 08:09:02.641626 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zpvms\" (UniqueName: \"kubernetes.io/projected/f37f5ce9-74c3-48fc-9080-ae81464dba49-kube-api-access-zpvms\") pod \"horizon-594c9f8cf-q282v\" (UID: \"f37f5ce9-74c3-48fc-9080-ae81464dba49\") " pod="openstack/horizon-594c9f8cf-q282v" Feb 27 08:09:02 crc kubenswrapper[4612]: I0227 08:09:02.784426 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-594c9f8cf-q282v" Feb 27 08:09:02 crc kubenswrapper[4612]: I0227 08:09:02.830164 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bbf5cc879-qrwqr" Feb 27 08:09:03 crc kubenswrapper[4612]: I0227 08:09:03.003493 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0a980065-12fa-4005-9bc7-cdde0637d18e-dns-swift-storage-0\") pod \"0a980065-12fa-4005-9bc7-cdde0637d18e\" (UID: \"0a980065-12fa-4005-9bc7-cdde0637d18e\") " Feb 27 08:09:03 crc kubenswrapper[4612]: I0227 08:09:03.004593 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0a980065-12fa-4005-9bc7-cdde0637d18e-config\") pod \"0a980065-12fa-4005-9bc7-cdde0637d18e\" (UID: \"0a980065-12fa-4005-9bc7-cdde0637d18e\") " Feb 27 08:09:03 crc kubenswrapper[4612]: I0227 08:09:03.004613 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0a980065-12fa-4005-9bc7-cdde0637d18e-ovsdbserver-sb\") pod \"0a980065-12fa-4005-9bc7-cdde0637d18e\" (UID: \"0a980065-12fa-4005-9bc7-cdde0637d18e\") " Feb 27 08:09:03 crc kubenswrapper[4612]: I0227 08:09:03.004654 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-klrr2\" (UniqueName: \"kubernetes.io/projected/0a980065-12fa-4005-9bc7-cdde0637d18e-kube-api-access-klrr2\") pod \"0a980065-12fa-4005-9bc7-cdde0637d18e\" (UID: \"0a980065-12fa-4005-9bc7-cdde0637d18e\") " Feb 27 08:09:03 crc kubenswrapper[4612]: I0227 08:09:03.004803 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0a980065-12fa-4005-9bc7-cdde0637d18e-ovsdbserver-nb\") pod \"0a980065-12fa-4005-9bc7-cdde0637d18e\" (UID: \"0a980065-12fa-4005-9bc7-cdde0637d18e\") " Feb 27 08:09:03 crc kubenswrapper[4612]: I0227 08:09:03.004860 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0a980065-12fa-4005-9bc7-cdde0637d18e-dns-svc\") pod \"0a980065-12fa-4005-9bc7-cdde0637d18e\" (UID: \"0a980065-12fa-4005-9bc7-cdde0637d18e\") " Feb 27 08:09:03 crc kubenswrapper[4612]: I0227 08:09:03.030127 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0a980065-12fa-4005-9bc7-cdde0637d18e-kube-api-access-klrr2" (OuterVolumeSpecName: "kube-api-access-klrr2") pod "0a980065-12fa-4005-9bc7-cdde0637d18e" (UID: "0a980065-12fa-4005-9bc7-cdde0637d18e"). InnerVolumeSpecName "kube-api-access-klrr2". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 08:09:03 crc kubenswrapper[4612]: I0227 08:09:03.048595 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0a980065-12fa-4005-9bc7-cdde0637d18e-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "0a980065-12fa-4005-9bc7-cdde0637d18e" (UID: "0a980065-12fa-4005-9bc7-cdde0637d18e"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 08:09:03 crc kubenswrapper[4612]: I0227 08:09:03.073335 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bbf5cc879-qrwqr" event={"ID":"0a980065-12fa-4005-9bc7-cdde0637d18e","Type":"ContainerDied","Data":"1b5437fb736591b4502a1ae18e859156022c7f2b227ab76c9fdab032d7d6e510"} Feb 27 08:09:03 crc kubenswrapper[4612]: I0227 08:09:03.073382 4612 scope.go:117] "RemoveContainer" containerID="552346eb6628f28a502393dc2fd40fdb5c913624fd6c002b985c10a0fb8d41a8" Feb 27 08:09:03 crc kubenswrapper[4612]: I0227 08:09:03.073492 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bbf5cc879-qrwqr" Feb 27 08:09:03 crc kubenswrapper[4612]: I0227 08:09:03.074375 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0a980065-12fa-4005-9bc7-cdde0637d18e-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "0a980065-12fa-4005-9bc7-cdde0637d18e" (UID: "0a980065-12fa-4005-9bc7-cdde0637d18e"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 08:09:03 crc kubenswrapper[4612]: I0227 08:09:03.079341 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0a980065-12fa-4005-9bc7-cdde0637d18e-config" (OuterVolumeSpecName: "config") pod "0a980065-12fa-4005-9bc7-cdde0637d18e" (UID: "0a980065-12fa-4005-9bc7-cdde0637d18e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 08:09:03 crc kubenswrapper[4612]: I0227 08:09:03.090565 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0a980065-12fa-4005-9bc7-cdde0637d18e-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "0a980065-12fa-4005-9bc7-cdde0637d18e" (UID: "0a980065-12fa-4005-9bc7-cdde0637d18e"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 08:09:03 crc kubenswrapper[4612]: I0227 08:09:03.093363 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0a980065-12fa-4005-9bc7-cdde0637d18e-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "0a980065-12fa-4005-9bc7-cdde0637d18e" (UID: "0a980065-12fa-4005-9bc7-cdde0637d18e"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 08:09:03 crc kubenswrapper[4612]: I0227 08:09:03.107640 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"7fc3d9c6-3a82-4dcf-bfb5-fbd7b8225f90","Type":"ContainerStarted","Data":"900b007f2caf8828e2d318815e2cbea0bd48685fbec9cf487f1e59bc1ed25dd1"} Feb 27 08:09:03 crc kubenswrapper[4612]: I0227 08:09:03.108299 4612 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0a980065-12fa-4005-9bc7-cdde0637d18e-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Feb 27 08:09:03 crc kubenswrapper[4612]: I0227 08:09:03.108328 4612 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0a980065-12fa-4005-9bc7-cdde0637d18e-config\") on node \"crc\" DevicePath \"\"" Feb 27 08:09:03 crc kubenswrapper[4612]: I0227 08:09:03.108342 4612 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0a980065-12fa-4005-9bc7-cdde0637d18e-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 27 08:09:03 crc kubenswrapper[4612]: I0227 08:09:03.108355 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-klrr2\" (UniqueName: \"kubernetes.io/projected/0a980065-12fa-4005-9bc7-cdde0637d18e-kube-api-access-klrr2\") on node \"crc\" DevicePath \"\"" Feb 27 08:09:03 crc kubenswrapper[4612]: I0227 08:09:03.108368 4612 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0a980065-12fa-4005-9bc7-cdde0637d18e-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 27 08:09:03 crc kubenswrapper[4612]: I0227 08:09:03.108380 4612 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0a980065-12fa-4005-9bc7-cdde0637d18e-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 27 08:09:03 crc kubenswrapper[4612]: I0227 08:09:03.122437 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9ab32d5a-a797-47a8-961a-a534f55fdfa7","Type":"ContainerStarted","Data":"0674026bc8c0b7f46f8592de3f04756fc96cc88411ef5b665e4e65ca68b325b1"} Feb 27 08:09:03 crc kubenswrapper[4612]: I0227 08:09:03.452134 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-bbf5cc879-qrwqr"] Feb 27 08:09:03 crc kubenswrapper[4612]: I0227 08:09:03.476834 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-bbf5cc879-qrwqr"] Feb 27 08:09:03 crc kubenswrapper[4612]: I0227 08:09:03.585125 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-594c9f8cf-q282v"] Feb 27 08:09:03 crc kubenswrapper[4612]: W0227 08:09:03.625835 4612 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf37f5ce9_74c3_48fc_9080_ae81464dba49.slice/crio-96a4e58df22136ff0feebad1d8d3ec302578a66e6437cd0adfd7785d7f586ee6 WatchSource:0}: Error finding container 96a4e58df22136ff0feebad1d8d3ec302578a66e6437cd0adfd7785d7f586ee6: Status 404 returned error can't find the container with id 96a4e58df22136ff0feebad1d8d3ec302578a66e6437cd0adfd7785d7f586ee6 Feb 27 08:09:04 crc kubenswrapper[4612]: I0227 08:09:04.091151 4612 scope.go:117] "RemoveContainer" containerID="524db539480c08de1db0e7eb21a0e4cc343c22d5c6b00e8818bb1778f73f9935" Feb 27 08:09:04 crc kubenswrapper[4612]: I0227 08:09:04.156064 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"7fc3d9c6-3a82-4dcf-bfb5-fbd7b8225f90","Type":"ContainerStarted","Data":"d3d04d402df81d47178f525ec89626c4c3774ab33d06c103ccd34cdf3c8720af"} Feb 27 08:09:04 crc kubenswrapper[4612]: I0227 08:09:04.158826 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"d5de72cb-312f-4631-b2e2-e0df3f759e90","Type":"ContainerStarted","Data":"87ccc96505e0e1cc4d27fc5ae561b2d18ece432441be4afda55ac54553c69d09"} Feb 27 08:09:04 crc kubenswrapper[4612]: I0227 08:09:04.161359 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56df8fb6b7-9p274" event={"ID":"fd3bd506-aad1-468a-b469-64ad78ec33fe","Type":"ContainerStarted","Data":"10abcf416e4e5487e724027816bc29df51d2715c2f37c2ad093bd0b055e40712"} Feb 27 08:09:04 crc kubenswrapper[4612]: I0227 08:09:04.161472 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-56df8fb6b7-9p274" Feb 27 08:09:04 crc kubenswrapper[4612]: I0227 08:09:04.163473 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-594c9f8cf-q282v" event={"ID":"f37f5ce9-74c3-48fc-9080-ae81464dba49","Type":"ContainerStarted","Data":"96a4e58df22136ff0feebad1d8d3ec302578a66e6437cd0adfd7785d7f586ee6"} Feb 27 08:09:04 crc kubenswrapper[4612]: I0227 08:09:04.189394 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-56df8fb6b7-9p274" podStartSLOduration=5.189375852 podStartE2EDuration="5.189375852s" podCreationTimestamp="2026-02-27 08:08:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 08:09:04.185305256 +0000 UTC m=+1202.039235254" watchObservedRunningTime="2026-02-27 08:09:04.189375852 +0000 UTC m=+1202.043305850" Feb 27 08:09:04 crc kubenswrapper[4612]: I0227 08:09:04.872678 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0a980065-12fa-4005-9bc7-cdde0637d18e" path="/var/lib/kubelet/pods/0a980065-12fa-4005-9bc7-cdde0637d18e/volumes" Feb 27 08:09:05 crc kubenswrapper[4612]: I0227 08:09:05.210564 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"d5de72cb-312f-4631-b2e2-e0df3f759e90","Type":"ContainerStarted","Data":"c3d7e4c6e41d6a457e050942c6ac447823a5e27cf10ae87bc84dfe43e55afc39"} Feb 27 08:09:05 crc kubenswrapper[4612]: I0227 08:09:05.210800 4612 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="d5de72cb-312f-4631-b2e2-e0df3f759e90" containerName="glance-log" containerID="cri-o://87ccc96505e0e1cc4d27fc5ae561b2d18ece432441be4afda55ac54553c69d09" gracePeriod=30 Feb 27 08:09:05 crc kubenswrapper[4612]: I0227 08:09:05.210969 4612 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="d5de72cb-312f-4631-b2e2-e0df3f759e90" containerName="glance-httpd" containerID="cri-o://c3d7e4c6e41d6a457e050942c6ac447823a5e27cf10ae87bc84dfe43e55afc39" gracePeriod=30 Feb 27 08:09:06 crc kubenswrapper[4612]: I0227 08:09:06.067091 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 27 08:09:06 crc kubenswrapper[4612]: I0227 08:09:06.196337 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h7q6m\" (UniqueName: \"kubernetes.io/projected/d5de72cb-312f-4631-b2e2-e0df3f759e90-kube-api-access-h7q6m\") pod \"d5de72cb-312f-4631-b2e2-e0df3f759e90\" (UID: \"d5de72cb-312f-4631-b2e2-e0df3f759e90\") " Feb 27 08:09:06 crc kubenswrapper[4612]: I0227 08:09:06.196413 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d5de72cb-312f-4631-b2e2-e0df3f759e90-config-data\") pod \"d5de72cb-312f-4631-b2e2-e0df3f759e90\" (UID: \"d5de72cb-312f-4631-b2e2-e0df3f759e90\") " Feb 27 08:09:06 crc kubenswrapper[4612]: I0227 08:09:06.196441 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d5de72cb-312f-4631-b2e2-e0df3f759e90-logs\") pod \"d5de72cb-312f-4631-b2e2-e0df3f759e90\" (UID: \"d5de72cb-312f-4631-b2e2-e0df3f759e90\") " Feb 27 08:09:06 crc kubenswrapper[4612]: I0227 08:09:06.196518 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"d5de72cb-312f-4631-b2e2-e0df3f759e90\" (UID: \"d5de72cb-312f-4631-b2e2-e0df3f759e90\") " Feb 27 08:09:06 crc kubenswrapper[4612]: I0227 08:09:06.196555 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d5de72cb-312f-4631-b2e2-e0df3f759e90-scripts\") pod \"d5de72cb-312f-4631-b2e2-e0df3f759e90\" (UID: \"d5de72cb-312f-4631-b2e2-e0df3f759e90\") " Feb 27 08:09:06 crc kubenswrapper[4612]: I0227 08:09:06.196585 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d5de72cb-312f-4631-b2e2-e0df3f759e90-combined-ca-bundle\") pod \"d5de72cb-312f-4631-b2e2-e0df3f759e90\" (UID: \"d5de72cb-312f-4631-b2e2-e0df3f759e90\") " Feb 27 08:09:06 crc kubenswrapper[4612]: I0227 08:09:06.196614 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d5de72cb-312f-4631-b2e2-e0df3f759e90-httpd-run\") pod \"d5de72cb-312f-4631-b2e2-e0df3f759e90\" (UID: \"d5de72cb-312f-4631-b2e2-e0df3f759e90\") " Feb 27 08:09:06 crc kubenswrapper[4612]: I0227 08:09:06.196774 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d5de72cb-312f-4631-b2e2-e0df3f759e90-public-tls-certs\") pod \"d5de72cb-312f-4631-b2e2-e0df3f759e90\" (UID: \"d5de72cb-312f-4631-b2e2-e0df3f759e90\") " Feb 27 08:09:06 crc kubenswrapper[4612]: I0227 08:09:06.198461 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d5de72cb-312f-4631-b2e2-e0df3f759e90-logs" (OuterVolumeSpecName: "logs") pod "d5de72cb-312f-4631-b2e2-e0df3f759e90" (UID: "d5de72cb-312f-4631-b2e2-e0df3f759e90"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 08:09:06 crc kubenswrapper[4612]: I0227 08:09:06.198482 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d5de72cb-312f-4631-b2e2-e0df3f759e90-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "d5de72cb-312f-4631-b2e2-e0df3f759e90" (UID: "d5de72cb-312f-4631-b2e2-e0df3f759e90"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 08:09:06 crc kubenswrapper[4612]: I0227 08:09:06.205928 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage03-crc" (OuterVolumeSpecName: "glance") pod "d5de72cb-312f-4631-b2e2-e0df3f759e90" (UID: "d5de72cb-312f-4631-b2e2-e0df3f759e90"). InnerVolumeSpecName "local-storage03-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Feb 27 08:09:06 crc kubenswrapper[4612]: I0227 08:09:06.218401 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d5de72cb-312f-4631-b2e2-e0df3f759e90-scripts" (OuterVolumeSpecName: "scripts") pod "d5de72cb-312f-4631-b2e2-e0df3f759e90" (UID: "d5de72cb-312f-4631-b2e2-e0df3f759e90"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:09:06 crc kubenswrapper[4612]: I0227 08:09:06.221149 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d5de72cb-312f-4631-b2e2-e0df3f759e90-kube-api-access-h7q6m" (OuterVolumeSpecName: "kube-api-access-h7q6m") pod "d5de72cb-312f-4631-b2e2-e0df3f759e90" (UID: "d5de72cb-312f-4631-b2e2-e0df3f759e90"). InnerVolumeSpecName "kube-api-access-h7q6m". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 08:09:06 crc kubenswrapper[4612]: I0227 08:09:06.236186 4612 generic.go:334] "Generic (PLEG): container finished" podID="d5de72cb-312f-4631-b2e2-e0df3f759e90" containerID="c3d7e4c6e41d6a457e050942c6ac447823a5e27cf10ae87bc84dfe43e55afc39" exitCode=143 Feb 27 08:09:06 crc kubenswrapper[4612]: I0227 08:09:06.236221 4612 generic.go:334] "Generic (PLEG): container finished" podID="d5de72cb-312f-4631-b2e2-e0df3f759e90" containerID="87ccc96505e0e1cc4d27fc5ae561b2d18ece432441be4afda55ac54553c69d09" exitCode=143 Feb 27 08:09:06 crc kubenswrapper[4612]: I0227 08:09:06.236264 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"d5de72cb-312f-4631-b2e2-e0df3f759e90","Type":"ContainerDied","Data":"c3d7e4c6e41d6a457e050942c6ac447823a5e27cf10ae87bc84dfe43e55afc39"} Feb 27 08:09:06 crc kubenswrapper[4612]: I0227 08:09:06.236293 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"d5de72cb-312f-4631-b2e2-e0df3f759e90","Type":"ContainerDied","Data":"87ccc96505e0e1cc4d27fc5ae561b2d18ece432441be4afda55ac54553c69d09"} Feb 27 08:09:06 crc kubenswrapper[4612]: I0227 08:09:06.236303 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"d5de72cb-312f-4631-b2e2-e0df3f759e90","Type":"ContainerDied","Data":"151dfb248c456fba8792f445e32b168295c2da2652c0a16d94350cf45392c2f6"} Feb 27 08:09:06 crc kubenswrapper[4612]: I0227 08:09:06.236320 4612 scope.go:117] "RemoveContainer" containerID="c3d7e4c6e41d6a457e050942c6ac447823a5e27cf10ae87bc84dfe43e55afc39" Feb 27 08:09:06 crc kubenswrapper[4612]: I0227 08:09:06.236459 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 27 08:09:06 crc kubenswrapper[4612]: I0227 08:09:06.237668 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d5de72cb-312f-4631-b2e2-e0df3f759e90-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d5de72cb-312f-4631-b2e2-e0df3f759e90" (UID: "d5de72cb-312f-4631-b2e2-e0df3f759e90"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:09:06 crc kubenswrapper[4612]: I0227 08:09:06.250586 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"7fc3d9c6-3a82-4dcf-bfb5-fbd7b8225f90","Type":"ContainerStarted","Data":"0d79762e8bbcf38c5d056d44b9fa86200838764e4ac3355e7d590c3d228be5d1"} Feb 27 08:09:06 crc kubenswrapper[4612]: I0227 08:09:06.250778 4612 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="7fc3d9c6-3a82-4dcf-bfb5-fbd7b8225f90" containerName="glance-log" containerID="cri-o://d3d04d402df81d47178f525ec89626c4c3774ab33d06c103ccd34cdf3c8720af" gracePeriod=30 Feb 27 08:09:06 crc kubenswrapper[4612]: I0227 08:09:06.251346 4612 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="7fc3d9c6-3a82-4dcf-bfb5-fbd7b8225f90" containerName="glance-httpd" containerID="cri-o://0d79762e8bbcf38c5d056d44b9fa86200838764e4ac3355e7d590c3d228be5d1" gracePeriod=30 Feb 27 08:09:06 crc kubenswrapper[4612]: I0227 08:09:06.302254 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=6.302234425 podStartE2EDuration="6.302234425s" podCreationTimestamp="2026-02-27 08:09:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 08:09:06.287484222 +0000 UTC m=+1204.141414220" watchObservedRunningTime="2026-02-27 08:09:06.302234425 +0000 UTC m=+1204.156164413" Feb 27 08:09:06 crc kubenswrapper[4612]: I0227 08:09:06.303321 4612 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d5de72cb-312f-4631-b2e2-e0df3f759e90-logs\") on node \"crc\" DevicePath \"\"" Feb 27 08:09:06 crc kubenswrapper[4612]: I0227 08:09:06.305259 4612 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" " Feb 27 08:09:06 crc kubenswrapper[4612]: I0227 08:09:06.305272 4612 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d5de72cb-312f-4631-b2e2-e0df3f759e90-scripts\") on node \"crc\" DevicePath \"\"" Feb 27 08:09:06 crc kubenswrapper[4612]: I0227 08:09:06.305282 4612 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d5de72cb-312f-4631-b2e2-e0df3f759e90-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 08:09:06 crc kubenswrapper[4612]: I0227 08:09:06.305292 4612 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d5de72cb-312f-4631-b2e2-e0df3f759e90-httpd-run\") on node \"crc\" DevicePath \"\"" Feb 27 08:09:06 crc kubenswrapper[4612]: I0227 08:09:06.305301 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h7q6m\" (UniqueName: \"kubernetes.io/projected/d5de72cb-312f-4631-b2e2-e0df3f759e90-kube-api-access-h7q6m\") on node \"crc\" DevicePath \"\"" Feb 27 08:09:06 crc kubenswrapper[4612]: I0227 08:09:06.322545 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d5de72cb-312f-4631-b2e2-e0df3f759e90-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "d5de72cb-312f-4631-b2e2-e0df3f759e90" (UID: "d5de72cb-312f-4631-b2e2-e0df3f759e90"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:09:06 crc kubenswrapper[4612]: I0227 08:09:06.330477 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d5de72cb-312f-4631-b2e2-e0df3f759e90-config-data" (OuterVolumeSpecName: "config-data") pod "d5de72cb-312f-4631-b2e2-e0df3f759e90" (UID: "d5de72cb-312f-4631-b2e2-e0df3f759e90"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:09:06 crc kubenswrapper[4612]: I0227 08:09:06.368285 4612 scope.go:117] "RemoveContainer" containerID="87ccc96505e0e1cc4d27fc5ae561b2d18ece432441be4afda55ac54553c69d09" Feb 27 08:09:06 crc kubenswrapper[4612]: I0227 08:09:06.377115 4612 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage03-crc" (UniqueName: "kubernetes.io/local-volume/local-storage03-crc") on node "crc" Feb 27 08:09:06 crc kubenswrapper[4612]: I0227 08:09:06.411962 4612 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d5de72cb-312f-4631-b2e2-e0df3f759e90-config-data\") on node \"crc\" DevicePath \"\"" Feb 27 08:09:06 crc kubenswrapper[4612]: I0227 08:09:06.414212 4612 reconciler_common.go:293] "Volume detached for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" DevicePath \"\"" Feb 27 08:09:06 crc kubenswrapper[4612]: I0227 08:09:06.414252 4612 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d5de72cb-312f-4631-b2e2-e0df3f759e90-public-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 27 08:09:06 crc kubenswrapper[4612]: I0227 08:09:06.526019 4612 scope.go:117] "RemoveContainer" containerID="c3d7e4c6e41d6a457e050942c6ac447823a5e27cf10ae87bc84dfe43e55afc39" Feb 27 08:09:06 crc kubenswrapper[4612]: E0227 08:09:06.527506 4612 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c3d7e4c6e41d6a457e050942c6ac447823a5e27cf10ae87bc84dfe43e55afc39\": container with ID starting with c3d7e4c6e41d6a457e050942c6ac447823a5e27cf10ae87bc84dfe43e55afc39 not found: ID does not exist" containerID="c3d7e4c6e41d6a457e050942c6ac447823a5e27cf10ae87bc84dfe43e55afc39" Feb 27 08:09:06 crc kubenswrapper[4612]: I0227 08:09:06.527601 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c3d7e4c6e41d6a457e050942c6ac447823a5e27cf10ae87bc84dfe43e55afc39"} err="failed to get container status \"c3d7e4c6e41d6a457e050942c6ac447823a5e27cf10ae87bc84dfe43e55afc39\": rpc error: code = NotFound desc = could not find container \"c3d7e4c6e41d6a457e050942c6ac447823a5e27cf10ae87bc84dfe43e55afc39\": container with ID starting with c3d7e4c6e41d6a457e050942c6ac447823a5e27cf10ae87bc84dfe43e55afc39 not found: ID does not exist" Feb 27 08:09:06 crc kubenswrapper[4612]: I0227 08:09:06.527646 4612 scope.go:117] "RemoveContainer" containerID="87ccc96505e0e1cc4d27fc5ae561b2d18ece432441be4afda55ac54553c69d09" Feb 27 08:09:06 crc kubenswrapper[4612]: E0227 08:09:06.533570 4612 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"87ccc96505e0e1cc4d27fc5ae561b2d18ece432441be4afda55ac54553c69d09\": container with ID starting with 87ccc96505e0e1cc4d27fc5ae561b2d18ece432441be4afda55ac54553c69d09 not found: ID does not exist" containerID="87ccc96505e0e1cc4d27fc5ae561b2d18ece432441be4afda55ac54553c69d09" Feb 27 08:09:06 crc kubenswrapper[4612]: I0227 08:09:06.533609 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"87ccc96505e0e1cc4d27fc5ae561b2d18ece432441be4afda55ac54553c69d09"} err="failed to get container status \"87ccc96505e0e1cc4d27fc5ae561b2d18ece432441be4afda55ac54553c69d09\": rpc error: code = NotFound desc = could not find container \"87ccc96505e0e1cc4d27fc5ae561b2d18ece432441be4afda55ac54553c69d09\": container with ID starting with 87ccc96505e0e1cc4d27fc5ae561b2d18ece432441be4afda55ac54553c69d09 not found: ID does not exist" Feb 27 08:09:06 crc kubenswrapper[4612]: I0227 08:09:06.533659 4612 scope.go:117] "RemoveContainer" containerID="c3d7e4c6e41d6a457e050942c6ac447823a5e27cf10ae87bc84dfe43e55afc39" Feb 27 08:09:06 crc kubenswrapper[4612]: I0227 08:09:06.534225 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c3d7e4c6e41d6a457e050942c6ac447823a5e27cf10ae87bc84dfe43e55afc39"} err="failed to get container status \"c3d7e4c6e41d6a457e050942c6ac447823a5e27cf10ae87bc84dfe43e55afc39\": rpc error: code = NotFound desc = could not find container \"c3d7e4c6e41d6a457e050942c6ac447823a5e27cf10ae87bc84dfe43e55afc39\": container with ID starting with c3d7e4c6e41d6a457e050942c6ac447823a5e27cf10ae87bc84dfe43e55afc39 not found: ID does not exist" Feb 27 08:09:06 crc kubenswrapper[4612]: I0227 08:09:06.534255 4612 scope.go:117] "RemoveContainer" containerID="87ccc96505e0e1cc4d27fc5ae561b2d18ece432441be4afda55ac54553c69d09" Feb 27 08:09:06 crc kubenswrapper[4612]: I0227 08:09:06.534784 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"87ccc96505e0e1cc4d27fc5ae561b2d18ece432441be4afda55ac54553c69d09"} err="failed to get container status \"87ccc96505e0e1cc4d27fc5ae561b2d18ece432441be4afda55ac54553c69d09\": rpc error: code = NotFound desc = could not find container \"87ccc96505e0e1cc4d27fc5ae561b2d18ece432441be4afda55ac54553c69d09\": container with ID starting with 87ccc96505e0e1cc4d27fc5ae561b2d18ece432441be4afda55ac54553c69d09 not found: ID does not exist" Feb 27 08:09:06 crc kubenswrapper[4612]: I0227 08:09:06.602760 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 27 08:09:06 crc kubenswrapper[4612]: I0227 08:09:06.615173 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 27 08:09:06 crc kubenswrapper[4612]: I0227 08:09:06.646878 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Feb 27 08:09:06 crc kubenswrapper[4612]: E0227 08:09:06.647326 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a980065-12fa-4005-9bc7-cdde0637d18e" containerName="init" Feb 27 08:09:06 crc kubenswrapper[4612]: I0227 08:09:06.647351 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a980065-12fa-4005-9bc7-cdde0637d18e" containerName="init" Feb 27 08:09:06 crc kubenswrapper[4612]: E0227 08:09:06.647367 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d5de72cb-312f-4631-b2e2-e0df3f759e90" containerName="glance-httpd" Feb 27 08:09:06 crc kubenswrapper[4612]: I0227 08:09:06.647375 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="d5de72cb-312f-4631-b2e2-e0df3f759e90" containerName="glance-httpd" Feb 27 08:09:06 crc kubenswrapper[4612]: E0227 08:09:06.647400 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d5de72cb-312f-4631-b2e2-e0df3f759e90" containerName="glance-log" Feb 27 08:09:06 crc kubenswrapper[4612]: I0227 08:09:06.647410 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="d5de72cb-312f-4631-b2e2-e0df3f759e90" containerName="glance-log" Feb 27 08:09:06 crc kubenswrapper[4612]: I0227 08:09:06.647626 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="d5de72cb-312f-4631-b2e2-e0df3f759e90" containerName="glance-httpd" Feb 27 08:09:06 crc kubenswrapper[4612]: I0227 08:09:06.647652 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="d5de72cb-312f-4631-b2e2-e0df3f759e90" containerName="glance-log" Feb 27 08:09:06 crc kubenswrapper[4612]: I0227 08:09:06.647671 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="0a980065-12fa-4005-9bc7-cdde0637d18e" containerName="init" Feb 27 08:09:06 crc kubenswrapper[4612]: I0227 08:09:06.648829 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 27 08:09:06 crc kubenswrapper[4612]: I0227 08:09:06.657097 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Feb 27 08:09:06 crc kubenswrapper[4612]: I0227 08:09:06.658005 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Feb 27 08:09:06 crc kubenswrapper[4612]: I0227 08:09:06.663626 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 27 08:09:06 crc kubenswrapper[4612]: I0227 08:09:06.819065 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d62745a5-cfe3-48b0-a167-76735bdfd8fe-logs\") pod \"glance-default-external-api-0\" (UID: \"d62745a5-cfe3-48b0-a167-76735bdfd8fe\") " pod="openstack/glance-default-external-api-0" Feb 27 08:09:06 crc kubenswrapper[4612]: I0227 08:09:06.819109 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d62745a5-cfe3-48b0-a167-76735bdfd8fe-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"d62745a5-cfe3-48b0-a167-76735bdfd8fe\") " pod="openstack/glance-default-external-api-0" Feb 27 08:09:06 crc kubenswrapper[4612]: I0227 08:09:06.819198 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d62745a5-cfe3-48b0-a167-76735bdfd8fe-config-data\") pod \"glance-default-external-api-0\" (UID: \"d62745a5-cfe3-48b0-a167-76735bdfd8fe\") " pod="openstack/glance-default-external-api-0" Feb 27 08:09:06 crc kubenswrapper[4612]: I0227 08:09:06.819224 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d62745a5-cfe3-48b0-a167-76735bdfd8fe-scripts\") pod \"glance-default-external-api-0\" (UID: \"d62745a5-cfe3-48b0-a167-76735bdfd8fe\") " pod="openstack/glance-default-external-api-0" Feb 27 08:09:06 crc kubenswrapper[4612]: I0227 08:09:06.819252 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2r2jz\" (UniqueName: \"kubernetes.io/projected/d62745a5-cfe3-48b0-a167-76735bdfd8fe-kube-api-access-2r2jz\") pod \"glance-default-external-api-0\" (UID: \"d62745a5-cfe3-48b0-a167-76735bdfd8fe\") " pod="openstack/glance-default-external-api-0" Feb 27 08:09:06 crc kubenswrapper[4612]: I0227 08:09:06.819275 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-external-api-0\" (UID: \"d62745a5-cfe3-48b0-a167-76735bdfd8fe\") " pod="openstack/glance-default-external-api-0" Feb 27 08:09:06 crc kubenswrapper[4612]: I0227 08:09:06.819294 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d62745a5-cfe3-48b0-a167-76735bdfd8fe-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"d62745a5-cfe3-48b0-a167-76735bdfd8fe\") " pod="openstack/glance-default-external-api-0" Feb 27 08:09:06 crc kubenswrapper[4612]: I0227 08:09:06.819355 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d62745a5-cfe3-48b0-a167-76735bdfd8fe-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"d62745a5-cfe3-48b0-a167-76735bdfd8fe\") " pod="openstack/glance-default-external-api-0" Feb 27 08:09:06 crc kubenswrapper[4612]: I0227 08:09:06.875327 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d5de72cb-312f-4631-b2e2-e0df3f759e90" path="/var/lib/kubelet/pods/d5de72cb-312f-4631-b2e2-e0df3f759e90/volumes" Feb 27 08:09:06 crc kubenswrapper[4612]: I0227 08:09:06.938646 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d62745a5-cfe3-48b0-a167-76735bdfd8fe-config-data\") pod \"glance-default-external-api-0\" (UID: \"d62745a5-cfe3-48b0-a167-76735bdfd8fe\") " pod="openstack/glance-default-external-api-0" Feb 27 08:09:06 crc kubenswrapper[4612]: I0227 08:09:06.938714 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d62745a5-cfe3-48b0-a167-76735bdfd8fe-scripts\") pod \"glance-default-external-api-0\" (UID: \"d62745a5-cfe3-48b0-a167-76735bdfd8fe\") " pod="openstack/glance-default-external-api-0" Feb 27 08:09:06 crc kubenswrapper[4612]: I0227 08:09:06.938754 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2r2jz\" (UniqueName: \"kubernetes.io/projected/d62745a5-cfe3-48b0-a167-76735bdfd8fe-kube-api-access-2r2jz\") pod \"glance-default-external-api-0\" (UID: \"d62745a5-cfe3-48b0-a167-76735bdfd8fe\") " pod="openstack/glance-default-external-api-0" Feb 27 08:09:06 crc kubenswrapper[4612]: I0227 08:09:06.941549 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-external-api-0\" (UID: \"d62745a5-cfe3-48b0-a167-76735bdfd8fe\") " pod="openstack/glance-default-external-api-0" Feb 27 08:09:06 crc kubenswrapper[4612]: I0227 08:09:06.941594 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d62745a5-cfe3-48b0-a167-76735bdfd8fe-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"d62745a5-cfe3-48b0-a167-76735bdfd8fe\") " pod="openstack/glance-default-external-api-0" Feb 27 08:09:06 crc kubenswrapper[4612]: I0227 08:09:06.941803 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d62745a5-cfe3-48b0-a167-76735bdfd8fe-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"d62745a5-cfe3-48b0-a167-76735bdfd8fe\") " pod="openstack/glance-default-external-api-0" Feb 27 08:09:06 crc kubenswrapper[4612]: I0227 08:09:06.941892 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d62745a5-cfe3-48b0-a167-76735bdfd8fe-logs\") pod \"glance-default-external-api-0\" (UID: \"d62745a5-cfe3-48b0-a167-76735bdfd8fe\") " pod="openstack/glance-default-external-api-0" Feb 27 08:09:06 crc kubenswrapper[4612]: I0227 08:09:06.941937 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d62745a5-cfe3-48b0-a167-76735bdfd8fe-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"d62745a5-cfe3-48b0-a167-76735bdfd8fe\") " pod="openstack/glance-default-external-api-0" Feb 27 08:09:06 crc kubenswrapper[4612]: I0227 08:09:06.943154 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d62745a5-cfe3-48b0-a167-76735bdfd8fe-logs\") pod \"glance-default-external-api-0\" (UID: \"d62745a5-cfe3-48b0-a167-76735bdfd8fe\") " pod="openstack/glance-default-external-api-0" Feb 27 08:09:06 crc kubenswrapper[4612]: I0227 08:09:06.943325 4612 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-external-api-0\" (UID: \"d62745a5-cfe3-48b0-a167-76735bdfd8fe\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/glance-default-external-api-0" Feb 27 08:09:06 crc kubenswrapper[4612]: I0227 08:09:06.943392 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d62745a5-cfe3-48b0-a167-76735bdfd8fe-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"d62745a5-cfe3-48b0-a167-76735bdfd8fe\") " pod="openstack/glance-default-external-api-0" Feb 27 08:09:06 crc kubenswrapper[4612]: I0227 08:09:06.961672 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d62745a5-cfe3-48b0-a167-76735bdfd8fe-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"d62745a5-cfe3-48b0-a167-76735bdfd8fe\") " pod="openstack/glance-default-external-api-0" Feb 27 08:09:06 crc kubenswrapper[4612]: I0227 08:09:06.961718 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d62745a5-cfe3-48b0-a167-76735bdfd8fe-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"d62745a5-cfe3-48b0-a167-76735bdfd8fe\") " pod="openstack/glance-default-external-api-0" Feb 27 08:09:06 crc kubenswrapper[4612]: I0227 08:09:06.963470 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d62745a5-cfe3-48b0-a167-76735bdfd8fe-config-data\") pod \"glance-default-external-api-0\" (UID: \"d62745a5-cfe3-48b0-a167-76735bdfd8fe\") " pod="openstack/glance-default-external-api-0" Feb 27 08:09:06 crc kubenswrapper[4612]: I0227 08:09:06.972854 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2r2jz\" (UniqueName: \"kubernetes.io/projected/d62745a5-cfe3-48b0-a167-76735bdfd8fe-kube-api-access-2r2jz\") pod \"glance-default-external-api-0\" (UID: \"d62745a5-cfe3-48b0-a167-76735bdfd8fe\") " pod="openstack/glance-default-external-api-0" Feb 27 08:09:06 crc kubenswrapper[4612]: I0227 08:09:06.980430 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d62745a5-cfe3-48b0-a167-76735bdfd8fe-scripts\") pod \"glance-default-external-api-0\" (UID: \"d62745a5-cfe3-48b0-a167-76735bdfd8fe\") " pod="openstack/glance-default-external-api-0" Feb 27 08:09:07 crc kubenswrapper[4612]: I0227 08:09:07.002162 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-external-api-0\" (UID: \"d62745a5-cfe3-48b0-a167-76735bdfd8fe\") " pod="openstack/glance-default-external-api-0" Feb 27 08:09:07 crc kubenswrapper[4612]: I0227 08:09:07.067915 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 27 08:09:07 crc kubenswrapper[4612]: I0227 08:09:07.144434 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7fc3d9c6-3a82-4dcf-bfb5-fbd7b8225f90-combined-ca-bundle\") pod \"7fc3d9c6-3a82-4dcf-bfb5-fbd7b8225f90\" (UID: \"7fc3d9c6-3a82-4dcf-bfb5-fbd7b8225f90\") " Feb 27 08:09:07 crc kubenswrapper[4612]: I0227 08:09:07.144681 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7fc3d9c6-3a82-4dcf-bfb5-fbd7b8225f90-httpd-run\") pod \"7fc3d9c6-3a82-4dcf-bfb5-fbd7b8225f90\" (UID: \"7fc3d9c6-3a82-4dcf-bfb5-fbd7b8225f90\") " Feb 27 08:09:07 crc kubenswrapper[4612]: I0227 08:09:07.144734 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7fc3d9c6-3a82-4dcf-bfb5-fbd7b8225f90-logs\") pod \"7fc3d9c6-3a82-4dcf-bfb5-fbd7b8225f90\" (UID: \"7fc3d9c6-3a82-4dcf-bfb5-fbd7b8225f90\") " Feb 27 08:09:07 crc kubenswrapper[4612]: I0227 08:09:07.144772 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7fc3d9c6-3a82-4dcf-bfb5-fbd7b8225f90-internal-tls-certs\") pod \"7fc3d9c6-3a82-4dcf-bfb5-fbd7b8225f90\" (UID: \"7fc3d9c6-3a82-4dcf-bfb5-fbd7b8225f90\") " Feb 27 08:09:07 crc kubenswrapper[4612]: I0227 08:09:07.144790 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h7m9b\" (UniqueName: \"kubernetes.io/projected/7fc3d9c6-3a82-4dcf-bfb5-fbd7b8225f90-kube-api-access-h7m9b\") pod \"7fc3d9c6-3a82-4dcf-bfb5-fbd7b8225f90\" (UID: \"7fc3d9c6-3a82-4dcf-bfb5-fbd7b8225f90\") " Feb 27 08:09:07 crc kubenswrapper[4612]: I0227 08:09:07.144811 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"7fc3d9c6-3a82-4dcf-bfb5-fbd7b8225f90\" (UID: \"7fc3d9c6-3a82-4dcf-bfb5-fbd7b8225f90\") " Feb 27 08:09:07 crc kubenswrapper[4612]: I0227 08:09:07.144836 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7fc3d9c6-3a82-4dcf-bfb5-fbd7b8225f90-config-data\") pod \"7fc3d9c6-3a82-4dcf-bfb5-fbd7b8225f90\" (UID: \"7fc3d9c6-3a82-4dcf-bfb5-fbd7b8225f90\") " Feb 27 08:09:07 crc kubenswrapper[4612]: I0227 08:09:07.144874 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7fc3d9c6-3a82-4dcf-bfb5-fbd7b8225f90-scripts\") pod \"7fc3d9c6-3a82-4dcf-bfb5-fbd7b8225f90\" (UID: \"7fc3d9c6-3a82-4dcf-bfb5-fbd7b8225f90\") " Feb 27 08:09:07 crc kubenswrapper[4612]: I0227 08:09:07.146912 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7fc3d9c6-3a82-4dcf-bfb5-fbd7b8225f90-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "7fc3d9c6-3a82-4dcf-bfb5-fbd7b8225f90" (UID: "7fc3d9c6-3a82-4dcf-bfb5-fbd7b8225f90"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 08:09:07 crc kubenswrapper[4612]: I0227 08:09:07.148619 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7fc3d9c6-3a82-4dcf-bfb5-fbd7b8225f90-logs" (OuterVolumeSpecName: "logs") pod "7fc3d9c6-3a82-4dcf-bfb5-fbd7b8225f90" (UID: "7fc3d9c6-3a82-4dcf-bfb5-fbd7b8225f90"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 08:09:07 crc kubenswrapper[4612]: I0227 08:09:07.150389 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7fc3d9c6-3a82-4dcf-bfb5-fbd7b8225f90-scripts" (OuterVolumeSpecName: "scripts") pod "7fc3d9c6-3a82-4dcf-bfb5-fbd7b8225f90" (UID: "7fc3d9c6-3a82-4dcf-bfb5-fbd7b8225f90"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:09:07 crc kubenswrapper[4612]: I0227 08:09:07.154127 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage04-crc" (OuterVolumeSpecName: "glance") pod "7fc3d9c6-3a82-4dcf-bfb5-fbd7b8225f90" (UID: "7fc3d9c6-3a82-4dcf-bfb5-fbd7b8225f90"). InnerVolumeSpecName "local-storage04-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Feb 27 08:09:07 crc kubenswrapper[4612]: I0227 08:09:07.154301 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7fc3d9c6-3a82-4dcf-bfb5-fbd7b8225f90-kube-api-access-h7m9b" (OuterVolumeSpecName: "kube-api-access-h7m9b") pod "7fc3d9c6-3a82-4dcf-bfb5-fbd7b8225f90" (UID: "7fc3d9c6-3a82-4dcf-bfb5-fbd7b8225f90"). InnerVolumeSpecName "kube-api-access-h7m9b". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 08:09:07 crc kubenswrapper[4612]: I0227 08:09:07.206440 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7fc3d9c6-3a82-4dcf-bfb5-fbd7b8225f90-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7fc3d9c6-3a82-4dcf-bfb5-fbd7b8225f90" (UID: "7fc3d9c6-3a82-4dcf-bfb5-fbd7b8225f90"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:09:07 crc kubenswrapper[4612]: I0227 08:09:07.215873 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7fc3d9c6-3a82-4dcf-bfb5-fbd7b8225f90-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "7fc3d9c6-3a82-4dcf-bfb5-fbd7b8225f90" (UID: "7fc3d9c6-3a82-4dcf-bfb5-fbd7b8225f90"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:09:07 crc kubenswrapper[4612]: I0227 08:09:07.245327 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7fc3d9c6-3a82-4dcf-bfb5-fbd7b8225f90-config-data" (OuterVolumeSpecName: "config-data") pod "7fc3d9c6-3a82-4dcf-bfb5-fbd7b8225f90" (UID: "7fc3d9c6-3a82-4dcf-bfb5-fbd7b8225f90"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:09:07 crc kubenswrapper[4612]: I0227 08:09:07.247111 4612 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7fc3d9c6-3a82-4dcf-bfb5-fbd7b8225f90-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 27 08:09:07 crc kubenswrapper[4612]: I0227 08:09:07.247137 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h7m9b\" (UniqueName: \"kubernetes.io/projected/7fc3d9c6-3a82-4dcf-bfb5-fbd7b8225f90-kube-api-access-h7m9b\") on node \"crc\" DevicePath \"\"" Feb 27 08:09:07 crc kubenswrapper[4612]: I0227 08:09:07.247169 4612 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" " Feb 27 08:09:07 crc kubenswrapper[4612]: I0227 08:09:07.247178 4612 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7fc3d9c6-3a82-4dcf-bfb5-fbd7b8225f90-config-data\") on node \"crc\" DevicePath \"\"" Feb 27 08:09:07 crc kubenswrapper[4612]: I0227 08:09:07.247186 4612 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7fc3d9c6-3a82-4dcf-bfb5-fbd7b8225f90-scripts\") on node \"crc\" DevicePath \"\"" Feb 27 08:09:07 crc kubenswrapper[4612]: I0227 08:09:07.247195 4612 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7fc3d9c6-3a82-4dcf-bfb5-fbd7b8225f90-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 08:09:07 crc kubenswrapper[4612]: I0227 08:09:07.247205 4612 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7fc3d9c6-3a82-4dcf-bfb5-fbd7b8225f90-httpd-run\") on node \"crc\" DevicePath \"\"" Feb 27 08:09:07 crc kubenswrapper[4612]: I0227 08:09:07.247217 4612 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7fc3d9c6-3a82-4dcf-bfb5-fbd7b8225f90-logs\") on node \"crc\" DevicePath \"\"" Feb 27 08:09:07 crc kubenswrapper[4612]: I0227 08:09:07.264530 4612 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage04-crc" (UniqueName: "kubernetes.io/local-volume/local-storage04-crc") on node "crc" Feb 27 08:09:07 crc kubenswrapper[4612]: I0227 08:09:07.273989 4612 generic.go:334] "Generic (PLEG): container finished" podID="7fc3d9c6-3a82-4dcf-bfb5-fbd7b8225f90" containerID="0d79762e8bbcf38c5d056d44b9fa86200838764e4ac3355e7d590c3d228be5d1" exitCode=0 Feb 27 08:09:07 crc kubenswrapper[4612]: I0227 08:09:07.274033 4612 generic.go:334] "Generic (PLEG): container finished" podID="7fc3d9c6-3a82-4dcf-bfb5-fbd7b8225f90" containerID="d3d04d402df81d47178f525ec89626c4c3774ab33d06c103ccd34cdf3c8720af" exitCode=143 Feb 27 08:09:07 crc kubenswrapper[4612]: I0227 08:09:07.274228 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"7fc3d9c6-3a82-4dcf-bfb5-fbd7b8225f90","Type":"ContainerDied","Data":"0d79762e8bbcf38c5d056d44b9fa86200838764e4ac3355e7d590c3d228be5d1"} Feb 27 08:09:07 crc kubenswrapper[4612]: I0227 08:09:07.274273 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"7fc3d9c6-3a82-4dcf-bfb5-fbd7b8225f90","Type":"ContainerDied","Data":"d3d04d402df81d47178f525ec89626c4c3774ab33d06c103ccd34cdf3c8720af"} Feb 27 08:09:07 crc kubenswrapper[4612]: I0227 08:09:07.274325 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"7fc3d9c6-3a82-4dcf-bfb5-fbd7b8225f90","Type":"ContainerDied","Data":"900b007f2caf8828e2d318815e2cbea0bd48685fbec9cf487f1e59bc1ed25dd1"} Feb 27 08:09:07 crc kubenswrapper[4612]: I0227 08:09:07.274348 4612 scope.go:117] "RemoveContainer" containerID="0d79762e8bbcf38c5d056d44b9fa86200838764e4ac3355e7d590c3d228be5d1" Feb 27 08:09:07 crc kubenswrapper[4612]: I0227 08:09:07.274590 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 27 08:09:07 crc kubenswrapper[4612]: I0227 08:09:07.283275 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 27 08:09:07 crc kubenswrapper[4612]: I0227 08:09:07.313774 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 27 08:09:07 crc kubenswrapper[4612]: I0227 08:09:07.321237 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 27 08:09:07 crc kubenswrapper[4612]: I0227 08:09:07.341582 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 27 08:09:07 crc kubenswrapper[4612]: E0227 08:09:07.342366 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7fc3d9c6-3a82-4dcf-bfb5-fbd7b8225f90" containerName="glance-httpd" Feb 27 08:09:07 crc kubenswrapper[4612]: I0227 08:09:07.342463 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="7fc3d9c6-3a82-4dcf-bfb5-fbd7b8225f90" containerName="glance-httpd" Feb 27 08:09:07 crc kubenswrapper[4612]: E0227 08:09:07.342558 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7fc3d9c6-3a82-4dcf-bfb5-fbd7b8225f90" containerName="glance-log" Feb 27 08:09:07 crc kubenswrapper[4612]: I0227 08:09:07.342834 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="7fc3d9c6-3a82-4dcf-bfb5-fbd7b8225f90" containerName="glance-log" Feb 27 08:09:07 crc kubenswrapper[4612]: I0227 08:09:07.343059 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="7fc3d9c6-3a82-4dcf-bfb5-fbd7b8225f90" containerName="glance-httpd" Feb 27 08:09:07 crc kubenswrapper[4612]: I0227 08:09:07.343146 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="7fc3d9c6-3a82-4dcf-bfb5-fbd7b8225f90" containerName="glance-log" Feb 27 08:09:07 crc kubenswrapper[4612]: I0227 08:09:07.344790 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 27 08:09:07 crc kubenswrapper[4612]: I0227 08:09:07.348136 4612 reconciler_common.go:293] "Volume detached for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" DevicePath \"\"" Feb 27 08:09:07 crc kubenswrapper[4612]: I0227 08:09:07.351798 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Feb 27 08:09:07 crc kubenswrapper[4612]: I0227 08:09:07.355249 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Feb 27 08:09:07 crc kubenswrapper[4612]: I0227 08:09:07.360860 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 27 08:09:07 crc kubenswrapper[4612]: I0227 08:09:07.438407 4612 scope.go:117] "RemoveContainer" containerID="d3d04d402df81d47178f525ec89626c4c3774ab33d06c103ccd34cdf3c8720af" Feb 27 08:09:07 crc kubenswrapper[4612]: I0227 08:09:07.477371 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4726ffcd-f407-4cd4-a6e4-c422801f2617-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"4726ffcd-f407-4cd4-a6e4-c422801f2617\") " pod="openstack/glance-default-internal-api-0" Feb 27 08:09:07 crc kubenswrapper[4612]: I0227 08:09:07.477605 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4726ffcd-f407-4cd4-a6e4-c422801f2617-logs\") pod \"glance-default-internal-api-0\" (UID: \"4726ffcd-f407-4cd4-a6e4-c422801f2617\") " pod="openstack/glance-default-internal-api-0" Feb 27 08:09:07 crc kubenswrapper[4612]: I0227 08:09:07.477784 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"4726ffcd-f407-4cd4-a6e4-c422801f2617\") " pod="openstack/glance-default-internal-api-0" Feb 27 08:09:07 crc kubenswrapper[4612]: I0227 08:09:07.477914 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4726ffcd-f407-4cd4-a6e4-c422801f2617-scripts\") pod \"glance-default-internal-api-0\" (UID: \"4726ffcd-f407-4cd4-a6e4-c422801f2617\") " pod="openstack/glance-default-internal-api-0" Feb 27 08:09:07 crc kubenswrapper[4612]: I0227 08:09:07.478027 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-645t5\" (UniqueName: \"kubernetes.io/projected/4726ffcd-f407-4cd4-a6e4-c422801f2617-kube-api-access-645t5\") pod \"glance-default-internal-api-0\" (UID: \"4726ffcd-f407-4cd4-a6e4-c422801f2617\") " pod="openstack/glance-default-internal-api-0" Feb 27 08:09:07 crc kubenswrapper[4612]: I0227 08:09:07.478131 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4726ffcd-f407-4cd4-a6e4-c422801f2617-config-data\") pod \"glance-default-internal-api-0\" (UID: \"4726ffcd-f407-4cd4-a6e4-c422801f2617\") " pod="openstack/glance-default-internal-api-0" Feb 27 08:09:07 crc kubenswrapper[4612]: I0227 08:09:07.478259 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/4726ffcd-f407-4cd4-a6e4-c422801f2617-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"4726ffcd-f407-4cd4-a6e4-c422801f2617\") " pod="openstack/glance-default-internal-api-0" Feb 27 08:09:07 crc kubenswrapper[4612]: I0227 08:09:07.478378 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4726ffcd-f407-4cd4-a6e4-c422801f2617-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"4726ffcd-f407-4cd4-a6e4-c422801f2617\") " pod="openstack/glance-default-internal-api-0" Feb 27 08:09:07 crc kubenswrapper[4612]: I0227 08:09:07.548049 4612 scope.go:117] "RemoveContainer" containerID="0d79762e8bbcf38c5d056d44b9fa86200838764e4ac3355e7d590c3d228be5d1" Feb 27 08:09:07 crc kubenswrapper[4612]: E0227 08:09:07.551592 4612 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0d79762e8bbcf38c5d056d44b9fa86200838764e4ac3355e7d590c3d228be5d1\": container with ID starting with 0d79762e8bbcf38c5d056d44b9fa86200838764e4ac3355e7d590c3d228be5d1 not found: ID does not exist" containerID="0d79762e8bbcf38c5d056d44b9fa86200838764e4ac3355e7d590c3d228be5d1" Feb 27 08:09:07 crc kubenswrapper[4612]: I0227 08:09:07.551639 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0d79762e8bbcf38c5d056d44b9fa86200838764e4ac3355e7d590c3d228be5d1"} err="failed to get container status \"0d79762e8bbcf38c5d056d44b9fa86200838764e4ac3355e7d590c3d228be5d1\": rpc error: code = NotFound desc = could not find container \"0d79762e8bbcf38c5d056d44b9fa86200838764e4ac3355e7d590c3d228be5d1\": container with ID starting with 0d79762e8bbcf38c5d056d44b9fa86200838764e4ac3355e7d590c3d228be5d1 not found: ID does not exist" Feb 27 08:09:07 crc kubenswrapper[4612]: I0227 08:09:07.551667 4612 scope.go:117] "RemoveContainer" containerID="d3d04d402df81d47178f525ec89626c4c3774ab33d06c103ccd34cdf3c8720af" Feb 27 08:09:07 crc kubenswrapper[4612]: E0227 08:09:07.554517 4612 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d3d04d402df81d47178f525ec89626c4c3774ab33d06c103ccd34cdf3c8720af\": container with ID starting with d3d04d402df81d47178f525ec89626c4c3774ab33d06c103ccd34cdf3c8720af not found: ID does not exist" containerID="d3d04d402df81d47178f525ec89626c4c3774ab33d06c103ccd34cdf3c8720af" Feb 27 08:09:07 crc kubenswrapper[4612]: I0227 08:09:07.554568 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d3d04d402df81d47178f525ec89626c4c3774ab33d06c103ccd34cdf3c8720af"} err="failed to get container status \"d3d04d402df81d47178f525ec89626c4c3774ab33d06c103ccd34cdf3c8720af\": rpc error: code = NotFound desc = could not find container \"d3d04d402df81d47178f525ec89626c4c3774ab33d06c103ccd34cdf3c8720af\": container with ID starting with d3d04d402df81d47178f525ec89626c4c3774ab33d06c103ccd34cdf3c8720af not found: ID does not exist" Feb 27 08:09:07 crc kubenswrapper[4612]: I0227 08:09:07.554601 4612 scope.go:117] "RemoveContainer" containerID="0d79762e8bbcf38c5d056d44b9fa86200838764e4ac3355e7d590c3d228be5d1" Feb 27 08:09:07 crc kubenswrapper[4612]: I0227 08:09:07.558536 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0d79762e8bbcf38c5d056d44b9fa86200838764e4ac3355e7d590c3d228be5d1"} err="failed to get container status \"0d79762e8bbcf38c5d056d44b9fa86200838764e4ac3355e7d590c3d228be5d1\": rpc error: code = NotFound desc = could not find container \"0d79762e8bbcf38c5d056d44b9fa86200838764e4ac3355e7d590c3d228be5d1\": container with ID starting with 0d79762e8bbcf38c5d056d44b9fa86200838764e4ac3355e7d590c3d228be5d1 not found: ID does not exist" Feb 27 08:09:07 crc kubenswrapper[4612]: I0227 08:09:07.558572 4612 scope.go:117] "RemoveContainer" containerID="d3d04d402df81d47178f525ec89626c4c3774ab33d06c103ccd34cdf3c8720af" Feb 27 08:09:07 crc kubenswrapper[4612]: I0227 08:09:07.560476 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d3d04d402df81d47178f525ec89626c4c3774ab33d06c103ccd34cdf3c8720af"} err="failed to get container status \"d3d04d402df81d47178f525ec89626c4c3774ab33d06c103ccd34cdf3c8720af\": rpc error: code = NotFound desc = could not find container \"d3d04d402df81d47178f525ec89626c4c3774ab33d06c103ccd34cdf3c8720af\": container with ID starting with d3d04d402df81d47178f525ec89626c4c3774ab33d06c103ccd34cdf3c8720af not found: ID does not exist" Feb 27 08:09:07 crc kubenswrapper[4612]: I0227 08:09:07.581755 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4726ffcd-f407-4cd4-a6e4-c422801f2617-scripts\") pod \"glance-default-internal-api-0\" (UID: \"4726ffcd-f407-4cd4-a6e4-c422801f2617\") " pod="openstack/glance-default-internal-api-0" Feb 27 08:09:07 crc kubenswrapper[4612]: I0227 08:09:07.581785 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-645t5\" (UniqueName: \"kubernetes.io/projected/4726ffcd-f407-4cd4-a6e4-c422801f2617-kube-api-access-645t5\") pod \"glance-default-internal-api-0\" (UID: \"4726ffcd-f407-4cd4-a6e4-c422801f2617\") " pod="openstack/glance-default-internal-api-0" Feb 27 08:09:07 crc kubenswrapper[4612]: I0227 08:09:07.581812 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4726ffcd-f407-4cd4-a6e4-c422801f2617-config-data\") pod \"glance-default-internal-api-0\" (UID: \"4726ffcd-f407-4cd4-a6e4-c422801f2617\") " pod="openstack/glance-default-internal-api-0" Feb 27 08:09:07 crc kubenswrapper[4612]: I0227 08:09:07.581854 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/4726ffcd-f407-4cd4-a6e4-c422801f2617-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"4726ffcd-f407-4cd4-a6e4-c422801f2617\") " pod="openstack/glance-default-internal-api-0" Feb 27 08:09:07 crc kubenswrapper[4612]: I0227 08:09:07.581888 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4726ffcd-f407-4cd4-a6e4-c422801f2617-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"4726ffcd-f407-4cd4-a6e4-c422801f2617\") " pod="openstack/glance-default-internal-api-0" Feb 27 08:09:07 crc kubenswrapper[4612]: I0227 08:09:07.582739 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4726ffcd-f407-4cd4-a6e4-c422801f2617-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"4726ffcd-f407-4cd4-a6e4-c422801f2617\") " pod="openstack/glance-default-internal-api-0" Feb 27 08:09:07 crc kubenswrapper[4612]: I0227 08:09:07.582763 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4726ffcd-f407-4cd4-a6e4-c422801f2617-logs\") pod \"glance-default-internal-api-0\" (UID: \"4726ffcd-f407-4cd4-a6e4-c422801f2617\") " pod="openstack/glance-default-internal-api-0" Feb 27 08:09:07 crc kubenswrapper[4612]: I0227 08:09:07.582792 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"4726ffcd-f407-4cd4-a6e4-c422801f2617\") " pod="openstack/glance-default-internal-api-0" Feb 27 08:09:07 crc kubenswrapper[4612]: I0227 08:09:07.583075 4612 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"4726ffcd-f407-4cd4-a6e4-c422801f2617\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/glance-default-internal-api-0" Feb 27 08:09:07 crc kubenswrapper[4612]: I0227 08:09:07.586188 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/4726ffcd-f407-4cd4-a6e4-c422801f2617-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"4726ffcd-f407-4cd4-a6e4-c422801f2617\") " pod="openstack/glance-default-internal-api-0" Feb 27 08:09:07 crc kubenswrapper[4612]: I0227 08:09:07.589226 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4726ffcd-f407-4cd4-a6e4-c422801f2617-logs\") pod \"glance-default-internal-api-0\" (UID: \"4726ffcd-f407-4cd4-a6e4-c422801f2617\") " pod="openstack/glance-default-internal-api-0" Feb 27 08:09:07 crc kubenswrapper[4612]: I0227 08:09:07.601425 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4726ffcd-f407-4cd4-a6e4-c422801f2617-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"4726ffcd-f407-4cd4-a6e4-c422801f2617\") " pod="openstack/glance-default-internal-api-0" Feb 27 08:09:07 crc kubenswrapper[4612]: I0227 08:09:07.603880 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4726ffcd-f407-4cd4-a6e4-c422801f2617-scripts\") pod \"glance-default-internal-api-0\" (UID: \"4726ffcd-f407-4cd4-a6e4-c422801f2617\") " pod="openstack/glance-default-internal-api-0" Feb 27 08:09:07 crc kubenswrapper[4612]: I0227 08:09:07.604790 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4726ffcd-f407-4cd4-a6e4-c422801f2617-config-data\") pod \"glance-default-internal-api-0\" (UID: \"4726ffcd-f407-4cd4-a6e4-c422801f2617\") " pod="openstack/glance-default-internal-api-0" Feb 27 08:09:07 crc kubenswrapper[4612]: I0227 08:09:07.611102 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4726ffcd-f407-4cd4-a6e4-c422801f2617-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"4726ffcd-f407-4cd4-a6e4-c422801f2617\") " pod="openstack/glance-default-internal-api-0" Feb 27 08:09:07 crc kubenswrapper[4612]: I0227 08:09:07.636065 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-645t5\" (UniqueName: \"kubernetes.io/projected/4726ffcd-f407-4cd4-a6e4-c422801f2617-kube-api-access-645t5\") pod \"glance-default-internal-api-0\" (UID: \"4726ffcd-f407-4cd4-a6e4-c422801f2617\") " pod="openstack/glance-default-internal-api-0" Feb 27 08:09:07 crc kubenswrapper[4612]: I0227 08:09:07.651314 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"4726ffcd-f407-4cd4-a6e4-c422801f2617\") " pod="openstack/glance-default-internal-api-0" Feb 27 08:09:07 crc kubenswrapper[4612]: I0227 08:09:07.742917 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 27 08:09:08 crc kubenswrapper[4612]: I0227 08:09:08.080766 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 27 08:09:08 crc kubenswrapper[4612]: I0227 08:09:08.331616 4612 generic.go:334] "Generic (PLEG): container finished" podID="dfebce69-1679-4abd-8b9d-270046157f42" containerID="9a40f17b5d0d9218b64aab79491862655274f46636610ed2f3fac9e5e60ebdf4" exitCode=0 Feb 27 08:09:08 crc kubenswrapper[4612]: I0227 08:09:08.331884 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-4nm8j" event={"ID":"dfebce69-1679-4abd-8b9d-270046157f42","Type":"ContainerDied","Data":"9a40f17b5d0d9218b64aab79491862655274f46636610ed2f3fac9e5e60ebdf4"} Feb 27 08:09:08 crc kubenswrapper[4612]: I0227 08:09:08.357484 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 27 08:09:08 crc kubenswrapper[4612]: I0227 08:09:08.360509 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"d62745a5-cfe3-48b0-a167-76735bdfd8fe","Type":"ContainerStarted","Data":"3d531517a4dce736d9caa8e8b6967b62d1fa7937af3dabd7ecc97ce41b5b7653"} Feb 27 08:09:08 crc kubenswrapper[4612]: I0227 08:09:08.716289 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-7f9b6cd66c-5lwzw"] Feb 27 08:09:08 crc kubenswrapper[4612]: I0227 08:09:08.765046 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-6b985547d4-mkfpv"] Feb 27 08:09:08 crc kubenswrapper[4612]: I0227 08:09:08.767579 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6b985547d4-mkfpv" Feb 27 08:09:08 crc kubenswrapper[4612]: I0227 08:09:08.775218 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-horizon-svc" Feb 27 08:09:08 crc kubenswrapper[4612]: I0227 08:09:08.777654 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-6b985547d4-mkfpv"] Feb 27 08:09:08 crc kubenswrapper[4612]: I0227 08:09:08.853173 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-594c9f8cf-q282v"] Feb 27 08:09:08 crc kubenswrapper[4612]: I0227 08:09:08.889217 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7fc3d9c6-3a82-4dcf-bfb5-fbd7b8225f90" path="/var/lib/kubelet/pods/7fc3d9c6-3a82-4dcf-bfb5-fbd7b8225f90/volumes" Feb 27 08:09:08 crc kubenswrapper[4612]: I0227 08:09:08.890109 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 27 08:09:08 crc kubenswrapper[4612]: I0227 08:09:08.890146 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-77948db5bb-97qh6"] Feb 27 08:09:08 crc kubenswrapper[4612]: I0227 08:09:08.894378 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-77948db5bb-97qh6" Feb 27 08:09:08 crc kubenswrapper[4612]: I0227 08:09:08.920032 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-77948db5bb-97qh6"] Feb 27 08:09:08 crc kubenswrapper[4612]: I0227 08:09:08.941362 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/76009243-16b3-4652-abfe-a6cdba363724-config-data\") pod \"horizon-6b985547d4-mkfpv\" (UID: \"76009243-16b3-4652-abfe-a6cdba363724\") " pod="openstack/horizon-6b985547d4-mkfpv" Feb 27 08:09:08 crc kubenswrapper[4612]: I0227 08:09:08.950291 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b72b541d-4661-44a7-a121-c2a8aef6db11-logs\") pod \"horizon-77948db5bb-97qh6\" (UID: \"b72b541d-4661-44a7-a121-c2a8aef6db11\") " pod="openstack/horizon-77948db5bb-97qh6" Feb 27 08:09:08 crc kubenswrapper[4612]: I0227 08:09:08.950563 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/76009243-16b3-4652-abfe-a6cdba363724-logs\") pod \"horizon-6b985547d4-mkfpv\" (UID: \"76009243-16b3-4652-abfe-a6cdba363724\") " pod="openstack/horizon-6b985547d4-mkfpv" Feb 27 08:09:08 crc kubenswrapper[4612]: I0227 08:09:08.950734 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/b72b541d-4661-44a7-a121-c2a8aef6db11-horizon-tls-certs\") pod \"horizon-77948db5bb-97qh6\" (UID: \"b72b541d-4661-44a7-a121-c2a8aef6db11\") " pod="openstack/horizon-77948db5bb-97qh6" Feb 27 08:09:08 crc kubenswrapper[4612]: I0227 08:09:08.950906 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nhcmx\" (UniqueName: \"kubernetes.io/projected/b72b541d-4661-44a7-a121-c2a8aef6db11-kube-api-access-nhcmx\") pod \"horizon-77948db5bb-97qh6\" (UID: \"b72b541d-4661-44a7-a121-c2a8aef6db11\") " pod="openstack/horizon-77948db5bb-97qh6" Feb 27 08:09:08 crc kubenswrapper[4612]: I0227 08:09:08.951076 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b72b541d-4661-44a7-a121-c2a8aef6db11-scripts\") pod \"horizon-77948db5bb-97qh6\" (UID: \"b72b541d-4661-44a7-a121-c2a8aef6db11\") " pod="openstack/horizon-77948db5bb-97qh6" Feb 27 08:09:08 crc kubenswrapper[4612]: I0227 08:09:08.951209 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/76009243-16b3-4652-abfe-a6cdba363724-horizon-secret-key\") pod \"horizon-6b985547d4-mkfpv\" (UID: \"76009243-16b3-4652-abfe-a6cdba363724\") " pod="openstack/horizon-6b985547d4-mkfpv" Feb 27 08:09:08 crc kubenswrapper[4612]: I0227 08:09:08.951300 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b72b541d-4661-44a7-a121-c2a8aef6db11-combined-ca-bundle\") pod \"horizon-77948db5bb-97qh6\" (UID: \"b72b541d-4661-44a7-a121-c2a8aef6db11\") " pod="openstack/horizon-77948db5bb-97qh6" Feb 27 08:09:08 crc kubenswrapper[4612]: I0227 08:09:08.951518 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/76009243-16b3-4652-abfe-a6cdba363724-scripts\") pod \"horizon-6b985547d4-mkfpv\" (UID: \"76009243-16b3-4652-abfe-a6cdba363724\") " pod="openstack/horizon-6b985547d4-mkfpv" Feb 27 08:09:08 crc kubenswrapper[4612]: I0227 08:09:08.952302 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b72b541d-4661-44a7-a121-c2a8aef6db11-config-data\") pod \"horizon-77948db5bb-97qh6\" (UID: \"b72b541d-4661-44a7-a121-c2a8aef6db11\") " pod="openstack/horizon-77948db5bb-97qh6" Feb 27 08:09:08 crc kubenswrapper[4612]: I0227 08:09:08.952466 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/b72b541d-4661-44a7-a121-c2a8aef6db11-horizon-secret-key\") pod \"horizon-77948db5bb-97qh6\" (UID: \"b72b541d-4661-44a7-a121-c2a8aef6db11\") " pod="openstack/horizon-77948db5bb-97qh6" Feb 27 08:09:08 crc kubenswrapper[4612]: I0227 08:09:08.952613 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76009243-16b3-4652-abfe-a6cdba363724-combined-ca-bundle\") pod \"horizon-6b985547d4-mkfpv\" (UID: \"76009243-16b3-4652-abfe-a6cdba363724\") " pod="openstack/horizon-6b985547d4-mkfpv" Feb 27 08:09:08 crc kubenswrapper[4612]: I0227 08:09:08.952953 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/76009243-16b3-4652-abfe-a6cdba363724-horizon-tls-certs\") pod \"horizon-6b985547d4-mkfpv\" (UID: \"76009243-16b3-4652-abfe-a6cdba363724\") " pod="openstack/horizon-6b985547d4-mkfpv" Feb 27 08:09:08 crc kubenswrapper[4612]: I0227 08:09:08.953115 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9mg4q\" (UniqueName: \"kubernetes.io/projected/76009243-16b3-4652-abfe-a6cdba363724-kube-api-access-9mg4q\") pod \"horizon-6b985547d4-mkfpv\" (UID: \"76009243-16b3-4652-abfe-a6cdba363724\") " pod="openstack/horizon-6b985547d4-mkfpv" Feb 27 08:09:08 crc kubenswrapper[4612]: I0227 08:09:08.975535 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 27 08:09:09 crc kubenswrapper[4612]: I0227 08:09:09.054402 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9mg4q\" (UniqueName: \"kubernetes.io/projected/76009243-16b3-4652-abfe-a6cdba363724-kube-api-access-9mg4q\") pod \"horizon-6b985547d4-mkfpv\" (UID: \"76009243-16b3-4652-abfe-a6cdba363724\") " pod="openstack/horizon-6b985547d4-mkfpv" Feb 27 08:09:09 crc kubenswrapper[4612]: I0227 08:09:09.054460 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/76009243-16b3-4652-abfe-a6cdba363724-config-data\") pod \"horizon-6b985547d4-mkfpv\" (UID: \"76009243-16b3-4652-abfe-a6cdba363724\") " pod="openstack/horizon-6b985547d4-mkfpv" Feb 27 08:09:09 crc kubenswrapper[4612]: I0227 08:09:09.054489 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b72b541d-4661-44a7-a121-c2a8aef6db11-logs\") pod \"horizon-77948db5bb-97qh6\" (UID: \"b72b541d-4661-44a7-a121-c2a8aef6db11\") " pod="openstack/horizon-77948db5bb-97qh6" Feb 27 08:09:09 crc kubenswrapper[4612]: I0227 08:09:09.054523 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/76009243-16b3-4652-abfe-a6cdba363724-logs\") pod \"horizon-6b985547d4-mkfpv\" (UID: \"76009243-16b3-4652-abfe-a6cdba363724\") " pod="openstack/horizon-6b985547d4-mkfpv" Feb 27 08:09:09 crc kubenswrapper[4612]: I0227 08:09:09.054537 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/b72b541d-4661-44a7-a121-c2a8aef6db11-horizon-tls-certs\") pod \"horizon-77948db5bb-97qh6\" (UID: \"b72b541d-4661-44a7-a121-c2a8aef6db11\") " pod="openstack/horizon-77948db5bb-97qh6" Feb 27 08:09:09 crc kubenswrapper[4612]: I0227 08:09:09.054550 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nhcmx\" (UniqueName: \"kubernetes.io/projected/b72b541d-4661-44a7-a121-c2a8aef6db11-kube-api-access-nhcmx\") pod \"horizon-77948db5bb-97qh6\" (UID: \"b72b541d-4661-44a7-a121-c2a8aef6db11\") " pod="openstack/horizon-77948db5bb-97qh6" Feb 27 08:09:09 crc kubenswrapper[4612]: I0227 08:09:09.054566 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b72b541d-4661-44a7-a121-c2a8aef6db11-scripts\") pod \"horizon-77948db5bb-97qh6\" (UID: \"b72b541d-4661-44a7-a121-c2a8aef6db11\") " pod="openstack/horizon-77948db5bb-97qh6" Feb 27 08:09:09 crc kubenswrapper[4612]: I0227 08:09:09.054590 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/76009243-16b3-4652-abfe-a6cdba363724-horizon-secret-key\") pod \"horizon-6b985547d4-mkfpv\" (UID: \"76009243-16b3-4652-abfe-a6cdba363724\") " pod="openstack/horizon-6b985547d4-mkfpv" Feb 27 08:09:09 crc kubenswrapper[4612]: I0227 08:09:09.054607 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b72b541d-4661-44a7-a121-c2a8aef6db11-combined-ca-bundle\") pod \"horizon-77948db5bb-97qh6\" (UID: \"b72b541d-4661-44a7-a121-c2a8aef6db11\") " pod="openstack/horizon-77948db5bb-97qh6" Feb 27 08:09:09 crc kubenswrapper[4612]: I0227 08:09:09.054649 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/76009243-16b3-4652-abfe-a6cdba363724-scripts\") pod \"horizon-6b985547d4-mkfpv\" (UID: \"76009243-16b3-4652-abfe-a6cdba363724\") " pod="openstack/horizon-6b985547d4-mkfpv" Feb 27 08:09:09 crc kubenswrapper[4612]: I0227 08:09:09.054665 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b72b541d-4661-44a7-a121-c2a8aef6db11-config-data\") pod \"horizon-77948db5bb-97qh6\" (UID: \"b72b541d-4661-44a7-a121-c2a8aef6db11\") " pod="openstack/horizon-77948db5bb-97qh6" Feb 27 08:09:09 crc kubenswrapper[4612]: I0227 08:09:09.054707 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/b72b541d-4661-44a7-a121-c2a8aef6db11-horizon-secret-key\") pod \"horizon-77948db5bb-97qh6\" (UID: \"b72b541d-4661-44a7-a121-c2a8aef6db11\") " pod="openstack/horizon-77948db5bb-97qh6" Feb 27 08:09:09 crc kubenswrapper[4612]: I0227 08:09:09.054731 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76009243-16b3-4652-abfe-a6cdba363724-combined-ca-bundle\") pod \"horizon-6b985547d4-mkfpv\" (UID: \"76009243-16b3-4652-abfe-a6cdba363724\") " pod="openstack/horizon-6b985547d4-mkfpv" Feb 27 08:09:09 crc kubenswrapper[4612]: I0227 08:09:09.054753 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/76009243-16b3-4652-abfe-a6cdba363724-horizon-tls-certs\") pod \"horizon-6b985547d4-mkfpv\" (UID: \"76009243-16b3-4652-abfe-a6cdba363724\") " pod="openstack/horizon-6b985547d4-mkfpv" Feb 27 08:09:09 crc kubenswrapper[4612]: I0227 08:09:09.057067 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/76009243-16b3-4652-abfe-a6cdba363724-config-data\") pod \"horizon-6b985547d4-mkfpv\" (UID: \"76009243-16b3-4652-abfe-a6cdba363724\") " pod="openstack/horizon-6b985547d4-mkfpv" Feb 27 08:09:09 crc kubenswrapper[4612]: I0227 08:09:09.058490 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b72b541d-4661-44a7-a121-c2a8aef6db11-scripts\") pod \"horizon-77948db5bb-97qh6\" (UID: \"b72b541d-4661-44a7-a121-c2a8aef6db11\") " pod="openstack/horizon-77948db5bb-97qh6" Feb 27 08:09:09 crc kubenswrapper[4612]: I0227 08:09:09.060175 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b72b541d-4661-44a7-a121-c2a8aef6db11-config-data\") pod \"horizon-77948db5bb-97qh6\" (UID: \"b72b541d-4661-44a7-a121-c2a8aef6db11\") " pod="openstack/horizon-77948db5bb-97qh6" Feb 27 08:09:09 crc kubenswrapper[4612]: I0227 08:09:09.060393 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/76009243-16b3-4652-abfe-a6cdba363724-logs\") pod \"horizon-6b985547d4-mkfpv\" (UID: \"76009243-16b3-4652-abfe-a6cdba363724\") " pod="openstack/horizon-6b985547d4-mkfpv" Feb 27 08:09:09 crc kubenswrapper[4612]: I0227 08:09:09.061042 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/76009243-16b3-4652-abfe-a6cdba363724-scripts\") pod \"horizon-6b985547d4-mkfpv\" (UID: \"76009243-16b3-4652-abfe-a6cdba363724\") " pod="openstack/horizon-6b985547d4-mkfpv" Feb 27 08:09:09 crc kubenswrapper[4612]: I0227 08:09:09.061363 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b72b541d-4661-44a7-a121-c2a8aef6db11-logs\") pod \"horizon-77948db5bb-97qh6\" (UID: \"b72b541d-4661-44a7-a121-c2a8aef6db11\") " pod="openstack/horizon-77948db5bb-97qh6" Feb 27 08:09:09 crc kubenswrapper[4612]: I0227 08:09:09.069129 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/b72b541d-4661-44a7-a121-c2a8aef6db11-horizon-tls-certs\") pod \"horizon-77948db5bb-97qh6\" (UID: \"b72b541d-4661-44a7-a121-c2a8aef6db11\") " pod="openstack/horizon-77948db5bb-97qh6" Feb 27 08:09:09 crc kubenswrapper[4612]: I0227 08:09:09.069239 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b72b541d-4661-44a7-a121-c2a8aef6db11-combined-ca-bundle\") pod \"horizon-77948db5bb-97qh6\" (UID: \"b72b541d-4661-44a7-a121-c2a8aef6db11\") " pod="openstack/horizon-77948db5bb-97qh6" Feb 27 08:09:09 crc kubenswrapper[4612]: I0227 08:09:09.069617 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/76009243-16b3-4652-abfe-a6cdba363724-horizon-secret-key\") pod \"horizon-6b985547d4-mkfpv\" (UID: \"76009243-16b3-4652-abfe-a6cdba363724\") " pod="openstack/horizon-6b985547d4-mkfpv" Feb 27 08:09:09 crc kubenswrapper[4612]: I0227 08:09:09.070433 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/76009243-16b3-4652-abfe-a6cdba363724-horizon-tls-certs\") pod \"horizon-6b985547d4-mkfpv\" (UID: \"76009243-16b3-4652-abfe-a6cdba363724\") " pod="openstack/horizon-6b985547d4-mkfpv" Feb 27 08:09:09 crc kubenswrapper[4612]: I0227 08:09:09.072250 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9mg4q\" (UniqueName: \"kubernetes.io/projected/76009243-16b3-4652-abfe-a6cdba363724-kube-api-access-9mg4q\") pod \"horizon-6b985547d4-mkfpv\" (UID: \"76009243-16b3-4652-abfe-a6cdba363724\") " pod="openstack/horizon-6b985547d4-mkfpv" Feb 27 08:09:09 crc kubenswrapper[4612]: I0227 08:09:09.073189 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/b72b541d-4661-44a7-a121-c2a8aef6db11-horizon-secret-key\") pod \"horizon-77948db5bb-97qh6\" (UID: \"b72b541d-4661-44a7-a121-c2a8aef6db11\") " pod="openstack/horizon-77948db5bb-97qh6" Feb 27 08:09:09 crc kubenswrapper[4612]: I0227 08:09:09.078899 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76009243-16b3-4652-abfe-a6cdba363724-combined-ca-bundle\") pod \"horizon-6b985547d4-mkfpv\" (UID: \"76009243-16b3-4652-abfe-a6cdba363724\") " pod="openstack/horizon-6b985547d4-mkfpv" Feb 27 08:09:09 crc kubenswrapper[4612]: I0227 08:09:09.084297 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nhcmx\" (UniqueName: \"kubernetes.io/projected/b72b541d-4661-44a7-a121-c2a8aef6db11-kube-api-access-nhcmx\") pod \"horizon-77948db5bb-97qh6\" (UID: \"b72b541d-4661-44a7-a121-c2a8aef6db11\") " pod="openstack/horizon-77948db5bb-97qh6" Feb 27 08:09:09 crc kubenswrapper[4612]: I0227 08:09:09.105957 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6b985547d4-mkfpv" Feb 27 08:09:09 crc kubenswrapper[4612]: I0227 08:09:09.271178 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-77948db5bb-97qh6" Feb 27 08:09:09 crc kubenswrapper[4612]: I0227 08:09:09.388080 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"d62745a5-cfe3-48b0-a167-76735bdfd8fe","Type":"ContainerStarted","Data":"d7beccc144fa9bf9b73e2f015fc0afe88018fd5883a580e0623b4c591da30e02"} Feb 27 08:09:10 crc kubenswrapper[4612]: I0227 08:09:10.247896 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-56df8fb6b7-9p274" Feb 27 08:09:10 crc kubenswrapper[4612]: I0227 08:09:10.317767 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5f59b8f679-j8tjt"] Feb 27 08:09:10 crc kubenswrapper[4612]: I0227 08:09:10.317999 4612 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5f59b8f679-j8tjt" podUID="df886311-677c-4060-8bc7-90198481530b" containerName="dnsmasq-dns" containerID="cri-o://97b0c335965700a7bc21108de21fdffbd2d00631135a6c4d577a8c6f1acbd803" gracePeriod=10 Feb 27 08:09:11 crc kubenswrapper[4612]: I0227 08:09:11.422049 4612 generic.go:334] "Generic (PLEG): container finished" podID="df886311-677c-4060-8bc7-90198481530b" containerID="97b0c335965700a7bc21108de21fdffbd2d00631135a6c4d577a8c6f1acbd803" exitCode=0 Feb 27 08:09:11 crc kubenswrapper[4612]: I0227 08:09:11.422401 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5f59b8f679-j8tjt" event={"ID":"df886311-677c-4060-8bc7-90198481530b","Type":"ContainerDied","Data":"97b0c335965700a7bc21108de21fdffbd2d00631135a6c4d577a8c6f1acbd803"} Feb 27 08:09:13 crc kubenswrapper[4612]: I0227 08:09:13.237682 4612 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-5f59b8f679-j8tjt" podUID="df886311-677c-4060-8bc7-90198481530b" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.129:5353: connect: connection refused" Feb 27 08:09:14 crc kubenswrapper[4612]: I0227 08:09:14.105855 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-4nm8j" Feb 27 08:09:14 crc kubenswrapper[4612]: I0227 08:09:14.160366 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dfebce69-1679-4abd-8b9d-270046157f42-scripts\") pod \"dfebce69-1679-4abd-8b9d-270046157f42\" (UID: \"dfebce69-1679-4abd-8b9d-270046157f42\") " Feb 27 08:09:14 crc kubenswrapper[4612]: I0227 08:09:14.160473 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/dfebce69-1679-4abd-8b9d-270046157f42-fernet-keys\") pod \"dfebce69-1679-4abd-8b9d-270046157f42\" (UID: \"dfebce69-1679-4abd-8b9d-270046157f42\") " Feb 27 08:09:14 crc kubenswrapper[4612]: I0227 08:09:14.160621 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dfebce69-1679-4abd-8b9d-270046157f42-config-data\") pod \"dfebce69-1679-4abd-8b9d-270046157f42\" (UID: \"dfebce69-1679-4abd-8b9d-270046157f42\") " Feb 27 08:09:14 crc kubenswrapper[4612]: I0227 08:09:14.160721 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-flp5k\" (UniqueName: \"kubernetes.io/projected/dfebce69-1679-4abd-8b9d-270046157f42-kube-api-access-flp5k\") pod \"dfebce69-1679-4abd-8b9d-270046157f42\" (UID: \"dfebce69-1679-4abd-8b9d-270046157f42\") " Feb 27 08:09:14 crc kubenswrapper[4612]: I0227 08:09:14.160838 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dfebce69-1679-4abd-8b9d-270046157f42-combined-ca-bundle\") pod \"dfebce69-1679-4abd-8b9d-270046157f42\" (UID: \"dfebce69-1679-4abd-8b9d-270046157f42\") " Feb 27 08:09:14 crc kubenswrapper[4612]: I0227 08:09:14.160905 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/dfebce69-1679-4abd-8b9d-270046157f42-credential-keys\") pod \"dfebce69-1679-4abd-8b9d-270046157f42\" (UID: \"dfebce69-1679-4abd-8b9d-270046157f42\") " Feb 27 08:09:14 crc kubenswrapper[4612]: I0227 08:09:14.165442 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dfebce69-1679-4abd-8b9d-270046157f42-kube-api-access-flp5k" (OuterVolumeSpecName: "kube-api-access-flp5k") pod "dfebce69-1679-4abd-8b9d-270046157f42" (UID: "dfebce69-1679-4abd-8b9d-270046157f42"). InnerVolumeSpecName "kube-api-access-flp5k". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 08:09:14 crc kubenswrapper[4612]: I0227 08:09:14.167430 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dfebce69-1679-4abd-8b9d-270046157f42-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "dfebce69-1679-4abd-8b9d-270046157f42" (UID: "dfebce69-1679-4abd-8b9d-270046157f42"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:09:14 crc kubenswrapper[4612]: I0227 08:09:14.168043 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dfebce69-1679-4abd-8b9d-270046157f42-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "dfebce69-1679-4abd-8b9d-270046157f42" (UID: "dfebce69-1679-4abd-8b9d-270046157f42"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:09:14 crc kubenswrapper[4612]: I0227 08:09:14.183161 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dfebce69-1679-4abd-8b9d-270046157f42-scripts" (OuterVolumeSpecName: "scripts") pod "dfebce69-1679-4abd-8b9d-270046157f42" (UID: "dfebce69-1679-4abd-8b9d-270046157f42"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:09:14 crc kubenswrapper[4612]: I0227 08:09:14.194907 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dfebce69-1679-4abd-8b9d-270046157f42-config-data" (OuterVolumeSpecName: "config-data") pod "dfebce69-1679-4abd-8b9d-270046157f42" (UID: "dfebce69-1679-4abd-8b9d-270046157f42"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:09:14 crc kubenswrapper[4612]: I0227 08:09:14.207859 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dfebce69-1679-4abd-8b9d-270046157f42-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "dfebce69-1679-4abd-8b9d-270046157f42" (UID: "dfebce69-1679-4abd-8b9d-270046157f42"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:09:14 crc kubenswrapper[4612]: I0227 08:09:14.263131 4612 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dfebce69-1679-4abd-8b9d-270046157f42-scripts\") on node \"crc\" DevicePath \"\"" Feb 27 08:09:14 crc kubenswrapper[4612]: I0227 08:09:14.263166 4612 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/dfebce69-1679-4abd-8b9d-270046157f42-fernet-keys\") on node \"crc\" DevicePath \"\"" Feb 27 08:09:14 crc kubenswrapper[4612]: I0227 08:09:14.263179 4612 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dfebce69-1679-4abd-8b9d-270046157f42-config-data\") on node \"crc\" DevicePath \"\"" Feb 27 08:09:14 crc kubenswrapper[4612]: I0227 08:09:14.263187 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-flp5k\" (UniqueName: \"kubernetes.io/projected/dfebce69-1679-4abd-8b9d-270046157f42-kube-api-access-flp5k\") on node \"crc\" DevicePath \"\"" Feb 27 08:09:14 crc kubenswrapper[4612]: I0227 08:09:14.263196 4612 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dfebce69-1679-4abd-8b9d-270046157f42-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 08:09:14 crc kubenswrapper[4612]: I0227 08:09:14.263204 4612 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/dfebce69-1679-4abd-8b9d-270046157f42-credential-keys\") on node \"crc\" DevicePath \"\"" Feb 27 08:09:14 crc kubenswrapper[4612]: I0227 08:09:14.448368 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-4nm8j" event={"ID":"dfebce69-1679-4abd-8b9d-270046157f42","Type":"ContainerDied","Data":"4fdf68d99fd0d497fa371456c87bcc3cc9674d0be1100134047f677eea693c47"} Feb 27 08:09:14 crc kubenswrapper[4612]: I0227 08:09:14.448443 4612 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4fdf68d99fd0d497fa371456c87bcc3cc9674d0be1100134047f677eea693c47" Feb 27 08:09:14 crc kubenswrapper[4612]: I0227 08:09:14.448495 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-4nm8j" Feb 27 08:09:15 crc kubenswrapper[4612]: I0227 08:09:15.181319 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-4nm8j"] Feb 27 08:09:15 crc kubenswrapper[4612]: I0227 08:09:15.188475 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-4nm8j"] Feb 27 08:09:15 crc kubenswrapper[4612]: I0227 08:09:15.296485 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-s5bjv"] Feb 27 08:09:15 crc kubenswrapper[4612]: E0227 08:09:15.296831 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dfebce69-1679-4abd-8b9d-270046157f42" containerName="keystone-bootstrap" Feb 27 08:09:15 crc kubenswrapper[4612]: I0227 08:09:15.296843 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="dfebce69-1679-4abd-8b9d-270046157f42" containerName="keystone-bootstrap" Feb 27 08:09:15 crc kubenswrapper[4612]: I0227 08:09:15.296991 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="dfebce69-1679-4abd-8b9d-270046157f42" containerName="keystone-bootstrap" Feb 27 08:09:15 crc kubenswrapper[4612]: I0227 08:09:15.297495 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-s5bjv" Feb 27 08:09:15 crc kubenswrapper[4612]: I0227 08:09:15.301237 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Feb 27 08:09:15 crc kubenswrapper[4612]: I0227 08:09:15.301539 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Feb 27 08:09:15 crc kubenswrapper[4612]: I0227 08:09:15.301755 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-zdkwc" Feb 27 08:09:15 crc kubenswrapper[4612]: I0227 08:09:15.304347 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Feb 27 08:09:15 crc kubenswrapper[4612]: I0227 08:09:15.314470 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Feb 27 08:09:15 crc kubenswrapper[4612]: I0227 08:09:15.319200 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-s5bjv"] Feb 27 08:09:15 crc kubenswrapper[4612]: I0227 08:09:15.385131 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1a05a7e5-8c46-441a-9faf-0fd3b41bbd98-scripts\") pod \"keystone-bootstrap-s5bjv\" (UID: \"1a05a7e5-8c46-441a-9faf-0fd3b41bbd98\") " pod="openstack/keystone-bootstrap-s5bjv" Feb 27 08:09:15 crc kubenswrapper[4612]: I0227 08:09:15.385277 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a05a7e5-8c46-441a-9faf-0fd3b41bbd98-combined-ca-bundle\") pod \"keystone-bootstrap-s5bjv\" (UID: \"1a05a7e5-8c46-441a-9faf-0fd3b41bbd98\") " pod="openstack/keystone-bootstrap-s5bjv" Feb 27 08:09:15 crc kubenswrapper[4612]: I0227 08:09:15.385307 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1a05a7e5-8c46-441a-9faf-0fd3b41bbd98-config-data\") pod \"keystone-bootstrap-s5bjv\" (UID: \"1a05a7e5-8c46-441a-9faf-0fd3b41bbd98\") " pod="openstack/keystone-bootstrap-s5bjv" Feb 27 08:09:15 crc kubenswrapper[4612]: I0227 08:09:15.385369 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b6hqv\" (UniqueName: \"kubernetes.io/projected/1a05a7e5-8c46-441a-9faf-0fd3b41bbd98-kube-api-access-b6hqv\") pod \"keystone-bootstrap-s5bjv\" (UID: \"1a05a7e5-8c46-441a-9faf-0fd3b41bbd98\") " pod="openstack/keystone-bootstrap-s5bjv" Feb 27 08:09:15 crc kubenswrapper[4612]: I0227 08:09:15.385392 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/1a05a7e5-8c46-441a-9faf-0fd3b41bbd98-credential-keys\") pod \"keystone-bootstrap-s5bjv\" (UID: \"1a05a7e5-8c46-441a-9faf-0fd3b41bbd98\") " pod="openstack/keystone-bootstrap-s5bjv" Feb 27 08:09:15 crc kubenswrapper[4612]: I0227 08:09:15.385446 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/1a05a7e5-8c46-441a-9faf-0fd3b41bbd98-fernet-keys\") pod \"keystone-bootstrap-s5bjv\" (UID: \"1a05a7e5-8c46-441a-9faf-0fd3b41bbd98\") " pod="openstack/keystone-bootstrap-s5bjv" Feb 27 08:09:15 crc kubenswrapper[4612]: I0227 08:09:15.486349 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1a05a7e5-8c46-441a-9faf-0fd3b41bbd98-scripts\") pod \"keystone-bootstrap-s5bjv\" (UID: \"1a05a7e5-8c46-441a-9faf-0fd3b41bbd98\") " pod="openstack/keystone-bootstrap-s5bjv" Feb 27 08:09:15 crc kubenswrapper[4612]: I0227 08:09:15.486494 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a05a7e5-8c46-441a-9faf-0fd3b41bbd98-combined-ca-bundle\") pod \"keystone-bootstrap-s5bjv\" (UID: \"1a05a7e5-8c46-441a-9faf-0fd3b41bbd98\") " pod="openstack/keystone-bootstrap-s5bjv" Feb 27 08:09:15 crc kubenswrapper[4612]: I0227 08:09:15.486526 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1a05a7e5-8c46-441a-9faf-0fd3b41bbd98-config-data\") pod \"keystone-bootstrap-s5bjv\" (UID: \"1a05a7e5-8c46-441a-9faf-0fd3b41bbd98\") " pod="openstack/keystone-bootstrap-s5bjv" Feb 27 08:09:15 crc kubenswrapper[4612]: I0227 08:09:15.486589 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b6hqv\" (UniqueName: \"kubernetes.io/projected/1a05a7e5-8c46-441a-9faf-0fd3b41bbd98-kube-api-access-b6hqv\") pod \"keystone-bootstrap-s5bjv\" (UID: \"1a05a7e5-8c46-441a-9faf-0fd3b41bbd98\") " pod="openstack/keystone-bootstrap-s5bjv" Feb 27 08:09:15 crc kubenswrapper[4612]: I0227 08:09:15.486620 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/1a05a7e5-8c46-441a-9faf-0fd3b41bbd98-credential-keys\") pod \"keystone-bootstrap-s5bjv\" (UID: \"1a05a7e5-8c46-441a-9faf-0fd3b41bbd98\") " pod="openstack/keystone-bootstrap-s5bjv" Feb 27 08:09:15 crc kubenswrapper[4612]: I0227 08:09:15.486669 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/1a05a7e5-8c46-441a-9faf-0fd3b41bbd98-fernet-keys\") pod \"keystone-bootstrap-s5bjv\" (UID: \"1a05a7e5-8c46-441a-9faf-0fd3b41bbd98\") " pod="openstack/keystone-bootstrap-s5bjv" Feb 27 08:09:15 crc kubenswrapper[4612]: I0227 08:09:15.491612 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/1a05a7e5-8c46-441a-9faf-0fd3b41bbd98-credential-keys\") pod \"keystone-bootstrap-s5bjv\" (UID: \"1a05a7e5-8c46-441a-9faf-0fd3b41bbd98\") " pod="openstack/keystone-bootstrap-s5bjv" Feb 27 08:09:15 crc kubenswrapper[4612]: I0227 08:09:15.493283 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a05a7e5-8c46-441a-9faf-0fd3b41bbd98-combined-ca-bundle\") pod \"keystone-bootstrap-s5bjv\" (UID: \"1a05a7e5-8c46-441a-9faf-0fd3b41bbd98\") " pod="openstack/keystone-bootstrap-s5bjv" Feb 27 08:09:15 crc kubenswrapper[4612]: I0227 08:09:15.493477 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/1a05a7e5-8c46-441a-9faf-0fd3b41bbd98-fernet-keys\") pod \"keystone-bootstrap-s5bjv\" (UID: \"1a05a7e5-8c46-441a-9faf-0fd3b41bbd98\") " pod="openstack/keystone-bootstrap-s5bjv" Feb 27 08:09:15 crc kubenswrapper[4612]: I0227 08:09:15.494604 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1a05a7e5-8c46-441a-9faf-0fd3b41bbd98-config-data\") pod \"keystone-bootstrap-s5bjv\" (UID: \"1a05a7e5-8c46-441a-9faf-0fd3b41bbd98\") " pod="openstack/keystone-bootstrap-s5bjv" Feb 27 08:09:15 crc kubenswrapper[4612]: I0227 08:09:15.500760 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1a05a7e5-8c46-441a-9faf-0fd3b41bbd98-scripts\") pod \"keystone-bootstrap-s5bjv\" (UID: \"1a05a7e5-8c46-441a-9faf-0fd3b41bbd98\") " pod="openstack/keystone-bootstrap-s5bjv" Feb 27 08:09:15 crc kubenswrapper[4612]: I0227 08:09:15.503764 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b6hqv\" (UniqueName: \"kubernetes.io/projected/1a05a7e5-8c46-441a-9faf-0fd3b41bbd98-kube-api-access-b6hqv\") pod \"keystone-bootstrap-s5bjv\" (UID: \"1a05a7e5-8c46-441a-9faf-0fd3b41bbd98\") " pod="openstack/keystone-bootstrap-s5bjv" Feb 27 08:09:15 crc kubenswrapper[4612]: I0227 08:09:15.629090 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-s5bjv" Feb 27 08:09:16 crc kubenswrapper[4612]: I0227 08:09:16.867562 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dfebce69-1679-4abd-8b9d-270046157f42" path="/var/lib/kubelet/pods/dfebce69-1679-4abd-8b9d-270046157f42/volumes" Feb 27 08:09:18 crc kubenswrapper[4612]: I0227 08:09:18.238176 4612 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-5f59b8f679-j8tjt" podUID="df886311-677c-4060-8bc7-90198481530b" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.129:5353: connect: connection refused" Feb 27 08:09:20 crc kubenswrapper[4612]: E0227 08:09:20.918170 4612 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-horizon:current-podified" Feb 27 08:09:20 crc kubenswrapper[4612]: E0227 08:09:20.918713 4612 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:quay.io/podified-antelope-centos9/openstack-horizon:current-podified,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n67dh6dh5bfh89h64ch568hb9hcch597h5h57ch5c8h78h5cch57bh678h67ch5dbh85hb7h5fbhd5h8dh598hf6h544hcchcch549h5d8h5ddhdcq,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:no,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-t2rpv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-58b79554cc-f9ms6_openstack(53ff41b8-ca85-4358-a8fe-cb490ec40793): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 27 08:09:20 crc kubenswrapper[4612]: E0227 08:09:20.928459 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizon:current-podified\\\"\"]" pod="openstack/horizon-58b79554cc-f9ms6" podUID="53ff41b8-ca85-4358-a8fe-cb490ec40793" Feb 27 08:09:22 crc kubenswrapper[4612]: W0227 08:09:22.655154 4612 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4726ffcd_f407_4cd4_a6e4_c422801f2617.slice/crio-a474aa8d3910d92396577ba685fe0afa2b21b30098d34cdf06cdf8e1e660a9c4 WatchSource:0}: Error finding container a474aa8d3910d92396577ba685fe0afa2b21b30098d34cdf06cdf8e1e660a9c4: Status 404 returned error can't find the container with id a474aa8d3910d92396577ba685fe0afa2b21b30098d34cdf06cdf8e1e660a9c4 Feb 27 08:09:23 crc kubenswrapper[4612]: E0227 08:09:23.501238 4612 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified" Feb 27 08:09:23 crc kubenswrapper[4612]: E0227 08:09:23.501502 4612 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:barbican-db-sync,Image:quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified,Command:[/bin/bash],Args:[-c barbican-manage db upgrade],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/barbican/barbican.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-625qb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42403,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42403,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod barbican-db-sync-r8dpp_openstack(87df2a96-7ae3-47af-92bb-97db038b957c): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 27 08:09:23 crc kubenswrapper[4612]: E0227 08:09:23.502716 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/barbican-db-sync-r8dpp" podUID="87df2a96-7ae3-47af-92bb-97db038b957c" Feb 27 08:09:23 crc kubenswrapper[4612]: I0227 08:09:23.523999 4612 generic.go:334] "Generic (PLEG): container finished" podID="f63a4b12-f7e9-43a4-91d2-fd46f65daf92" containerID="8b548933be8af10568a38da48000e3a51d2fe9ff3b04b577671c700e4196b621" exitCode=0 Feb 27 08:09:23 crc kubenswrapper[4612]: I0227 08:09:23.524068 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-4ljbd" event={"ID":"f63a4b12-f7e9-43a4-91d2-fd46f65daf92","Type":"ContainerDied","Data":"8b548933be8af10568a38da48000e3a51d2fe9ff3b04b577671c700e4196b621"} Feb 27 08:09:23 crc kubenswrapper[4612]: I0227 08:09:23.527111 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"4726ffcd-f407-4cd4-a6e4-c422801f2617","Type":"ContainerStarted","Data":"a474aa8d3910d92396577ba685fe0afa2b21b30098d34cdf06cdf8e1e660a9c4"} Feb 27 08:09:23 crc kubenswrapper[4612]: E0227 08:09:23.535347 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified\\\"\"" pod="openstack/barbican-db-sync-r8dpp" podUID="87df2a96-7ae3-47af-92bb-97db038b957c" Feb 27 08:09:23 crc kubenswrapper[4612]: I0227 08:09:23.622464 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5f59b8f679-j8tjt" Feb 27 08:09:23 crc kubenswrapper[4612]: I0227 08:09:23.726853 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/df886311-677c-4060-8bc7-90198481530b-dns-swift-storage-0\") pod \"df886311-677c-4060-8bc7-90198481530b\" (UID: \"df886311-677c-4060-8bc7-90198481530b\") " Feb 27 08:09:23 crc kubenswrapper[4612]: I0227 08:09:23.727364 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/df886311-677c-4060-8bc7-90198481530b-ovsdbserver-nb\") pod \"df886311-677c-4060-8bc7-90198481530b\" (UID: \"df886311-677c-4060-8bc7-90198481530b\") " Feb 27 08:09:23 crc kubenswrapper[4612]: I0227 08:09:23.727415 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/df886311-677c-4060-8bc7-90198481530b-ovsdbserver-sb\") pod \"df886311-677c-4060-8bc7-90198481530b\" (UID: \"df886311-677c-4060-8bc7-90198481530b\") " Feb 27 08:09:23 crc kubenswrapper[4612]: I0227 08:09:23.727460 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/df886311-677c-4060-8bc7-90198481530b-config\") pod \"df886311-677c-4060-8bc7-90198481530b\" (UID: \"df886311-677c-4060-8bc7-90198481530b\") " Feb 27 08:09:23 crc kubenswrapper[4612]: I0227 08:09:23.727486 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/df886311-677c-4060-8bc7-90198481530b-dns-svc\") pod \"df886311-677c-4060-8bc7-90198481530b\" (UID: \"df886311-677c-4060-8bc7-90198481530b\") " Feb 27 08:09:23 crc kubenswrapper[4612]: I0227 08:09:23.727599 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q2kt8\" (UniqueName: \"kubernetes.io/projected/df886311-677c-4060-8bc7-90198481530b-kube-api-access-q2kt8\") pod \"df886311-677c-4060-8bc7-90198481530b\" (UID: \"df886311-677c-4060-8bc7-90198481530b\") " Feb 27 08:09:23 crc kubenswrapper[4612]: I0227 08:09:23.740979 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/df886311-677c-4060-8bc7-90198481530b-kube-api-access-q2kt8" (OuterVolumeSpecName: "kube-api-access-q2kt8") pod "df886311-677c-4060-8bc7-90198481530b" (UID: "df886311-677c-4060-8bc7-90198481530b"). InnerVolumeSpecName "kube-api-access-q2kt8". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 08:09:23 crc kubenswrapper[4612]: I0227 08:09:23.776703 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/df886311-677c-4060-8bc7-90198481530b-config" (OuterVolumeSpecName: "config") pod "df886311-677c-4060-8bc7-90198481530b" (UID: "df886311-677c-4060-8bc7-90198481530b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 08:09:23 crc kubenswrapper[4612]: I0227 08:09:23.780811 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/df886311-677c-4060-8bc7-90198481530b-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "df886311-677c-4060-8bc7-90198481530b" (UID: "df886311-677c-4060-8bc7-90198481530b"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 08:09:23 crc kubenswrapper[4612]: I0227 08:09:23.785315 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/df886311-677c-4060-8bc7-90198481530b-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "df886311-677c-4060-8bc7-90198481530b" (UID: "df886311-677c-4060-8bc7-90198481530b"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 08:09:23 crc kubenswrapper[4612]: I0227 08:09:23.788118 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/df886311-677c-4060-8bc7-90198481530b-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "df886311-677c-4060-8bc7-90198481530b" (UID: "df886311-677c-4060-8bc7-90198481530b"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 08:09:23 crc kubenswrapper[4612]: I0227 08:09:23.799367 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/df886311-677c-4060-8bc7-90198481530b-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "df886311-677c-4060-8bc7-90198481530b" (UID: "df886311-677c-4060-8bc7-90198481530b"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 08:09:23 crc kubenswrapper[4612]: I0227 08:09:23.830030 4612 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/df886311-677c-4060-8bc7-90198481530b-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 27 08:09:23 crc kubenswrapper[4612]: I0227 08:09:23.830064 4612 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/df886311-677c-4060-8bc7-90198481530b-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 27 08:09:23 crc kubenswrapper[4612]: I0227 08:09:23.830074 4612 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/df886311-677c-4060-8bc7-90198481530b-config\") on node \"crc\" DevicePath \"\"" Feb 27 08:09:23 crc kubenswrapper[4612]: I0227 08:09:23.830082 4612 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/df886311-677c-4060-8bc7-90198481530b-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 27 08:09:23 crc kubenswrapper[4612]: I0227 08:09:23.830092 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q2kt8\" (UniqueName: \"kubernetes.io/projected/df886311-677c-4060-8bc7-90198481530b-kube-api-access-q2kt8\") on node \"crc\" DevicePath \"\"" Feb 27 08:09:23 crc kubenswrapper[4612]: I0227 08:09:23.830102 4612 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/df886311-677c-4060-8bc7-90198481530b-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Feb 27 08:09:24 crc kubenswrapper[4612]: I0227 08:09:24.551750 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5f59b8f679-j8tjt" event={"ID":"df886311-677c-4060-8bc7-90198481530b","Type":"ContainerDied","Data":"a290ea972eddf4b4c16051587218a705c9b71b91918480048ea775ae359439fa"} Feb 27 08:09:24 crc kubenswrapper[4612]: I0227 08:09:24.551809 4612 scope.go:117] "RemoveContainer" containerID="97b0c335965700a7bc21108de21fdffbd2d00631135a6c4d577a8c6f1acbd803" Feb 27 08:09:24 crc kubenswrapper[4612]: I0227 08:09:24.551823 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5f59b8f679-j8tjt" Feb 27 08:09:24 crc kubenswrapper[4612]: I0227 08:09:24.594604 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5f59b8f679-j8tjt"] Feb 27 08:09:24 crc kubenswrapper[4612]: I0227 08:09:24.602000 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5f59b8f679-j8tjt"] Feb 27 08:09:24 crc kubenswrapper[4612]: I0227 08:09:24.865126 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="df886311-677c-4060-8bc7-90198481530b" path="/var/lib/kubelet/pods/df886311-677c-4060-8bc7-90198481530b/volumes" Feb 27 08:09:28 crc kubenswrapper[4612]: I0227 08:09:28.238520 4612 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-5f59b8f679-j8tjt" podUID="df886311-677c-4060-8bc7-90198481530b" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.129:5353: i/o timeout" Feb 27 08:09:33 crc kubenswrapper[4612]: I0227 08:09:33.041275 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-58b79554cc-f9ms6" Feb 27 08:09:33 crc kubenswrapper[4612]: E0227 08:09:33.109349 4612 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-horizon:current-podified" Feb 27 08:09:33 crc kubenswrapper[4612]: E0227 08:09:33.109557 4612 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:quay.io/podified-antelope-centos9/openstack-horizon:current-podified,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n678h648h677hd7h5b9h556hfh5c4h98h7ch558h5f4hb4h65fh5b6h5b7h79hdbh77hdfh54bhdfh56dh9ch544hdfh65h67ch7chf9h5dh8q,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:no,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-zpvms,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-594c9f8cf-q282v_openstack(f37f5ce9-74c3-48fc-9080-ae81464dba49): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 27 08:09:33 crc kubenswrapper[4612]: E0227 08:09:33.111862 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizon:current-podified\\\"\"]" pod="openstack/horizon-594c9f8cf-q282v" podUID="f37f5ce9-74c3-48fc-9080-ae81464dba49" Feb 27 08:09:33 crc kubenswrapper[4612]: I0227 08:09:33.214371 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/53ff41b8-ca85-4358-a8fe-cb490ec40793-logs\") pod \"53ff41b8-ca85-4358-a8fe-cb490ec40793\" (UID: \"53ff41b8-ca85-4358-a8fe-cb490ec40793\") " Feb 27 08:09:33 crc kubenswrapper[4612]: I0227 08:09:33.214468 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/53ff41b8-ca85-4358-a8fe-cb490ec40793-config-data\") pod \"53ff41b8-ca85-4358-a8fe-cb490ec40793\" (UID: \"53ff41b8-ca85-4358-a8fe-cb490ec40793\") " Feb 27 08:09:33 crc kubenswrapper[4612]: I0227 08:09:33.214547 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/53ff41b8-ca85-4358-a8fe-cb490ec40793-scripts\") pod \"53ff41b8-ca85-4358-a8fe-cb490ec40793\" (UID: \"53ff41b8-ca85-4358-a8fe-cb490ec40793\") " Feb 27 08:09:33 crc kubenswrapper[4612]: I0227 08:09:33.214596 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t2rpv\" (UniqueName: \"kubernetes.io/projected/53ff41b8-ca85-4358-a8fe-cb490ec40793-kube-api-access-t2rpv\") pod \"53ff41b8-ca85-4358-a8fe-cb490ec40793\" (UID: \"53ff41b8-ca85-4358-a8fe-cb490ec40793\") " Feb 27 08:09:33 crc kubenswrapper[4612]: I0227 08:09:33.214649 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/53ff41b8-ca85-4358-a8fe-cb490ec40793-horizon-secret-key\") pod \"53ff41b8-ca85-4358-a8fe-cb490ec40793\" (UID: \"53ff41b8-ca85-4358-a8fe-cb490ec40793\") " Feb 27 08:09:33 crc kubenswrapper[4612]: I0227 08:09:33.214862 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/53ff41b8-ca85-4358-a8fe-cb490ec40793-logs" (OuterVolumeSpecName: "logs") pod "53ff41b8-ca85-4358-a8fe-cb490ec40793" (UID: "53ff41b8-ca85-4358-a8fe-cb490ec40793"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 08:09:33 crc kubenswrapper[4612]: I0227 08:09:33.215120 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/53ff41b8-ca85-4358-a8fe-cb490ec40793-scripts" (OuterVolumeSpecName: "scripts") pod "53ff41b8-ca85-4358-a8fe-cb490ec40793" (UID: "53ff41b8-ca85-4358-a8fe-cb490ec40793"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 08:09:33 crc kubenswrapper[4612]: I0227 08:09:33.215185 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/53ff41b8-ca85-4358-a8fe-cb490ec40793-config-data" (OuterVolumeSpecName: "config-data") pod "53ff41b8-ca85-4358-a8fe-cb490ec40793" (UID: "53ff41b8-ca85-4358-a8fe-cb490ec40793"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 08:09:33 crc kubenswrapper[4612]: I0227 08:09:33.216176 4612 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/53ff41b8-ca85-4358-a8fe-cb490ec40793-logs\") on node \"crc\" DevicePath \"\"" Feb 27 08:09:33 crc kubenswrapper[4612]: I0227 08:09:33.216202 4612 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/53ff41b8-ca85-4358-a8fe-cb490ec40793-config-data\") on node \"crc\" DevicePath \"\"" Feb 27 08:09:33 crc kubenswrapper[4612]: I0227 08:09:33.216213 4612 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/53ff41b8-ca85-4358-a8fe-cb490ec40793-scripts\") on node \"crc\" DevicePath \"\"" Feb 27 08:09:33 crc kubenswrapper[4612]: I0227 08:09:33.222117 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/53ff41b8-ca85-4358-a8fe-cb490ec40793-kube-api-access-t2rpv" (OuterVolumeSpecName: "kube-api-access-t2rpv") pod "53ff41b8-ca85-4358-a8fe-cb490ec40793" (UID: "53ff41b8-ca85-4358-a8fe-cb490ec40793"). InnerVolumeSpecName "kube-api-access-t2rpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 08:09:33 crc kubenswrapper[4612]: I0227 08:09:33.227905 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/53ff41b8-ca85-4358-a8fe-cb490ec40793-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "53ff41b8-ca85-4358-a8fe-cb490ec40793" (UID: "53ff41b8-ca85-4358-a8fe-cb490ec40793"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:09:33 crc kubenswrapper[4612]: I0227 08:09:33.318093 4612 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/53ff41b8-ca85-4358-a8fe-cb490ec40793-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Feb 27 08:09:33 crc kubenswrapper[4612]: I0227 08:09:33.318127 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t2rpv\" (UniqueName: \"kubernetes.io/projected/53ff41b8-ca85-4358-a8fe-cb490ec40793-kube-api-access-t2rpv\") on node \"crc\" DevicePath \"\"" Feb 27 08:09:33 crc kubenswrapper[4612]: I0227 08:09:33.632293 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-58b79554cc-f9ms6" Feb 27 08:09:33 crc kubenswrapper[4612]: I0227 08:09:33.632303 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-58b79554cc-f9ms6" event={"ID":"53ff41b8-ca85-4358-a8fe-cb490ec40793","Type":"ContainerDied","Data":"5272d46ea77a4e7e73f94df6aed6387d54219176f3bcd356974de1c682c0856a"} Feb 27 08:09:33 crc kubenswrapper[4612]: I0227 08:09:33.723995 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-58b79554cc-f9ms6"] Feb 27 08:09:33 crc kubenswrapper[4612]: I0227 08:09:33.732115 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-58b79554cc-f9ms6"] Feb 27 08:09:33 crc kubenswrapper[4612]: E0227 08:09:33.872245 4612 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-ceilometer-central:current-podified" Feb 27 08:09:33 crc kubenswrapper[4612]: E0227 08:09:33.872435 4612 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:ceilometer-central-agent,Image:quay.io/podified-antelope-centos9/openstack-ceilometer-central:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n5f7h56bh594h88h98h584h5d9hdch676h75h65dh59h5d7hcdh67bh654h697h58fh88h8dhfdh656h8bh586h5b7h5dfh56dh544hc9h68ch5f5h57bq,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/var/lib/openstack/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:ceilometer-central-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-sh9th,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/python3 /var/lib/openstack/bin/centralhealth.py],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:300,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ceilometer-0_openstack(9ab32d5a-a797-47a8-961a-a534f55fdfa7): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 27 08:09:33 crc kubenswrapper[4612]: E0227 08:09:33.881568 4612 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-horizon:current-podified" Feb 27 08:09:33 crc kubenswrapper[4612]: E0227 08:09:33.881830 4612 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:quay.io/podified-antelope-centos9/openstack-horizon:current-podified,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n5d9h8hc7h9h7bh5cfh87hf8h97hddh556h669h8fh64ch664h678h55ch67h95hc7h68dh5c5h596h597h5dbh6ch57dh5b8h55ch5c7h6ch99q,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:no,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-9zthz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-7f9b6cd66c-5lwzw_openstack(a10f6a85-63bd-4aa7-91a6-b3aec6ecb9d4): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 27 08:09:33 crc kubenswrapper[4612]: E0227 08:09:33.886149 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizon:current-podified\\\"\"]" pod="openstack/horizon-7f9b6cd66c-5lwzw" podUID="a10f6a85-63bd-4aa7-91a6-b3aec6ecb9d4" Feb 27 08:09:33 crc kubenswrapper[4612]: I0227 08:09:33.932728 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-4ljbd" Feb 27 08:09:34 crc kubenswrapper[4612]: I0227 08:09:34.029287 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f63a4b12-f7e9-43a4-91d2-fd46f65daf92-combined-ca-bundle\") pod \"f63a4b12-f7e9-43a4-91d2-fd46f65daf92\" (UID: \"f63a4b12-f7e9-43a4-91d2-fd46f65daf92\") " Feb 27 08:09:34 crc kubenswrapper[4612]: I0227 08:09:34.029403 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/f63a4b12-f7e9-43a4-91d2-fd46f65daf92-config\") pod \"f63a4b12-f7e9-43a4-91d2-fd46f65daf92\" (UID: \"f63a4b12-f7e9-43a4-91d2-fd46f65daf92\") " Feb 27 08:09:34 crc kubenswrapper[4612]: I0227 08:09:34.029452 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zrxvv\" (UniqueName: \"kubernetes.io/projected/f63a4b12-f7e9-43a4-91d2-fd46f65daf92-kube-api-access-zrxvv\") pod \"f63a4b12-f7e9-43a4-91d2-fd46f65daf92\" (UID: \"f63a4b12-f7e9-43a4-91d2-fd46f65daf92\") " Feb 27 08:09:34 crc kubenswrapper[4612]: I0227 08:09:34.035098 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f63a4b12-f7e9-43a4-91d2-fd46f65daf92-kube-api-access-zrxvv" (OuterVolumeSpecName: "kube-api-access-zrxvv") pod "f63a4b12-f7e9-43a4-91d2-fd46f65daf92" (UID: "f63a4b12-f7e9-43a4-91d2-fd46f65daf92"). InnerVolumeSpecName "kube-api-access-zrxvv". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 08:09:34 crc kubenswrapper[4612]: I0227 08:09:34.055708 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f63a4b12-f7e9-43a4-91d2-fd46f65daf92-config" (OuterVolumeSpecName: "config") pod "f63a4b12-f7e9-43a4-91d2-fd46f65daf92" (UID: "f63a4b12-f7e9-43a4-91d2-fd46f65daf92"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:09:34 crc kubenswrapper[4612]: I0227 08:09:34.061882 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f63a4b12-f7e9-43a4-91d2-fd46f65daf92-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f63a4b12-f7e9-43a4-91d2-fd46f65daf92" (UID: "f63a4b12-f7e9-43a4-91d2-fd46f65daf92"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:09:34 crc kubenswrapper[4612]: I0227 08:09:34.131391 4612 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f63a4b12-f7e9-43a4-91d2-fd46f65daf92-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 08:09:34 crc kubenswrapper[4612]: I0227 08:09:34.131426 4612 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/f63a4b12-f7e9-43a4-91d2-fd46f65daf92-config\") on node \"crc\" DevicePath \"\"" Feb 27 08:09:34 crc kubenswrapper[4612]: I0227 08:09:34.131437 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zrxvv\" (UniqueName: \"kubernetes.io/projected/f63a4b12-f7e9-43a4-91d2-fd46f65daf92-kube-api-access-zrxvv\") on node \"crc\" DevicePath \"\"" Feb 27 08:09:34 crc kubenswrapper[4612]: I0227 08:09:34.643550 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-4ljbd" event={"ID":"f63a4b12-f7e9-43a4-91d2-fd46f65daf92","Type":"ContainerDied","Data":"e12eddf1ed7f59c3a3f7e5ee4f4984107cdd6c22f0eebd0205e1ad560b240c98"} Feb 27 08:09:34 crc kubenswrapper[4612]: I0227 08:09:34.643578 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-4ljbd" Feb 27 08:09:34 crc kubenswrapper[4612]: I0227 08:09:34.643591 4612 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e12eddf1ed7f59c3a3f7e5ee4f4984107cdd6c22f0eebd0205e1ad560b240c98" Feb 27 08:09:34 crc kubenswrapper[4612]: I0227 08:09:34.861772 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="53ff41b8-ca85-4358-a8fe-cb490ec40793" path="/var/lib/kubelet/pods/53ff41b8-ca85-4358-a8fe-cb490ec40793/volumes" Feb 27 08:09:35 crc kubenswrapper[4612]: I0227 08:09:35.106505 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6b7b667979-pm5fl"] Feb 27 08:09:35 crc kubenswrapper[4612]: E0227 08:09:35.106884 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f63a4b12-f7e9-43a4-91d2-fd46f65daf92" containerName="neutron-db-sync" Feb 27 08:09:35 crc kubenswrapper[4612]: I0227 08:09:35.106895 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="f63a4b12-f7e9-43a4-91d2-fd46f65daf92" containerName="neutron-db-sync" Feb 27 08:09:35 crc kubenswrapper[4612]: E0227 08:09:35.106913 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="df886311-677c-4060-8bc7-90198481530b" containerName="init" Feb 27 08:09:35 crc kubenswrapper[4612]: I0227 08:09:35.106920 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="df886311-677c-4060-8bc7-90198481530b" containerName="init" Feb 27 08:09:35 crc kubenswrapper[4612]: E0227 08:09:35.106933 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="df886311-677c-4060-8bc7-90198481530b" containerName="dnsmasq-dns" Feb 27 08:09:35 crc kubenswrapper[4612]: I0227 08:09:35.106940 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="df886311-677c-4060-8bc7-90198481530b" containerName="dnsmasq-dns" Feb 27 08:09:35 crc kubenswrapper[4612]: I0227 08:09:35.107098 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="df886311-677c-4060-8bc7-90198481530b" containerName="dnsmasq-dns" Feb 27 08:09:35 crc kubenswrapper[4612]: I0227 08:09:35.107109 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="f63a4b12-f7e9-43a4-91d2-fd46f65daf92" containerName="neutron-db-sync" Feb 27 08:09:35 crc kubenswrapper[4612]: I0227 08:09:35.107938 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b7b667979-pm5fl" Feb 27 08:09:35 crc kubenswrapper[4612]: I0227 08:09:35.128934 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6b7b667979-pm5fl"] Feb 27 08:09:35 crc kubenswrapper[4612]: I0227 08:09:35.168588 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bf325571-0b74-447f-bd04-213d54a8437e-config\") pod \"dnsmasq-dns-6b7b667979-pm5fl\" (UID: \"bf325571-0b74-447f-bd04-213d54a8437e\") " pod="openstack/dnsmasq-dns-6b7b667979-pm5fl" Feb 27 08:09:35 crc kubenswrapper[4612]: I0227 08:09:35.168667 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bf325571-0b74-447f-bd04-213d54a8437e-ovsdbserver-nb\") pod \"dnsmasq-dns-6b7b667979-pm5fl\" (UID: \"bf325571-0b74-447f-bd04-213d54a8437e\") " pod="openstack/dnsmasq-dns-6b7b667979-pm5fl" Feb 27 08:09:35 crc kubenswrapper[4612]: I0227 08:09:35.168705 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8zzrm\" (UniqueName: \"kubernetes.io/projected/bf325571-0b74-447f-bd04-213d54a8437e-kube-api-access-8zzrm\") pod \"dnsmasq-dns-6b7b667979-pm5fl\" (UID: \"bf325571-0b74-447f-bd04-213d54a8437e\") " pod="openstack/dnsmasq-dns-6b7b667979-pm5fl" Feb 27 08:09:35 crc kubenswrapper[4612]: I0227 08:09:35.168760 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/bf325571-0b74-447f-bd04-213d54a8437e-dns-swift-storage-0\") pod \"dnsmasq-dns-6b7b667979-pm5fl\" (UID: \"bf325571-0b74-447f-bd04-213d54a8437e\") " pod="openstack/dnsmasq-dns-6b7b667979-pm5fl" Feb 27 08:09:35 crc kubenswrapper[4612]: I0227 08:09:35.168780 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bf325571-0b74-447f-bd04-213d54a8437e-ovsdbserver-sb\") pod \"dnsmasq-dns-6b7b667979-pm5fl\" (UID: \"bf325571-0b74-447f-bd04-213d54a8437e\") " pod="openstack/dnsmasq-dns-6b7b667979-pm5fl" Feb 27 08:09:35 crc kubenswrapper[4612]: I0227 08:09:35.168796 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bf325571-0b74-447f-bd04-213d54a8437e-dns-svc\") pod \"dnsmasq-dns-6b7b667979-pm5fl\" (UID: \"bf325571-0b74-447f-bd04-213d54a8437e\") " pod="openstack/dnsmasq-dns-6b7b667979-pm5fl" Feb 27 08:09:35 crc kubenswrapper[4612]: I0227 08:09:35.254854 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-77854b564-8nkbp"] Feb 27 08:09:35 crc kubenswrapper[4612]: I0227 08:09:35.256357 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-77854b564-8nkbp" Feb 27 08:09:35 crc kubenswrapper[4612]: I0227 08:09:35.259175 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Feb 27 08:09:35 crc kubenswrapper[4612]: I0227 08:09:35.265470 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Feb 27 08:09:35 crc kubenswrapper[4612]: I0227 08:09:35.268545 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-2j2mr" Feb 27 08:09:35 crc kubenswrapper[4612]: I0227 08:09:35.269469 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Feb 27 08:09:35 crc kubenswrapper[4612]: I0227 08:09:35.270254 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bf325571-0b74-447f-bd04-213d54a8437e-config\") pod \"dnsmasq-dns-6b7b667979-pm5fl\" (UID: \"bf325571-0b74-447f-bd04-213d54a8437e\") " pod="openstack/dnsmasq-dns-6b7b667979-pm5fl" Feb 27 08:09:35 crc kubenswrapper[4612]: I0227 08:09:35.270360 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bf325571-0b74-447f-bd04-213d54a8437e-ovsdbserver-nb\") pod \"dnsmasq-dns-6b7b667979-pm5fl\" (UID: \"bf325571-0b74-447f-bd04-213d54a8437e\") " pod="openstack/dnsmasq-dns-6b7b667979-pm5fl" Feb 27 08:09:35 crc kubenswrapper[4612]: I0227 08:09:35.270385 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8zzrm\" (UniqueName: \"kubernetes.io/projected/bf325571-0b74-447f-bd04-213d54a8437e-kube-api-access-8zzrm\") pod \"dnsmasq-dns-6b7b667979-pm5fl\" (UID: \"bf325571-0b74-447f-bd04-213d54a8437e\") " pod="openstack/dnsmasq-dns-6b7b667979-pm5fl" Feb 27 08:09:35 crc kubenswrapper[4612]: I0227 08:09:35.270445 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/bf325571-0b74-447f-bd04-213d54a8437e-dns-swift-storage-0\") pod \"dnsmasq-dns-6b7b667979-pm5fl\" (UID: \"bf325571-0b74-447f-bd04-213d54a8437e\") " pod="openstack/dnsmasq-dns-6b7b667979-pm5fl" Feb 27 08:09:35 crc kubenswrapper[4612]: I0227 08:09:35.270467 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bf325571-0b74-447f-bd04-213d54a8437e-ovsdbserver-sb\") pod \"dnsmasq-dns-6b7b667979-pm5fl\" (UID: \"bf325571-0b74-447f-bd04-213d54a8437e\") " pod="openstack/dnsmasq-dns-6b7b667979-pm5fl" Feb 27 08:09:35 crc kubenswrapper[4612]: I0227 08:09:35.270487 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bf325571-0b74-447f-bd04-213d54a8437e-dns-svc\") pod \"dnsmasq-dns-6b7b667979-pm5fl\" (UID: \"bf325571-0b74-447f-bd04-213d54a8437e\") " pod="openstack/dnsmasq-dns-6b7b667979-pm5fl" Feb 27 08:09:35 crc kubenswrapper[4612]: I0227 08:09:35.275515 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-77854b564-8nkbp"] Feb 27 08:09:35 crc kubenswrapper[4612]: I0227 08:09:35.283451 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bf325571-0b74-447f-bd04-213d54a8437e-ovsdbserver-nb\") pod \"dnsmasq-dns-6b7b667979-pm5fl\" (UID: \"bf325571-0b74-447f-bd04-213d54a8437e\") " pod="openstack/dnsmasq-dns-6b7b667979-pm5fl" Feb 27 08:09:35 crc kubenswrapper[4612]: I0227 08:09:35.284263 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bf325571-0b74-447f-bd04-213d54a8437e-config\") pod \"dnsmasq-dns-6b7b667979-pm5fl\" (UID: \"bf325571-0b74-447f-bd04-213d54a8437e\") " pod="openstack/dnsmasq-dns-6b7b667979-pm5fl" Feb 27 08:09:35 crc kubenswrapper[4612]: I0227 08:09:35.284810 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bf325571-0b74-447f-bd04-213d54a8437e-dns-svc\") pod \"dnsmasq-dns-6b7b667979-pm5fl\" (UID: \"bf325571-0b74-447f-bd04-213d54a8437e\") " pod="openstack/dnsmasq-dns-6b7b667979-pm5fl" Feb 27 08:09:35 crc kubenswrapper[4612]: I0227 08:09:35.285376 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/bf325571-0b74-447f-bd04-213d54a8437e-dns-swift-storage-0\") pod \"dnsmasq-dns-6b7b667979-pm5fl\" (UID: \"bf325571-0b74-447f-bd04-213d54a8437e\") " pod="openstack/dnsmasq-dns-6b7b667979-pm5fl" Feb 27 08:09:35 crc kubenswrapper[4612]: I0227 08:09:35.286408 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bf325571-0b74-447f-bd04-213d54a8437e-ovsdbserver-sb\") pod \"dnsmasq-dns-6b7b667979-pm5fl\" (UID: \"bf325571-0b74-447f-bd04-213d54a8437e\") " pod="openstack/dnsmasq-dns-6b7b667979-pm5fl" Feb 27 08:09:35 crc kubenswrapper[4612]: I0227 08:09:35.304752 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8zzrm\" (UniqueName: \"kubernetes.io/projected/bf325571-0b74-447f-bd04-213d54a8437e-kube-api-access-8zzrm\") pod \"dnsmasq-dns-6b7b667979-pm5fl\" (UID: \"bf325571-0b74-447f-bd04-213d54a8437e\") " pod="openstack/dnsmasq-dns-6b7b667979-pm5fl" Feb 27 08:09:35 crc kubenswrapper[4612]: I0227 08:09:35.372826 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/a9397555-a3e3-4a49-b725-882006d2ef0d-ovndb-tls-certs\") pod \"neutron-77854b564-8nkbp\" (UID: \"a9397555-a3e3-4a49-b725-882006d2ef0d\") " pod="openstack/neutron-77854b564-8nkbp" Feb 27 08:09:35 crc kubenswrapper[4612]: I0227 08:09:35.372875 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/a9397555-a3e3-4a49-b725-882006d2ef0d-config\") pod \"neutron-77854b564-8nkbp\" (UID: \"a9397555-a3e3-4a49-b725-882006d2ef0d\") " pod="openstack/neutron-77854b564-8nkbp" Feb 27 08:09:35 crc kubenswrapper[4612]: I0227 08:09:35.373017 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-njhrs\" (UniqueName: \"kubernetes.io/projected/a9397555-a3e3-4a49-b725-882006d2ef0d-kube-api-access-njhrs\") pod \"neutron-77854b564-8nkbp\" (UID: \"a9397555-a3e3-4a49-b725-882006d2ef0d\") " pod="openstack/neutron-77854b564-8nkbp" Feb 27 08:09:35 crc kubenswrapper[4612]: I0227 08:09:35.373052 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/a9397555-a3e3-4a49-b725-882006d2ef0d-httpd-config\") pod \"neutron-77854b564-8nkbp\" (UID: \"a9397555-a3e3-4a49-b725-882006d2ef0d\") " pod="openstack/neutron-77854b564-8nkbp" Feb 27 08:09:35 crc kubenswrapper[4612]: I0227 08:09:35.373075 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a9397555-a3e3-4a49-b725-882006d2ef0d-combined-ca-bundle\") pod \"neutron-77854b564-8nkbp\" (UID: \"a9397555-a3e3-4a49-b725-882006d2ef0d\") " pod="openstack/neutron-77854b564-8nkbp" Feb 27 08:09:35 crc kubenswrapper[4612]: I0227 08:09:35.428563 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b7b667979-pm5fl" Feb 27 08:09:35 crc kubenswrapper[4612]: I0227 08:09:35.474622 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-njhrs\" (UniqueName: \"kubernetes.io/projected/a9397555-a3e3-4a49-b725-882006d2ef0d-kube-api-access-njhrs\") pod \"neutron-77854b564-8nkbp\" (UID: \"a9397555-a3e3-4a49-b725-882006d2ef0d\") " pod="openstack/neutron-77854b564-8nkbp" Feb 27 08:09:35 crc kubenswrapper[4612]: I0227 08:09:35.474792 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/a9397555-a3e3-4a49-b725-882006d2ef0d-httpd-config\") pod \"neutron-77854b564-8nkbp\" (UID: \"a9397555-a3e3-4a49-b725-882006d2ef0d\") " pod="openstack/neutron-77854b564-8nkbp" Feb 27 08:09:35 crc kubenswrapper[4612]: I0227 08:09:35.474819 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a9397555-a3e3-4a49-b725-882006d2ef0d-combined-ca-bundle\") pod \"neutron-77854b564-8nkbp\" (UID: \"a9397555-a3e3-4a49-b725-882006d2ef0d\") " pod="openstack/neutron-77854b564-8nkbp" Feb 27 08:09:35 crc kubenswrapper[4612]: I0227 08:09:35.474872 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/a9397555-a3e3-4a49-b725-882006d2ef0d-ovndb-tls-certs\") pod \"neutron-77854b564-8nkbp\" (UID: \"a9397555-a3e3-4a49-b725-882006d2ef0d\") " pod="openstack/neutron-77854b564-8nkbp" Feb 27 08:09:35 crc kubenswrapper[4612]: I0227 08:09:35.474895 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/a9397555-a3e3-4a49-b725-882006d2ef0d-config\") pod \"neutron-77854b564-8nkbp\" (UID: \"a9397555-a3e3-4a49-b725-882006d2ef0d\") " pod="openstack/neutron-77854b564-8nkbp" Feb 27 08:09:35 crc kubenswrapper[4612]: I0227 08:09:35.478268 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/a9397555-a3e3-4a49-b725-882006d2ef0d-config\") pod \"neutron-77854b564-8nkbp\" (UID: \"a9397555-a3e3-4a49-b725-882006d2ef0d\") " pod="openstack/neutron-77854b564-8nkbp" Feb 27 08:09:35 crc kubenswrapper[4612]: I0227 08:09:35.481418 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a9397555-a3e3-4a49-b725-882006d2ef0d-combined-ca-bundle\") pod \"neutron-77854b564-8nkbp\" (UID: \"a9397555-a3e3-4a49-b725-882006d2ef0d\") " pod="openstack/neutron-77854b564-8nkbp" Feb 27 08:09:35 crc kubenswrapper[4612]: I0227 08:09:35.494955 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/a9397555-a3e3-4a49-b725-882006d2ef0d-httpd-config\") pod \"neutron-77854b564-8nkbp\" (UID: \"a9397555-a3e3-4a49-b725-882006d2ef0d\") " pod="openstack/neutron-77854b564-8nkbp" Feb 27 08:09:35 crc kubenswrapper[4612]: I0227 08:09:35.499261 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/a9397555-a3e3-4a49-b725-882006d2ef0d-ovndb-tls-certs\") pod \"neutron-77854b564-8nkbp\" (UID: \"a9397555-a3e3-4a49-b725-882006d2ef0d\") " pod="openstack/neutron-77854b564-8nkbp" Feb 27 08:09:35 crc kubenswrapper[4612]: I0227 08:09:35.500183 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-njhrs\" (UniqueName: \"kubernetes.io/projected/a9397555-a3e3-4a49-b725-882006d2ef0d-kube-api-access-njhrs\") pod \"neutron-77854b564-8nkbp\" (UID: \"a9397555-a3e3-4a49-b725-882006d2ef0d\") " pod="openstack/neutron-77854b564-8nkbp" Feb 27 08:09:35 crc kubenswrapper[4612]: I0227 08:09:35.576711 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-77854b564-8nkbp" Feb 27 08:09:37 crc kubenswrapper[4612]: I0227 08:09:37.037186 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-5ddbbb8689-bm5cp"] Feb 27 08:09:37 crc kubenswrapper[4612]: I0227 08:09:37.039281 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5ddbbb8689-bm5cp" Feb 27 08:09:37 crc kubenswrapper[4612]: I0227 08:09:37.041770 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Feb 27 08:09:37 crc kubenswrapper[4612]: I0227 08:09:37.045331 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Feb 27 08:09:37 crc kubenswrapper[4612]: I0227 08:09:37.056332 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-5ddbbb8689-bm5cp"] Feb 27 08:09:37 crc kubenswrapper[4612]: I0227 08:09:37.131935 4612 scope.go:117] "RemoveContainer" containerID="1036f4eb8af1f65a4cb1647bbdf944639cd97efc7f52a091258964de332daa16" Feb 27 08:09:37 crc kubenswrapper[4612]: I0227 08:09:37.199941 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/649e0df0-b37c-42f5-8af0-c804d4f467d4-config\") pod \"neutron-5ddbbb8689-bm5cp\" (UID: \"649e0df0-b37c-42f5-8af0-c804d4f467d4\") " pod="openstack/neutron-5ddbbb8689-bm5cp" Feb 27 08:09:37 crc kubenswrapper[4612]: I0227 08:09:37.200015 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/649e0df0-b37c-42f5-8af0-c804d4f467d4-httpd-config\") pod \"neutron-5ddbbb8689-bm5cp\" (UID: \"649e0df0-b37c-42f5-8af0-c804d4f467d4\") " pod="openstack/neutron-5ddbbb8689-bm5cp" Feb 27 08:09:37 crc kubenswrapper[4612]: I0227 08:09:37.200047 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kl8ws\" (UniqueName: \"kubernetes.io/projected/649e0df0-b37c-42f5-8af0-c804d4f467d4-kube-api-access-kl8ws\") pod \"neutron-5ddbbb8689-bm5cp\" (UID: \"649e0df0-b37c-42f5-8af0-c804d4f467d4\") " pod="openstack/neutron-5ddbbb8689-bm5cp" Feb 27 08:09:37 crc kubenswrapper[4612]: I0227 08:09:37.200077 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/649e0df0-b37c-42f5-8af0-c804d4f467d4-ovndb-tls-certs\") pod \"neutron-5ddbbb8689-bm5cp\" (UID: \"649e0df0-b37c-42f5-8af0-c804d4f467d4\") " pod="openstack/neutron-5ddbbb8689-bm5cp" Feb 27 08:09:37 crc kubenswrapper[4612]: I0227 08:09:37.200117 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/649e0df0-b37c-42f5-8af0-c804d4f467d4-public-tls-certs\") pod \"neutron-5ddbbb8689-bm5cp\" (UID: \"649e0df0-b37c-42f5-8af0-c804d4f467d4\") " pod="openstack/neutron-5ddbbb8689-bm5cp" Feb 27 08:09:37 crc kubenswrapper[4612]: I0227 08:09:37.200202 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/649e0df0-b37c-42f5-8af0-c804d4f467d4-combined-ca-bundle\") pod \"neutron-5ddbbb8689-bm5cp\" (UID: \"649e0df0-b37c-42f5-8af0-c804d4f467d4\") " pod="openstack/neutron-5ddbbb8689-bm5cp" Feb 27 08:09:37 crc kubenswrapper[4612]: I0227 08:09:37.200260 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/649e0df0-b37c-42f5-8af0-c804d4f467d4-internal-tls-certs\") pod \"neutron-5ddbbb8689-bm5cp\" (UID: \"649e0df0-b37c-42f5-8af0-c804d4f467d4\") " pod="openstack/neutron-5ddbbb8689-bm5cp" Feb 27 08:09:37 crc kubenswrapper[4612]: E0227 08:09:37.278164 4612 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified" Feb 27 08:09:37 crc kubenswrapper[4612]: E0227 08:09:37.278293 4612 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-fhfxv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-l5z5s_openstack(0c9482e8-9eb1-4c30-95e9-bdaa5ccd239b): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 27 08:09:37 crc kubenswrapper[4612]: E0227 08:09:37.279575 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-l5z5s" podUID="0c9482e8-9eb1-4c30-95e9-bdaa5ccd239b" Feb 27 08:09:37 crc kubenswrapper[4612]: I0227 08:09:37.281204 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-594c9f8cf-q282v" Feb 27 08:09:37 crc kubenswrapper[4612]: I0227 08:09:37.301296 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/649e0df0-b37c-42f5-8af0-c804d4f467d4-combined-ca-bundle\") pod \"neutron-5ddbbb8689-bm5cp\" (UID: \"649e0df0-b37c-42f5-8af0-c804d4f467d4\") " pod="openstack/neutron-5ddbbb8689-bm5cp" Feb 27 08:09:37 crc kubenswrapper[4612]: I0227 08:09:37.301353 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/649e0df0-b37c-42f5-8af0-c804d4f467d4-internal-tls-certs\") pod \"neutron-5ddbbb8689-bm5cp\" (UID: \"649e0df0-b37c-42f5-8af0-c804d4f467d4\") " pod="openstack/neutron-5ddbbb8689-bm5cp" Feb 27 08:09:37 crc kubenswrapper[4612]: I0227 08:09:37.301406 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/649e0df0-b37c-42f5-8af0-c804d4f467d4-config\") pod \"neutron-5ddbbb8689-bm5cp\" (UID: \"649e0df0-b37c-42f5-8af0-c804d4f467d4\") " pod="openstack/neutron-5ddbbb8689-bm5cp" Feb 27 08:09:37 crc kubenswrapper[4612]: I0227 08:09:37.301435 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/649e0df0-b37c-42f5-8af0-c804d4f467d4-httpd-config\") pod \"neutron-5ddbbb8689-bm5cp\" (UID: \"649e0df0-b37c-42f5-8af0-c804d4f467d4\") " pod="openstack/neutron-5ddbbb8689-bm5cp" Feb 27 08:09:37 crc kubenswrapper[4612]: I0227 08:09:37.301451 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kl8ws\" (UniqueName: \"kubernetes.io/projected/649e0df0-b37c-42f5-8af0-c804d4f467d4-kube-api-access-kl8ws\") pod \"neutron-5ddbbb8689-bm5cp\" (UID: \"649e0df0-b37c-42f5-8af0-c804d4f467d4\") " pod="openstack/neutron-5ddbbb8689-bm5cp" Feb 27 08:09:37 crc kubenswrapper[4612]: I0227 08:09:37.301468 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/649e0df0-b37c-42f5-8af0-c804d4f467d4-ovndb-tls-certs\") pod \"neutron-5ddbbb8689-bm5cp\" (UID: \"649e0df0-b37c-42f5-8af0-c804d4f467d4\") " pod="openstack/neutron-5ddbbb8689-bm5cp" Feb 27 08:09:37 crc kubenswrapper[4612]: I0227 08:09:37.301495 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/649e0df0-b37c-42f5-8af0-c804d4f467d4-public-tls-certs\") pod \"neutron-5ddbbb8689-bm5cp\" (UID: \"649e0df0-b37c-42f5-8af0-c804d4f467d4\") " pod="openstack/neutron-5ddbbb8689-bm5cp" Feb 27 08:09:37 crc kubenswrapper[4612]: I0227 08:09:37.308373 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/649e0df0-b37c-42f5-8af0-c804d4f467d4-ovndb-tls-certs\") pod \"neutron-5ddbbb8689-bm5cp\" (UID: \"649e0df0-b37c-42f5-8af0-c804d4f467d4\") " pod="openstack/neutron-5ddbbb8689-bm5cp" Feb 27 08:09:37 crc kubenswrapper[4612]: I0227 08:09:37.316307 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/649e0df0-b37c-42f5-8af0-c804d4f467d4-public-tls-certs\") pod \"neutron-5ddbbb8689-bm5cp\" (UID: \"649e0df0-b37c-42f5-8af0-c804d4f467d4\") " pod="openstack/neutron-5ddbbb8689-bm5cp" Feb 27 08:09:37 crc kubenswrapper[4612]: I0227 08:09:37.316493 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/649e0df0-b37c-42f5-8af0-c804d4f467d4-combined-ca-bundle\") pod \"neutron-5ddbbb8689-bm5cp\" (UID: \"649e0df0-b37c-42f5-8af0-c804d4f467d4\") " pod="openstack/neutron-5ddbbb8689-bm5cp" Feb 27 08:09:37 crc kubenswrapper[4612]: I0227 08:09:37.317141 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/649e0df0-b37c-42f5-8af0-c804d4f467d4-httpd-config\") pod \"neutron-5ddbbb8689-bm5cp\" (UID: \"649e0df0-b37c-42f5-8af0-c804d4f467d4\") " pod="openstack/neutron-5ddbbb8689-bm5cp" Feb 27 08:09:37 crc kubenswrapper[4612]: I0227 08:09:37.323651 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/649e0df0-b37c-42f5-8af0-c804d4f467d4-internal-tls-certs\") pod \"neutron-5ddbbb8689-bm5cp\" (UID: \"649e0df0-b37c-42f5-8af0-c804d4f467d4\") " pod="openstack/neutron-5ddbbb8689-bm5cp" Feb 27 08:09:37 crc kubenswrapper[4612]: I0227 08:09:37.323867 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7f9b6cd66c-5lwzw" Feb 27 08:09:37 crc kubenswrapper[4612]: I0227 08:09:37.375521 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kl8ws\" (UniqueName: \"kubernetes.io/projected/649e0df0-b37c-42f5-8af0-c804d4f467d4-kube-api-access-kl8ws\") pod \"neutron-5ddbbb8689-bm5cp\" (UID: \"649e0df0-b37c-42f5-8af0-c804d4f467d4\") " pod="openstack/neutron-5ddbbb8689-bm5cp" Feb 27 08:09:37 crc kubenswrapper[4612]: I0227 08:09:37.406584 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f37f5ce9-74c3-48fc-9080-ae81464dba49-logs\") pod \"f37f5ce9-74c3-48fc-9080-ae81464dba49\" (UID: \"f37f5ce9-74c3-48fc-9080-ae81464dba49\") " Feb 27 08:09:37 crc kubenswrapper[4612]: I0227 08:09:37.406663 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/f37f5ce9-74c3-48fc-9080-ae81464dba49-horizon-secret-key\") pod \"f37f5ce9-74c3-48fc-9080-ae81464dba49\" (UID: \"f37f5ce9-74c3-48fc-9080-ae81464dba49\") " Feb 27 08:09:37 crc kubenswrapper[4612]: I0227 08:09:37.406709 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a10f6a85-63bd-4aa7-91a6-b3aec6ecb9d4-scripts\") pod \"a10f6a85-63bd-4aa7-91a6-b3aec6ecb9d4\" (UID: \"a10f6a85-63bd-4aa7-91a6-b3aec6ecb9d4\") " Feb 27 08:09:37 crc kubenswrapper[4612]: I0227 08:09:37.410840 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a10f6a85-63bd-4aa7-91a6-b3aec6ecb9d4-logs\") pod \"a10f6a85-63bd-4aa7-91a6-b3aec6ecb9d4\" (UID: \"a10f6a85-63bd-4aa7-91a6-b3aec6ecb9d4\") " Feb 27 08:09:37 crc kubenswrapper[4612]: I0227 08:09:37.410897 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a10f6a85-63bd-4aa7-91a6-b3aec6ecb9d4-config-data\") pod \"a10f6a85-63bd-4aa7-91a6-b3aec6ecb9d4\" (UID: \"a10f6a85-63bd-4aa7-91a6-b3aec6ecb9d4\") " Feb 27 08:09:37 crc kubenswrapper[4612]: I0227 08:09:37.410943 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f37f5ce9-74c3-48fc-9080-ae81464dba49-scripts\") pod \"f37f5ce9-74c3-48fc-9080-ae81464dba49\" (UID: \"f37f5ce9-74c3-48fc-9080-ae81464dba49\") " Feb 27 08:09:37 crc kubenswrapper[4612]: I0227 08:09:37.410994 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f37f5ce9-74c3-48fc-9080-ae81464dba49-config-data\") pod \"f37f5ce9-74c3-48fc-9080-ae81464dba49\" (UID: \"f37f5ce9-74c3-48fc-9080-ae81464dba49\") " Feb 27 08:09:37 crc kubenswrapper[4612]: I0227 08:09:37.411027 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9zthz\" (UniqueName: \"kubernetes.io/projected/a10f6a85-63bd-4aa7-91a6-b3aec6ecb9d4-kube-api-access-9zthz\") pod \"a10f6a85-63bd-4aa7-91a6-b3aec6ecb9d4\" (UID: \"a10f6a85-63bd-4aa7-91a6-b3aec6ecb9d4\") " Feb 27 08:09:37 crc kubenswrapper[4612]: I0227 08:09:37.411115 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zpvms\" (UniqueName: \"kubernetes.io/projected/f37f5ce9-74c3-48fc-9080-ae81464dba49-kube-api-access-zpvms\") pod \"f37f5ce9-74c3-48fc-9080-ae81464dba49\" (UID: \"f37f5ce9-74c3-48fc-9080-ae81464dba49\") " Feb 27 08:09:37 crc kubenswrapper[4612]: I0227 08:09:37.411175 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/a10f6a85-63bd-4aa7-91a6-b3aec6ecb9d4-horizon-secret-key\") pod \"a10f6a85-63bd-4aa7-91a6-b3aec6ecb9d4\" (UID: \"a10f6a85-63bd-4aa7-91a6-b3aec6ecb9d4\") " Feb 27 08:09:37 crc kubenswrapper[4612]: I0227 08:09:37.412679 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a10f6a85-63bd-4aa7-91a6-b3aec6ecb9d4-config-data" (OuterVolumeSpecName: "config-data") pod "a10f6a85-63bd-4aa7-91a6-b3aec6ecb9d4" (UID: "a10f6a85-63bd-4aa7-91a6-b3aec6ecb9d4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 08:09:37 crc kubenswrapper[4612]: I0227 08:09:37.413360 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a10f6a85-63bd-4aa7-91a6-b3aec6ecb9d4-logs" (OuterVolumeSpecName: "logs") pod "a10f6a85-63bd-4aa7-91a6-b3aec6ecb9d4" (UID: "a10f6a85-63bd-4aa7-91a6-b3aec6ecb9d4"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 08:09:37 crc kubenswrapper[4612]: I0227 08:09:37.415305 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f37f5ce9-74c3-48fc-9080-ae81464dba49-scripts" (OuterVolumeSpecName: "scripts") pod "f37f5ce9-74c3-48fc-9080-ae81464dba49" (UID: "f37f5ce9-74c3-48fc-9080-ae81464dba49"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 08:09:37 crc kubenswrapper[4612]: I0227 08:09:37.466053 4612 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a10f6a85-63bd-4aa7-91a6-b3aec6ecb9d4-logs\") on node \"crc\" DevicePath \"\"" Feb 27 08:09:37 crc kubenswrapper[4612]: I0227 08:09:37.466264 4612 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a10f6a85-63bd-4aa7-91a6-b3aec6ecb9d4-config-data\") on node \"crc\" DevicePath \"\"" Feb 27 08:09:37 crc kubenswrapper[4612]: I0227 08:09:37.466278 4612 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f37f5ce9-74c3-48fc-9080-ae81464dba49-scripts\") on node \"crc\" DevicePath \"\"" Feb 27 08:09:37 crc kubenswrapper[4612]: I0227 08:09:37.468968 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f37f5ce9-74c3-48fc-9080-ae81464dba49-config-data" (OuterVolumeSpecName: "config-data") pod "f37f5ce9-74c3-48fc-9080-ae81464dba49" (UID: "f37f5ce9-74c3-48fc-9080-ae81464dba49"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 08:09:37 crc kubenswrapper[4612]: I0227 08:09:37.469993 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a10f6a85-63bd-4aa7-91a6-b3aec6ecb9d4-scripts" (OuterVolumeSpecName: "scripts") pod "a10f6a85-63bd-4aa7-91a6-b3aec6ecb9d4" (UID: "a10f6a85-63bd-4aa7-91a6-b3aec6ecb9d4"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 08:09:37 crc kubenswrapper[4612]: I0227 08:09:37.470037 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f37f5ce9-74c3-48fc-9080-ae81464dba49-logs" (OuterVolumeSpecName: "logs") pod "f37f5ce9-74c3-48fc-9080-ae81464dba49" (UID: "f37f5ce9-74c3-48fc-9080-ae81464dba49"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 08:09:37 crc kubenswrapper[4612]: I0227 08:09:37.474686 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f37f5ce9-74c3-48fc-9080-ae81464dba49-kube-api-access-zpvms" (OuterVolumeSpecName: "kube-api-access-zpvms") pod "f37f5ce9-74c3-48fc-9080-ae81464dba49" (UID: "f37f5ce9-74c3-48fc-9080-ae81464dba49"). InnerVolumeSpecName "kube-api-access-zpvms". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 08:09:37 crc kubenswrapper[4612]: I0227 08:09:37.475548 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/649e0df0-b37c-42f5-8af0-c804d4f467d4-config\") pod \"neutron-5ddbbb8689-bm5cp\" (UID: \"649e0df0-b37c-42f5-8af0-c804d4f467d4\") " pod="openstack/neutron-5ddbbb8689-bm5cp" Feb 27 08:09:37 crc kubenswrapper[4612]: I0227 08:09:37.478535 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f37f5ce9-74c3-48fc-9080-ae81464dba49-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "f37f5ce9-74c3-48fc-9080-ae81464dba49" (UID: "f37f5ce9-74c3-48fc-9080-ae81464dba49"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:09:37 crc kubenswrapper[4612]: I0227 08:09:37.486137 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a10f6a85-63bd-4aa7-91a6-b3aec6ecb9d4-kube-api-access-9zthz" (OuterVolumeSpecName: "kube-api-access-9zthz") pod "a10f6a85-63bd-4aa7-91a6-b3aec6ecb9d4" (UID: "a10f6a85-63bd-4aa7-91a6-b3aec6ecb9d4"). InnerVolumeSpecName "kube-api-access-9zthz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 08:09:37 crc kubenswrapper[4612]: I0227 08:09:37.490561 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a10f6a85-63bd-4aa7-91a6-b3aec6ecb9d4-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "a10f6a85-63bd-4aa7-91a6-b3aec6ecb9d4" (UID: "a10f6a85-63bd-4aa7-91a6-b3aec6ecb9d4"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:09:37 crc kubenswrapper[4612]: I0227 08:09:37.571635 4612 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a10f6a85-63bd-4aa7-91a6-b3aec6ecb9d4-scripts\") on node \"crc\" DevicePath \"\"" Feb 27 08:09:37 crc kubenswrapper[4612]: I0227 08:09:37.571671 4612 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/f37f5ce9-74c3-48fc-9080-ae81464dba49-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Feb 27 08:09:37 crc kubenswrapper[4612]: I0227 08:09:37.571685 4612 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f37f5ce9-74c3-48fc-9080-ae81464dba49-config-data\") on node \"crc\" DevicePath \"\"" Feb 27 08:09:37 crc kubenswrapper[4612]: I0227 08:09:37.571723 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9zthz\" (UniqueName: \"kubernetes.io/projected/a10f6a85-63bd-4aa7-91a6-b3aec6ecb9d4-kube-api-access-9zthz\") on node \"crc\" DevicePath \"\"" Feb 27 08:09:37 crc kubenswrapper[4612]: I0227 08:09:37.571736 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zpvms\" (UniqueName: \"kubernetes.io/projected/f37f5ce9-74c3-48fc-9080-ae81464dba49-kube-api-access-zpvms\") on node \"crc\" DevicePath \"\"" Feb 27 08:09:37 crc kubenswrapper[4612]: I0227 08:09:37.571749 4612 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/a10f6a85-63bd-4aa7-91a6-b3aec6ecb9d4-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Feb 27 08:09:37 crc kubenswrapper[4612]: I0227 08:09:37.571760 4612 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f37f5ce9-74c3-48fc-9080-ae81464dba49-logs\") on node \"crc\" DevicePath \"\"" Feb 27 08:09:37 crc kubenswrapper[4612]: I0227 08:09:37.581033 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5ddbbb8689-bm5cp" Feb 27 08:09:37 crc kubenswrapper[4612]: I0227 08:09:37.729943 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-594c9f8cf-q282v" event={"ID":"f37f5ce9-74c3-48fc-9080-ae81464dba49","Type":"ContainerDied","Data":"96a4e58df22136ff0feebad1d8d3ec302578a66e6437cd0adfd7785d7f586ee6"} Feb 27 08:09:37 crc kubenswrapper[4612]: I0227 08:09:37.730212 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-594c9f8cf-q282v" Feb 27 08:09:37 crc kubenswrapper[4612]: I0227 08:09:37.746457 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7f9b6cd66c-5lwzw" event={"ID":"a10f6a85-63bd-4aa7-91a6-b3aec6ecb9d4","Type":"ContainerDied","Data":"edb7c7f2a1e1708daa090ff5bc33972634b6a95993156afb29b1a25cf6beaa21"} Feb 27 08:09:37 crc kubenswrapper[4612]: I0227 08:09:37.746553 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7f9b6cd66c-5lwzw" Feb 27 08:09:37 crc kubenswrapper[4612]: E0227 08:09:37.758926 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified\\\"\"" pod="openstack/cinder-db-sync-l5z5s" podUID="0c9482e8-9eb1-4c30-95e9-bdaa5ccd239b" Feb 27 08:09:37 crc kubenswrapper[4612]: I0227 08:09:37.841541 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-594c9f8cf-q282v"] Feb 27 08:09:37 crc kubenswrapper[4612]: I0227 08:09:37.857118 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-594c9f8cf-q282v"] Feb 27 08:09:37 crc kubenswrapper[4612]: I0227 08:09:37.930588 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-7f9b6cd66c-5lwzw"] Feb 27 08:09:37 crc kubenswrapper[4612]: I0227 08:09:37.946289 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-7f9b6cd66c-5lwzw"] Feb 27 08:09:37 crc kubenswrapper[4612]: I0227 08:09:37.956020 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-77948db5bb-97qh6"] Feb 27 08:09:38 crc kubenswrapper[4612]: I0227 08:09:38.002803 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-s5bjv"] Feb 27 08:09:38 crc kubenswrapper[4612]: W0227 08:09:38.023911 4612 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1a05a7e5_8c46_441a_9faf_0fd3b41bbd98.slice/crio-df93dc184990b247c92ef9f46224a41ef8a99e51093c8ececf390e03ff9d0660 WatchSource:0}: Error finding container df93dc184990b247c92ef9f46224a41ef8a99e51093c8ececf390e03ff9d0660: Status 404 returned error can't find the container with id df93dc184990b247c92ef9f46224a41ef8a99e51093c8ececf390e03ff9d0660 Feb 27 08:09:38 crc kubenswrapper[4612]: I0227 08:09:38.188579 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6b7b667979-pm5fl"] Feb 27 08:09:38 crc kubenswrapper[4612]: I0227 08:09:38.198181 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-6b985547d4-mkfpv"] Feb 27 08:09:38 crc kubenswrapper[4612]: W0227 08:09:38.214251 4612 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbf325571_0b74_447f_bd04_213d54a8437e.slice/crio-26d204ca3654b872870f1ae7b2046125ba43c32433373e33bbf690ab1087c66c WatchSource:0}: Error finding container 26d204ca3654b872870f1ae7b2046125ba43c32433373e33bbf690ab1087c66c: Status 404 returned error can't find the container with id 26d204ca3654b872870f1ae7b2046125ba43c32433373e33bbf690ab1087c66c Feb 27 08:09:38 crc kubenswrapper[4612]: I0227 08:09:38.431832 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-77854b564-8nkbp"] Feb 27 08:09:38 crc kubenswrapper[4612]: W0227 08:09:38.434062 4612 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda9397555_a3e3_4a49_b725_882006d2ef0d.slice/crio-0b2aa2610321a1eab88845caf61ec74398c53cf0faad98a6a45c83957ad8e63e WatchSource:0}: Error finding container 0b2aa2610321a1eab88845caf61ec74398c53cf0faad98a6a45c83957ad8e63e: Status 404 returned error can't find the container with id 0b2aa2610321a1eab88845caf61ec74398c53cf0faad98a6a45c83957ad8e63e Feb 27 08:09:38 crc kubenswrapper[4612]: I0227 08:09:38.766073 4612 generic.go:334] "Generic (PLEG): container finished" podID="bf325571-0b74-447f-bd04-213d54a8437e" containerID="273c781089e6725f6eb0a109249acabcde54e129dd2aef517f776687ba0f1701" exitCode=0 Feb 27 08:09:38 crc kubenswrapper[4612]: I0227 08:09:38.766374 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b7b667979-pm5fl" event={"ID":"bf325571-0b74-447f-bd04-213d54a8437e","Type":"ContainerDied","Data":"273c781089e6725f6eb0a109249acabcde54e129dd2aef517f776687ba0f1701"} Feb 27 08:09:38 crc kubenswrapper[4612]: I0227 08:09:38.766400 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b7b667979-pm5fl" event={"ID":"bf325571-0b74-447f-bd04-213d54a8437e","Type":"ContainerStarted","Data":"26d204ca3654b872870f1ae7b2046125ba43c32433373e33bbf690ab1087c66c"} Feb 27 08:09:38 crc kubenswrapper[4612]: I0227 08:09:38.770497 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"d62745a5-cfe3-48b0-a167-76735bdfd8fe","Type":"ContainerStarted","Data":"01f1519d415fe7fb5962939cdb8fc37629da75d9f2d2645dce4cad73db91f1bb"} Feb 27 08:09:38 crc kubenswrapper[4612]: I0227 08:09:38.770669 4612 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="d62745a5-cfe3-48b0-a167-76735bdfd8fe" containerName="glance-log" containerID="cri-o://d7beccc144fa9bf9b73e2f015fc0afe88018fd5883a580e0623b4c591da30e02" gracePeriod=30 Feb 27 08:09:38 crc kubenswrapper[4612]: I0227 08:09:38.770773 4612 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="d62745a5-cfe3-48b0-a167-76735bdfd8fe" containerName="glance-httpd" containerID="cri-o://01f1519d415fe7fb5962939cdb8fc37629da75d9f2d2645dce4cad73db91f1bb" gracePeriod=30 Feb 27 08:09:38 crc kubenswrapper[4612]: I0227 08:09:38.773559 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6b985547d4-mkfpv" event={"ID":"76009243-16b3-4652-abfe-a6cdba363724","Type":"ContainerStarted","Data":"6222dbd09c56ca68cdf2e41ea355c459ebd6a049c501036b2337e5fe88da1409"} Feb 27 08:09:38 crc kubenswrapper[4612]: I0227 08:09:38.775398 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-s5bjv" event={"ID":"1a05a7e5-8c46-441a-9faf-0fd3b41bbd98","Type":"ContainerStarted","Data":"d7fbb0369e2f86e02466cb49c5f6a3bba32f32d2e771e988cd25937fb4d8dd80"} Feb 27 08:09:38 crc kubenswrapper[4612]: I0227 08:09:38.775428 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-s5bjv" event={"ID":"1a05a7e5-8c46-441a-9faf-0fd3b41bbd98","Type":"ContainerStarted","Data":"df93dc184990b247c92ef9f46224a41ef8a99e51093c8ececf390e03ff9d0660"} Feb 27 08:09:38 crc kubenswrapper[4612]: I0227 08:09:38.779225 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"4726ffcd-f407-4cd4-a6e4-c422801f2617","Type":"ContainerStarted","Data":"a661b93501b59c43ff694deb74a3b3bb3649f7a43b08aa70241c6752498b2d38"} Feb 27 08:09:38 crc kubenswrapper[4612]: I0227 08:09:38.780528 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-77948db5bb-97qh6" event={"ID":"b72b541d-4661-44a7-a121-c2a8aef6db11","Type":"ContainerStarted","Data":"33d7a7cef6907aa26e035be71f0c2a511861dcde3791484be5c89774f538fe0d"} Feb 27 08:09:38 crc kubenswrapper[4612]: I0227 08:09:38.783215 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-8zrcl" event={"ID":"36b549a3-c6e9-4f7e-8cff-85216c8e0e3d","Type":"ContainerStarted","Data":"a6d45103e694a574568b137e43fc7bcdcd0e18154a2c6b7ba6c90c509ea0da15"} Feb 27 08:09:38 crc kubenswrapper[4612]: I0227 08:09:38.802660 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-77854b564-8nkbp" event={"ID":"a9397555-a3e3-4a49-b725-882006d2ef0d","Type":"ContainerStarted","Data":"0b2aa2610321a1eab88845caf61ec74398c53cf0faad98a6a45c83957ad8e63e"} Feb 27 08:09:38 crc kubenswrapper[4612]: I0227 08:09:38.856425 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=32.856407291 podStartE2EDuration="32.856407291s" podCreationTimestamp="2026-02-27 08:09:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 08:09:38.820118091 +0000 UTC m=+1236.674048089" watchObservedRunningTime="2026-02-27 08:09:38.856407291 +0000 UTC m=+1236.710337289" Feb 27 08:09:38 crc kubenswrapper[4612]: I0227 08:09:38.857882 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-s5bjv" podStartSLOduration=23.857875033 podStartE2EDuration="23.857875033s" podCreationTimestamp="2026-02-27 08:09:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 08:09:38.848922526 +0000 UTC m=+1236.702852524" watchObservedRunningTime="2026-02-27 08:09:38.857875033 +0000 UTC m=+1236.711805031" Feb 27 08:09:38 crc kubenswrapper[4612]: I0227 08:09:38.878917 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a10f6a85-63bd-4aa7-91a6-b3aec6ecb9d4" path="/var/lib/kubelet/pods/a10f6a85-63bd-4aa7-91a6-b3aec6ecb9d4/volumes" Feb 27 08:09:38 crc kubenswrapper[4612]: I0227 08:09:38.879456 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f37f5ce9-74c3-48fc-9080-ae81464dba49" path="/var/lib/kubelet/pods/f37f5ce9-74c3-48fc-9080-ae81464dba49/volumes" Feb 27 08:09:38 crc kubenswrapper[4612]: I0227 08:09:38.900069 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-8zrcl" podStartSLOduration=4.438864878 podStartE2EDuration="39.900046681s" podCreationTimestamp="2026-02-27 08:08:59 +0000 UTC" firstStartedPulling="2026-02-27 08:09:01.595329893 +0000 UTC m=+1199.449259891" lastFinishedPulling="2026-02-27 08:09:37.056511696 +0000 UTC m=+1234.910441694" observedRunningTime="2026-02-27 08:09:38.870957578 +0000 UTC m=+1236.724887576" watchObservedRunningTime="2026-02-27 08:09:38.900046681 +0000 UTC m=+1236.753976679" Feb 27 08:09:39 crc kubenswrapper[4612]: I0227 08:09:39.535103 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-5ddbbb8689-bm5cp"] Feb 27 08:09:39 crc kubenswrapper[4612]: I0227 08:09:39.776615 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 27 08:09:39 crc kubenswrapper[4612]: I0227 08:09:39.841224 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-77948db5bb-97qh6" event={"ID":"b72b541d-4661-44a7-a121-c2a8aef6db11","Type":"ContainerStarted","Data":"74ca1600814f168714e4e6229ada6995b0eed2c976241b09d104b489d4efb1dc"} Feb 27 08:09:39 crc kubenswrapper[4612]: I0227 08:09:39.847800 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-77854b564-8nkbp" event={"ID":"a9397555-a3e3-4a49-b725-882006d2ef0d","Type":"ContainerStarted","Data":"a282f0ed98a69a8ea94fc9525a85432465089708bc9f78755024b91a5fd52a03"} Feb 27 08:09:39 crc kubenswrapper[4612]: I0227 08:09:39.872467 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5ddbbb8689-bm5cp" event={"ID":"649e0df0-b37c-42f5-8af0-c804d4f467d4","Type":"ContainerStarted","Data":"0974569712b15ea4dcc9022105f7deb4a6621d88d67b162ab1e1ec41c914fe97"} Feb 27 08:09:39 crc kubenswrapper[4612]: I0227 08:09:39.879188 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-r8dpp" event={"ID":"87df2a96-7ae3-47af-92bb-97db038b957c","Type":"ContainerStarted","Data":"7a14e07a854b7c34d66faacc2713fc09fcde83a58ea6b802875545f39c503506"} Feb 27 08:09:39 crc kubenswrapper[4612]: I0227 08:09:39.896378 4612 generic.go:334] "Generic (PLEG): container finished" podID="d62745a5-cfe3-48b0-a167-76735bdfd8fe" containerID="01f1519d415fe7fb5962939cdb8fc37629da75d9f2d2645dce4cad73db91f1bb" exitCode=143 Feb 27 08:09:39 crc kubenswrapper[4612]: I0227 08:09:39.896409 4612 generic.go:334] "Generic (PLEG): container finished" podID="d62745a5-cfe3-48b0-a167-76735bdfd8fe" containerID="d7beccc144fa9bf9b73e2f015fc0afe88018fd5883a580e0623b4c591da30e02" exitCode=143 Feb 27 08:09:39 crc kubenswrapper[4612]: I0227 08:09:39.897372 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 27 08:09:39 crc kubenswrapper[4612]: I0227 08:09:39.900745 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-r8dpp" podStartSLOduration=3.717096906 podStartE2EDuration="40.900726914s" podCreationTimestamp="2026-02-27 08:08:59 +0000 UTC" firstStartedPulling="2026-02-27 08:09:01.338179435 +0000 UTC m=+1199.192109433" lastFinishedPulling="2026-02-27 08:09:38.521809443 +0000 UTC m=+1236.375739441" observedRunningTime="2026-02-27 08:09:39.895102563 +0000 UTC m=+1237.749032561" watchObservedRunningTime="2026-02-27 08:09:39.900726914 +0000 UTC m=+1237.754656912" Feb 27 08:09:39 crc kubenswrapper[4612]: I0227 08:09:39.901135 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"d62745a5-cfe3-48b0-a167-76735bdfd8fe","Type":"ContainerDied","Data":"01f1519d415fe7fb5962939cdb8fc37629da75d9f2d2645dce4cad73db91f1bb"} Feb 27 08:09:39 crc kubenswrapper[4612]: I0227 08:09:39.901171 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"d62745a5-cfe3-48b0-a167-76735bdfd8fe","Type":"ContainerDied","Data":"d7beccc144fa9bf9b73e2f015fc0afe88018fd5883a580e0623b4c591da30e02"} Feb 27 08:09:39 crc kubenswrapper[4612]: I0227 08:09:39.901229 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"d62745a5-cfe3-48b0-a167-76735bdfd8fe","Type":"ContainerDied","Data":"3d531517a4dce736d9caa8e8b6967b62d1fa7937af3dabd7ecc97ce41b5b7653"} Feb 27 08:09:39 crc kubenswrapper[4612]: I0227 08:09:39.901246 4612 scope.go:117] "RemoveContainer" containerID="01f1519d415fe7fb5962939cdb8fc37629da75d9f2d2645dce4cad73db91f1bb" Feb 27 08:09:39 crc kubenswrapper[4612]: I0227 08:09:39.928366 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d62745a5-cfe3-48b0-a167-76735bdfd8fe-combined-ca-bundle\") pod \"d62745a5-cfe3-48b0-a167-76735bdfd8fe\" (UID: \"d62745a5-cfe3-48b0-a167-76735bdfd8fe\") " Feb 27 08:09:39 crc kubenswrapper[4612]: I0227 08:09:39.928565 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d62745a5-cfe3-48b0-a167-76735bdfd8fe-scripts\") pod \"d62745a5-cfe3-48b0-a167-76735bdfd8fe\" (UID: \"d62745a5-cfe3-48b0-a167-76735bdfd8fe\") " Feb 27 08:09:39 crc kubenswrapper[4612]: I0227 08:09:39.928600 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d62745a5-cfe3-48b0-a167-76735bdfd8fe-logs\") pod \"d62745a5-cfe3-48b0-a167-76735bdfd8fe\" (UID: \"d62745a5-cfe3-48b0-a167-76735bdfd8fe\") " Feb 27 08:09:39 crc kubenswrapper[4612]: I0227 08:09:39.931527 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d62745a5-cfe3-48b0-a167-76735bdfd8fe-logs" (OuterVolumeSpecName: "logs") pod "d62745a5-cfe3-48b0-a167-76735bdfd8fe" (UID: "d62745a5-cfe3-48b0-a167-76735bdfd8fe"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 08:09:39 crc kubenswrapper[4612]: I0227 08:09:39.932042 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2r2jz\" (UniqueName: \"kubernetes.io/projected/d62745a5-cfe3-48b0-a167-76735bdfd8fe-kube-api-access-2r2jz\") pod \"d62745a5-cfe3-48b0-a167-76735bdfd8fe\" (UID: \"d62745a5-cfe3-48b0-a167-76735bdfd8fe\") " Feb 27 08:09:39 crc kubenswrapper[4612]: I0227 08:09:39.932072 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d62745a5-cfe3-48b0-a167-76735bdfd8fe-httpd-run\") pod \"d62745a5-cfe3-48b0-a167-76735bdfd8fe\" (UID: \"d62745a5-cfe3-48b0-a167-76735bdfd8fe\") " Feb 27 08:09:39 crc kubenswrapper[4612]: I0227 08:09:39.932111 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"d62745a5-cfe3-48b0-a167-76735bdfd8fe\" (UID: \"d62745a5-cfe3-48b0-a167-76735bdfd8fe\") " Feb 27 08:09:39 crc kubenswrapper[4612]: I0227 08:09:39.932154 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d62745a5-cfe3-48b0-a167-76735bdfd8fe-public-tls-certs\") pod \"d62745a5-cfe3-48b0-a167-76735bdfd8fe\" (UID: \"d62745a5-cfe3-48b0-a167-76735bdfd8fe\") " Feb 27 08:09:39 crc kubenswrapper[4612]: I0227 08:09:39.932196 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d62745a5-cfe3-48b0-a167-76735bdfd8fe-config-data\") pod \"d62745a5-cfe3-48b0-a167-76735bdfd8fe\" (UID: \"d62745a5-cfe3-48b0-a167-76735bdfd8fe\") " Feb 27 08:09:39 crc kubenswrapper[4612]: I0227 08:09:39.932549 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d62745a5-cfe3-48b0-a167-76735bdfd8fe-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "d62745a5-cfe3-48b0-a167-76735bdfd8fe" (UID: "d62745a5-cfe3-48b0-a167-76735bdfd8fe"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 08:09:39 crc kubenswrapper[4612]: I0227 08:09:39.934664 4612 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d62745a5-cfe3-48b0-a167-76735bdfd8fe-logs\") on node \"crc\" DevicePath \"\"" Feb 27 08:09:39 crc kubenswrapper[4612]: I0227 08:09:39.934750 4612 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d62745a5-cfe3-48b0-a167-76735bdfd8fe-httpd-run\") on node \"crc\" DevicePath \"\"" Feb 27 08:09:39 crc kubenswrapper[4612]: I0227 08:09:39.934891 4612 scope.go:117] "RemoveContainer" containerID="d7beccc144fa9bf9b73e2f015fc0afe88018fd5883a580e0623b4c591da30e02" Feb 27 08:09:39 crc kubenswrapper[4612]: I0227 08:09:39.935150 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d62745a5-cfe3-48b0-a167-76735bdfd8fe-scripts" (OuterVolumeSpecName: "scripts") pod "d62745a5-cfe3-48b0-a167-76735bdfd8fe" (UID: "d62745a5-cfe3-48b0-a167-76735bdfd8fe"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:09:39 crc kubenswrapper[4612]: I0227 08:09:39.937385 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d62745a5-cfe3-48b0-a167-76735bdfd8fe-kube-api-access-2r2jz" (OuterVolumeSpecName: "kube-api-access-2r2jz") pod "d62745a5-cfe3-48b0-a167-76735bdfd8fe" (UID: "d62745a5-cfe3-48b0-a167-76735bdfd8fe"). InnerVolumeSpecName "kube-api-access-2r2jz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 08:09:39 crc kubenswrapper[4612]: I0227 08:09:39.947177 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage03-crc" (OuterVolumeSpecName: "glance") pod "d62745a5-cfe3-48b0-a167-76735bdfd8fe" (UID: "d62745a5-cfe3-48b0-a167-76735bdfd8fe"). InnerVolumeSpecName "local-storage03-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Feb 27 08:09:39 crc kubenswrapper[4612]: I0227 08:09:39.966313 4612 scope.go:117] "RemoveContainer" containerID="01f1519d415fe7fb5962939cdb8fc37629da75d9f2d2645dce4cad73db91f1bb" Feb 27 08:09:39 crc kubenswrapper[4612]: E0227 08:09:39.967376 4612 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"01f1519d415fe7fb5962939cdb8fc37629da75d9f2d2645dce4cad73db91f1bb\": container with ID starting with 01f1519d415fe7fb5962939cdb8fc37629da75d9f2d2645dce4cad73db91f1bb not found: ID does not exist" containerID="01f1519d415fe7fb5962939cdb8fc37629da75d9f2d2645dce4cad73db91f1bb" Feb 27 08:09:39 crc kubenswrapper[4612]: I0227 08:09:39.967617 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"01f1519d415fe7fb5962939cdb8fc37629da75d9f2d2645dce4cad73db91f1bb"} err="failed to get container status \"01f1519d415fe7fb5962939cdb8fc37629da75d9f2d2645dce4cad73db91f1bb\": rpc error: code = NotFound desc = could not find container \"01f1519d415fe7fb5962939cdb8fc37629da75d9f2d2645dce4cad73db91f1bb\": container with ID starting with 01f1519d415fe7fb5962939cdb8fc37629da75d9f2d2645dce4cad73db91f1bb not found: ID does not exist" Feb 27 08:09:39 crc kubenswrapper[4612]: I0227 08:09:39.967649 4612 scope.go:117] "RemoveContainer" containerID="d7beccc144fa9bf9b73e2f015fc0afe88018fd5883a580e0623b4c591da30e02" Feb 27 08:09:39 crc kubenswrapper[4612]: E0227 08:09:39.967974 4612 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d7beccc144fa9bf9b73e2f015fc0afe88018fd5883a580e0623b4c591da30e02\": container with ID starting with d7beccc144fa9bf9b73e2f015fc0afe88018fd5883a580e0623b4c591da30e02 not found: ID does not exist" containerID="d7beccc144fa9bf9b73e2f015fc0afe88018fd5883a580e0623b4c591da30e02" Feb 27 08:09:39 crc kubenswrapper[4612]: I0227 08:09:39.967996 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d7beccc144fa9bf9b73e2f015fc0afe88018fd5883a580e0623b4c591da30e02"} err="failed to get container status \"d7beccc144fa9bf9b73e2f015fc0afe88018fd5883a580e0623b4c591da30e02\": rpc error: code = NotFound desc = could not find container \"d7beccc144fa9bf9b73e2f015fc0afe88018fd5883a580e0623b4c591da30e02\": container with ID starting with d7beccc144fa9bf9b73e2f015fc0afe88018fd5883a580e0623b4c591da30e02 not found: ID does not exist" Feb 27 08:09:39 crc kubenswrapper[4612]: I0227 08:09:39.968010 4612 scope.go:117] "RemoveContainer" containerID="01f1519d415fe7fb5962939cdb8fc37629da75d9f2d2645dce4cad73db91f1bb" Feb 27 08:09:39 crc kubenswrapper[4612]: I0227 08:09:39.968217 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"01f1519d415fe7fb5962939cdb8fc37629da75d9f2d2645dce4cad73db91f1bb"} err="failed to get container status \"01f1519d415fe7fb5962939cdb8fc37629da75d9f2d2645dce4cad73db91f1bb\": rpc error: code = NotFound desc = could not find container \"01f1519d415fe7fb5962939cdb8fc37629da75d9f2d2645dce4cad73db91f1bb\": container with ID starting with 01f1519d415fe7fb5962939cdb8fc37629da75d9f2d2645dce4cad73db91f1bb not found: ID does not exist" Feb 27 08:09:39 crc kubenswrapper[4612]: I0227 08:09:39.968235 4612 scope.go:117] "RemoveContainer" containerID="d7beccc144fa9bf9b73e2f015fc0afe88018fd5883a580e0623b4c591da30e02" Feb 27 08:09:39 crc kubenswrapper[4612]: I0227 08:09:39.969286 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d7beccc144fa9bf9b73e2f015fc0afe88018fd5883a580e0623b4c591da30e02"} err="failed to get container status \"d7beccc144fa9bf9b73e2f015fc0afe88018fd5883a580e0623b4c591da30e02\": rpc error: code = NotFound desc = could not find container \"d7beccc144fa9bf9b73e2f015fc0afe88018fd5883a580e0623b4c591da30e02\": container with ID starting with d7beccc144fa9bf9b73e2f015fc0afe88018fd5883a580e0623b4c591da30e02 not found: ID does not exist" Feb 27 08:09:40 crc kubenswrapper[4612]: I0227 08:09:40.040200 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2r2jz\" (UniqueName: \"kubernetes.io/projected/d62745a5-cfe3-48b0-a167-76735bdfd8fe-kube-api-access-2r2jz\") on node \"crc\" DevicePath \"\"" Feb 27 08:09:40 crc kubenswrapper[4612]: I0227 08:09:40.040232 4612 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" " Feb 27 08:09:40 crc kubenswrapper[4612]: I0227 08:09:40.040243 4612 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d62745a5-cfe3-48b0-a167-76735bdfd8fe-scripts\") on node \"crc\" DevicePath \"\"" Feb 27 08:09:40 crc kubenswrapper[4612]: I0227 08:09:40.126348 4612 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage03-crc" (UniqueName: "kubernetes.io/local-volume/local-storage03-crc") on node "crc" Feb 27 08:09:40 crc kubenswrapper[4612]: I0227 08:09:40.141327 4612 reconciler_common.go:293] "Volume detached for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" DevicePath \"\"" Feb 27 08:09:40 crc kubenswrapper[4612]: I0227 08:09:40.177661 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d62745a5-cfe3-48b0-a167-76735bdfd8fe-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d62745a5-cfe3-48b0-a167-76735bdfd8fe" (UID: "d62745a5-cfe3-48b0-a167-76735bdfd8fe"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:09:40 crc kubenswrapper[4612]: I0227 08:09:40.221024 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d62745a5-cfe3-48b0-a167-76735bdfd8fe-config-data" (OuterVolumeSpecName: "config-data") pod "d62745a5-cfe3-48b0-a167-76735bdfd8fe" (UID: "d62745a5-cfe3-48b0-a167-76735bdfd8fe"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:09:40 crc kubenswrapper[4612]: I0227 08:09:40.221861 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d62745a5-cfe3-48b0-a167-76735bdfd8fe-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "d62745a5-cfe3-48b0-a167-76735bdfd8fe" (UID: "d62745a5-cfe3-48b0-a167-76735bdfd8fe"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:09:40 crc kubenswrapper[4612]: I0227 08:09:40.247313 4612 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d62745a5-cfe3-48b0-a167-76735bdfd8fe-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 08:09:40 crc kubenswrapper[4612]: I0227 08:09:40.247567 4612 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d62745a5-cfe3-48b0-a167-76735bdfd8fe-public-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 27 08:09:40 crc kubenswrapper[4612]: I0227 08:09:40.247579 4612 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d62745a5-cfe3-48b0-a167-76735bdfd8fe-config-data\") on node \"crc\" DevicePath \"\"" Feb 27 08:09:40 crc kubenswrapper[4612]: I0227 08:09:40.529175 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 27 08:09:40 crc kubenswrapper[4612]: I0227 08:09:40.553866 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 27 08:09:40 crc kubenswrapper[4612]: I0227 08:09:40.595159 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Feb 27 08:09:40 crc kubenswrapper[4612]: E0227 08:09:40.595627 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d62745a5-cfe3-48b0-a167-76735bdfd8fe" containerName="glance-log" Feb 27 08:09:40 crc kubenswrapper[4612]: I0227 08:09:40.595649 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="d62745a5-cfe3-48b0-a167-76735bdfd8fe" containerName="glance-log" Feb 27 08:09:40 crc kubenswrapper[4612]: E0227 08:09:40.595686 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d62745a5-cfe3-48b0-a167-76735bdfd8fe" containerName="glance-httpd" Feb 27 08:09:40 crc kubenswrapper[4612]: I0227 08:09:40.595713 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="d62745a5-cfe3-48b0-a167-76735bdfd8fe" containerName="glance-httpd" Feb 27 08:09:40 crc kubenswrapper[4612]: I0227 08:09:40.595972 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="d62745a5-cfe3-48b0-a167-76735bdfd8fe" containerName="glance-log" Feb 27 08:09:40 crc kubenswrapper[4612]: I0227 08:09:40.596033 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="d62745a5-cfe3-48b0-a167-76735bdfd8fe" containerName="glance-httpd" Feb 27 08:09:40 crc kubenswrapper[4612]: I0227 08:09:40.608255 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 27 08:09:40 crc kubenswrapper[4612]: I0227 08:09:40.611602 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 27 08:09:40 crc kubenswrapper[4612]: I0227 08:09:40.616420 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Feb 27 08:09:40 crc kubenswrapper[4612]: I0227 08:09:40.616617 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Feb 27 08:09:40 crc kubenswrapper[4612]: I0227 08:09:40.656907 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/16f91699-dd9d-4e71-8ff2-2bb767b3655c-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"16f91699-dd9d-4e71-8ff2-2bb767b3655c\") " pod="openstack/glance-default-external-api-0" Feb 27 08:09:40 crc kubenswrapper[4612]: I0227 08:09:40.656951 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/16f91699-dd9d-4e71-8ff2-2bb767b3655c-logs\") pod \"glance-default-external-api-0\" (UID: \"16f91699-dd9d-4e71-8ff2-2bb767b3655c\") " pod="openstack/glance-default-external-api-0" Feb 27 08:09:40 crc kubenswrapper[4612]: I0227 08:09:40.656976 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/16f91699-dd9d-4e71-8ff2-2bb767b3655c-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"16f91699-dd9d-4e71-8ff2-2bb767b3655c\") " pod="openstack/glance-default-external-api-0" Feb 27 08:09:40 crc kubenswrapper[4612]: I0227 08:09:40.656996 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-external-api-0\" (UID: \"16f91699-dd9d-4e71-8ff2-2bb767b3655c\") " pod="openstack/glance-default-external-api-0" Feb 27 08:09:40 crc kubenswrapper[4612]: I0227 08:09:40.657030 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/16f91699-dd9d-4e71-8ff2-2bb767b3655c-config-data\") pod \"glance-default-external-api-0\" (UID: \"16f91699-dd9d-4e71-8ff2-2bb767b3655c\") " pod="openstack/glance-default-external-api-0" Feb 27 08:09:40 crc kubenswrapper[4612]: I0227 08:09:40.657048 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/16f91699-dd9d-4e71-8ff2-2bb767b3655c-scripts\") pod \"glance-default-external-api-0\" (UID: \"16f91699-dd9d-4e71-8ff2-2bb767b3655c\") " pod="openstack/glance-default-external-api-0" Feb 27 08:09:40 crc kubenswrapper[4612]: I0227 08:09:40.657071 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/16f91699-dd9d-4e71-8ff2-2bb767b3655c-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"16f91699-dd9d-4e71-8ff2-2bb767b3655c\") " pod="openstack/glance-default-external-api-0" Feb 27 08:09:40 crc kubenswrapper[4612]: I0227 08:09:40.657113 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lbkk7\" (UniqueName: \"kubernetes.io/projected/16f91699-dd9d-4e71-8ff2-2bb767b3655c-kube-api-access-lbkk7\") pod \"glance-default-external-api-0\" (UID: \"16f91699-dd9d-4e71-8ff2-2bb767b3655c\") " pod="openstack/glance-default-external-api-0" Feb 27 08:09:40 crc kubenswrapper[4612]: I0227 08:09:40.758936 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/16f91699-dd9d-4e71-8ff2-2bb767b3655c-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"16f91699-dd9d-4e71-8ff2-2bb767b3655c\") " pod="openstack/glance-default-external-api-0" Feb 27 08:09:40 crc kubenswrapper[4612]: I0227 08:09:40.758997 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/16f91699-dd9d-4e71-8ff2-2bb767b3655c-logs\") pod \"glance-default-external-api-0\" (UID: \"16f91699-dd9d-4e71-8ff2-2bb767b3655c\") " pod="openstack/glance-default-external-api-0" Feb 27 08:09:40 crc kubenswrapper[4612]: I0227 08:09:40.759019 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/16f91699-dd9d-4e71-8ff2-2bb767b3655c-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"16f91699-dd9d-4e71-8ff2-2bb767b3655c\") " pod="openstack/glance-default-external-api-0" Feb 27 08:09:40 crc kubenswrapper[4612]: I0227 08:09:40.759039 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-external-api-0\" (UID: \"16f91699-dd9d-4e71-8ff2-2bb767b3655c\") " pod="openstack/glance-default-external-api-0" Feb 27 08:09:40 crc kubenswrapper[4612]: I0227 08:09:40.759082 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/16f91699-dd9d-4e71-8ff2-2bb767b3655c-config-data\") pod \"glance-default-external-api-0\" (UID: \"16f91699-dd9d-4e71-8ff2-2bb767b3655c\") " pod="openstack/glance-default-external-api-0" Feb 27 08:09:40 crc kubenswrapper[4612]: I0227 08:09:40.759104 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/16f91699-dd9d-4e71-8ff2-2bb767b3655c-scripts\") pod \"glance-default-external-api-0\" (UID: \"16f91699-dd9d-4e71-8ff2-2bb767b3655c\") " pod="openstack/glance-default-external-api-0" Feb 27 08:09:40 crc kubenswrapper[4612]: I0227 08:09:40.759128 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/16f91699-dd9d-4e71-8ff2-2bb767b3655c-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"16f91699-dd9d-4e71-8ff2-2bb767b3655c\") " pod="openstack/glance-default-external-api-0" Feb 27 08:09:40 crc kubenswrapper[4612]: I0227 08:09:40.759176 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lbkk7\" (UniqueName: \"kubernetes.io/projected/16f91699-dd9d-4e71-8ff2-2bb767b3655c-kube-api-access-lbkk7\") pod \"glance-default-external-api-0\" (UID: \"16f91699-dd9d-4e71-8ff2-2bb767b3655c\") " pod="openstack/glance-default-external-api-0" Feb 27 08:09:40 crc kubenswrapper[4612]: I0227 08:09:40.759525 4612 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-external-api-0\" (UID: \"16f91699-dd9d-4e71-8ff2-2bb767b3655c\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/glance-default-external-api-0" Feb 27 08:09:40 crc kubenswrapper[4612]: I0227 08:09:40.763387 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/16f91699-dd9d-4e71-8ff2-2bb767b3655c-logs\") pod \"glance-default-external-api-0\" (UID: \"16f91699-dd9d-4e71-8ff2-2bb767b3655c\") " pod="openstack/glance-default-external-api-0" Feb 27 08:09:40 crc kubenswrapper[4612]: I0227 08:09:40.763637 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/16f91699-dd9d-4e71-8ff2-2bb767b3655c-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"16f91699-dd9d-4e71-8ff2-2bb767b3655c\") " pod="openstack/glance-default-external-api-0" Feb 27 08:09:40 crc kubenswrapper[4612]: I0227 08:09:40.767417 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/16f91699-dd9d-4e71-8ff2-2bb767b3655c-config-data\") pod \"glance-default-external-api-0\" (UID: \"16f91699-dd9d-4e71-8ff2-2bb767b3655c\") " pod="openstack/glance-default-external-api-0" Feb 27 08:09:40 crc kubenswrapper[4612]: I0227 08:09:40.768426 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/16f91699-dd9d-4e71-8ff2-2bb767b3655c-scripts\") pod \"glance-default-external-api-0\" (UID: \"16f91699-dd9d-4e71-8ff2-2bb767b3655c\") " pod="openstack/glance-default-external-api-0" Feb 27 08:09:40 crc kubenswrapper[4612]: I0227 08:09:40.771325 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/16f91699-dd9d-4e71-8ff2-2bb767b3655c-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"16f91699-dd9d-4e71-8ff2-2bb767b3655c\") " pod="openstack/glance-default-external-api-0" Feb 27 08:09:40 crc kubenswrapper[4612]: I0227 08:09:40.785375 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/16f91699-dd9d-4e71-8ff2-2bb767b3655c-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"16f91699-dd9d-4e71-8ff2-2bb767b3655c\") " pod="openstack/glance-default-external-api-0" Feb 27 08:09:40 crc kubenswrapper[4612]: I0227 08:09:40.791353 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lbkk7\" (UniqueName: \"kubernetes.io/projected/16f91699-dd9d-4e71-8ff2-2bb767b3655c-kube-api-access-lbkk7\") pod \"glance-default-external-api-0\" (UID: \"16f91699-dd9d-4e71-8ff2-2bb767b3655c\") " pod="openstack/glance-default-external-api-0" Feb 27 08:09:40 crc kubenswrapper[4612]: I0227 08:09:40.815905 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-external-api-0\" (UID: \"16f91699-dd9d-4e71-8ff2-2bb767b3655c\") " pod="openstack/glance-default-external-api-0" Feb 27 08:09:40 crc kubenswrapper[4612]: I0227 08:09:40.866154 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d62745a5-cfe3-48b0-a167-76735bdfd8fe" path="/var/lib/kubelet/pods/d62745a5-cfe3-48b0-a167-76735bdfd8fe/volumes" Feb 27 08:09:40 crc kubenswrapper[4612]: I0227 08:09:40.910583 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-77854b564-8nkbp" event={"ID":"a9397555-a3e3-4a49-b725-882006d2ef0d","Type":"ContainerStarted","Data":"99db33fb196282606d81b167cb3314cc2ecbee32c696435c83491070eba30343"} Feb 27 08:09:40 crc kubenswrapper[4612]: I0227 08:09:40.910875 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-77854b564-8nkbp" Feb 27 08:09:40 crc kubenswrapper[4612]: I0227 08:09:40.914391 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b7b667979-pm5fl" event={"ID":"bf325571-0b74-447f-bd04-213d54a8437e","Type":"ContainerStarted","Data":"cd920bf9cdd1f21efee933e8fb8dc84fd5aac93ceafa5fc01745e88c2e9aaf86"} Feb 27 08:09:40 crc kubenswrapper[4612]: I0227 08:09:40.915067 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6b7b667979-pm5fl" Feb 27 08:09:40 crc kubenswrapper[4612]: I0227 08:09:40.918195 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9ab32d5a-a797-47a8-961a-a534f55fdfa7","Type":"ContainerStarted","Data":"0b87c11e9c758edb6f988ef0ee90e9d6d19fa4c1d46725a5fdb737090d12c6be"} Feb 27 08:09:40 crc kubenswrapper[4612]: I0227 08:09:40.919831 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6b985547d4-mkfpv" event={"ID":"76009243-16b3-4652-abfe-a6cdba363724","Type":"ContainerStarted","Data":"a0baf671a66706916e8b02ab97ccdead951de20e3f62e64d913a8590aa359cb3"} Feb 27 08:09:40 crc kubenswrapper[4612]: I0227 08:09:40.919939 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6b985547d4-mkfpv" event={"ID":"76009243-16b3-4652-abfe-a6cdba363724","Type":"ContainerStarted","Data":"deeb662d0bcfb1868b750b03f6073e2ff029f55cf5b6b403283389a06b1b3b17"} Feb 27 08:09:40 crc kubenswrapper[4612]: I0227 08:09:40.924156 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"4726ffcd-f407-4cd4-a6e4-c422801f2617","Type":"ContainerStarted","Data":"6aad69d61ff9ee83297a8b21d4c7bff687f56c29c744c4b9faf9c1c955d94f06"} Feb 27 08:09:40 crc kubenswrapper[4612]: I0227 08:09:40.924300 4612 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="4726ffcd-f407-4cd4-a6e4-c422801f2617" containerName="glance-log" containerID="cri-o://a661b93501b59c43ff694deb74a3b3bb3649f7a43b08aa70241c6752498b2d38" gracePeriod=30 Feb 27 08:09:40 crc kubenswrapper[4612]: I0227 08:09:40.924583 4612 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="4726ffcd-f407-4cd4-a6e4-c422801f2617" containerName="glance-httpd" containerID="cri-o://6aad69d61ff9ee83297a8b21d4c7bff687f56c29c744c4b9faf9c1c955d94f06" gracePeriod=30 Feb 27 08:09:40 crc kubenswrapper[4612]: I0227 08:09:40.935784 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 27 08:09:40 crc kubenswrapper[4612]: I0227 08:09:40.936101 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-77854b564-8nkbp" podStartSLOduration=5.936079661 podStartE2EDuration="5.936079661s" podCreationTimestamp="2026-02-27 08:09:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 08:09:40.932368195 +0000 UTC m=+1238.786298193" watchObservedRunningTime="2026-02-27 08:09:40.936079661 +0000 UTC m=+1238.790009669" Feb 27 08:09:40 crc kubenswrapper[4612]: I0227 08:09:40.936783 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-77948db5bb-97qh6" event={"ID":"b72b541d-4661-44a7-a121-c2a8aef6db11","Type":"ContainerStarted","Data":"92102897b17cb0c0c81984cbb9f2014fd0591e9bae23a94bbb5532b66e484fd8"} Feb 27 08:09:40 crc kubenswrapper[4612]: I0227 08:09:40.945304 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5ddbbb8689-bm5cp" event={"ID":"649e0df0-b37c-42f5-8af0-c804d4f467d4","Type":"ContainerStarted","Data":"399439f0cfe95ee5a2d4886d8c116dcf58ccc56b31529381b60a5bb7558643db"} Feb 27 08:09:40 crc kubenswrapper[4612]: I0227 08:09:40.945349 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5ddbbb8689-bm5cp" event={"ID":"649e0df0-b37c-42f5-8af0-c804d4f467d4","Type":"ContainerStarted","Data":"c8e24f4c991021623da34d4abac610956069b675ba545401121c750c3d4a1070"} Feb 27 08:09:40 crc kubenswrapper[4612]: I0227 08:09:40.945527 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-5ddbbb8689-bm5cp" Feb 27 08:09:40 crc kubenswrapper[4612]: I0227 08:09:40.973052 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-6b985547d4-mkfpv" podStartSLOduration=31.871399905 podStartE2EDuration="32.97303401s" podCreationTimestamp="2026-02-27 08:09:08 +0000 UTC" firstStartedPulling="2026-02-27 08:09:38.220470369 +0000 UTC m=+1236.074400367" lastFinishedPulling="2026-02-27 08:09:39.322104474 +0000 UTC m=+1237.176034472" observedRunningTime="2026-02-27 08:09:40.96219946 +0000 UTC m=+1238.816129458" watchObservedRunningTime="2026-02-27 08:09:40.97303401 +0000 UTC m=+1238.826964008" Feb 27 08:09:41 crc kubenswrapper[4612]: I0227 08:09:41.007541 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6b7b667979-pm5fl" podStartSLOduration=6.007520898 podStartE2EDuration="6.007520898s" podCreationTimestamp="2026-02-27 08:09:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 08:09:40.992172489 +0000 UTC m=+1238.846102487" watchObservedRunningTime="2026-02-27 08:09:41.007520898 +0000 UTC m=+1238.861450896" Feb 27 08:09:41 crc kubenswrapper[4612]: I0227 08:09:41.026623 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=34.026602865 podStartE2EDuration="34.026602865s" podCreationTimestamp="2026-02-27 08:09:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 08:09:41.022898089 +0000 UTC m=+1238.876828097" watchObservedRunningTime="2026-02-27 08:09:41.026602865 +0000 UTC m=+1238.880532873" Feb 27 08:09:41 crc kubenswrapper[4612]: I0227 08:09:41.064312 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-77948db5bb-97qh6" podStartSLOduration=32.45096423 podStartE2EDuration="33.064296715s" podCreationTimestamp="2026-02-27 08:09:08 +0000 UTC" firstStartedPulling="2026-02-27 08:09:37.912652958 +0000 UTC m=+1235.766582956" lastFinishedPulling="2026-02-27 08:09:38.525985443 +0000 UTC m=+1236.379915441" observedRunningTime="2026-02-27 08:09:41.055949606 +0000 UTC m=+1238.909879604" watchObservedRunningTime="2026-02-27 08:09:41.064296715 +0000 UTC m=+1238.918226713" Feb 27 08:09:41 crc kubenswrapper[4612]: I0227 08:09:41.146164 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-5ddbbb8689-bm5cp" podStartSLOduration=5.14614324 podStartE2EDuration="5.14614324s" podCreationTimestamp="2026-02-27 08:09:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 08:09:41.098250228 +0000 UTC m=+1238.952180226" watchObservedRunningTime="2026-02-27 08:09:41.14614324 +0000 UTC m=+1239.000073238" Feb 27 08:09:41 crc kubenswrapper[4612]: I0227 08:09:41.589684 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 27 08:09:41 crc kubenswrapper[4612]: I0227 08:09:41.963943 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"16f91699-dd9d-4e71-8ff2-2bb767b3655c","Type":"ContainerStarted","Data":"53130813b4018d05f1ac9e30b5ca1a3ee60f45537c457fde1fe6161d6a333c54"} Feb 27 08:09:41 crc kubenswrapper[4612]: I0227 08:09:41.968775 4612 generic.go:334] "Generic (PLEG): container finished" podID="4726ffcd-f407-4cd4-a6e4-c422801f2617" containerID="6aad69d61ff9ee83297a8b21d4c7bff687f56c29c744c4b9faf9c1c955d94f06" exitCode=0 Feb 27 08:09:41 crc kubenswrapper[4612]: I0227 08:09:41.968810 4612 generic.go:334] "Generic (PLEG): container finished" podID="4726ffcd-f407-4cd4-a6e4-c422801f2617" containerID="a661b93501b59c43ff694deb74a3b3bb3649f7a43b08aa70241c6752498b2d38" exitCode=143 Feb 27 08:09:41 crc kubenswrapper[4612]: I0227 08:09:41.969668 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"4726ffcd-f407-4cd4-a6e4-c422801f2617","Type":"ContainerDied","Data":"6aad69d61ff9ee83297a8b21d4c7bff687f56c29c744c4b9faf9c1c955d94f06"} Feb 27 08:09:41 crc kubenswrapper[4612]: I0227 08:09:41.969710 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"4726ffcd-f407-4cd4-a6e4-c422801f2617","Type":"ContainerDied","Data":"a661b93501b59c43ff694deb74a3b3bb3649f7a43b08aa70241c6752498b2d38"} Feb 27 08:09:42 crc kubenswrapper[4612]: I0227 08:09:42.905598 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 27 08:09:42 crc kubenswrapper[4612]: I0227 08:09:42.996522 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"16f91699-dd9d-4e71-8ff2-2bb767b3655c","Type":"ContainerStarted","Data":"6539429dfe0f2fb854db7c705771db9d5e7a1a793224317e2b3dee4af19646b5"} Feb 27 08:09:43 crc kubenswrapper[4612]: I0227 08:09:43.001855 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"4726ffcd-f407-4cd4-a6e4-c422801f2617","Type":"ContainerDied","Data":"a474aa8d3910d92396577ba685fe0afa2b21b30098d34cdf06cdf8e1e660a9c4"} Feb 27 08:09:43 crc kubenswrapper[4612]: I0227 08:09:43.001910 4612 scope.go:117] "RemoveContainer" containerID="6aad69d61ff9ee83297a8b21d4c7bff687f56c29c744c4b9faf9c1c955d94f06" Feb 27 08:09:43 crc kubenswrapper[4612]: I0227 08:09:43.001953 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 27 08:09:43 crc kubenswrapper[4612]: I0227 08:09:43.021304 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4726ffcd-f407-4cd4-a6e4-c422801f2617-config-data\") pod \"4726ffcd-f407-4cd4-a6e4-c422801f2617\" (UID: \"4726ffcd-f407-4cd4-a6e4-c422801f2617\") " Feb 27 08:09:43 crc kubenswrapper[4612]: I0227 08:09:43.021378 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"4726ffcd-f407-4cd4-a6e4-c422801f2617\" (UID: \"4726ffcd-f407-4cd4-a6e4-c422801f2617\") " Feb 27 08:09:43 crc kubenswrapper[4612]: I0227 08:09:43.021425 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4726ffcd-f407-4cd4-a6e4-c422801f2617-scripts\") pod \"4726ffcd-f407-4cd4-a6e4-c422801f2617\" (UID: \"4726ffcd-f407-4cd4-a6e4-c422801f2617\") " Feb 27 08:09:43 crc kubenswrapper[4612]: I0227 08:09:43.021468 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4726ffcd-f407-4cd4-a6e4-c422801f2617-combined-ca-bundle\") pod \"4726ffcd-f407-4cd4-a6e4-c422801f2617\" (UID: \"4726ffcd-f407-4cd4-a6e4-c422801f2617\") " Feb 27 08:09:43 crc kubenswrapper[4612]: I0227 08:09:43.021555 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-645t5\" (UniqueName: \"kubernetes.io/projected/4726ffcd-f407-4cd4-a6e4-c422801f2617-kube-api-access-645t5\") pod \"4726ffcd-f407-4cd4-a6e4-c422801f2617\" (UID: \"4726ffcd-f407-4cd4-a6e4-c422801f2617\") " Feb 27 08:09:43 crc kubenswrapper[4612]: I0227 08:09:43.021592 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4726ffcd-f407-4cd4-a6e4-c422801f2617-internal-tls-certs\") pod \"4726ffcd-f407-4cd4-a6e4-c422801f2617\" (UID: \"4726ffcd-f407-4cd4-a6e4-c422801f2617\") " Feb 27 08:09:43 crc kubenswrapper[4612]: I0227 08:09:43.021618 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/4726ffcd-f407-4cd4-a6e4-c422801f2617-httpd-run\") pod \"4726ffcd-f407-4cd4-a6e4-c422801f2617\" (UID: \"4726ffcd-f407-4cd4-a6e4-c422801f2617\") " Feb 27 08:09:43 crc kubenswrapper[4612]: I0227 08:09:43.021641 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4726ffcd-f407-4cd4-a6e4-c422801f2617-logs\") pod \"4726ffcd-f407-4cd4-a6e4-c422801f2617\" (UID: \"4726ffcd-f407-4cd4-a6e4-c422801f2617\") " Feb 27 08:09:43 crc kubenswrapper[4612]: I0227 08:09:43.023767 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4726ffcd-f407-4cd4-a6e4-c422801f2617-logs" (OuterVolumeSpecName: "logs") pod "4726ffcd-f407-4cd4-a6e4-c422801f2617" (UID: "4726ffcd-f407-4cd4-a6e4-c422801f2617"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 08:09:43 crc kubenswrapper[4612]: I0227 08:09:43.023974 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4726ffcd-f407-4cd4-a6e4-c422801f2617-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "4726ffcd-f407-4cd4-a6e4-c422801f2617" (UID: "4726ffcd-f407-4cd4-a6e4-c422801f2617"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 08:09:43 crc kubenswrapper[4612]: I0227 08:09:43.029805 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4726ffcd-f407-4cd4-a6e4-c422801f2617-kube-api-access-645t5" (OuterVolumeSpecName: "kube-api-access-645t5") pod "4726ffcd-f407-4cd4-a6e4-c422801f2617" (UID: "4726ffcd-f407-4cd4-a6e4-c422801f2617"). InnerVolumeSpecName "kube-api-access-645t5". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 08:09:43 crc kubenswrapper[4612]: I0227 08:09:43.031116 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4726ffcd-f407-4cd4-a6e4-c422801f2617-scripts" (OuterVolumeSpecName: "scripts") pod "4726ffcd-f407-4cd4-a6e4-c422801f2617" (UID: "4726ffcd-f407-4cd4-a6e4-c422801f2617"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:09:43 crc kubenswrapper[4612]: I0227 08:09:43.052063 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage04-crc" (OuterVolumeSpecName: "glance") pod "4726ffcd-f407-4cd4-a6e4-c422801f2617" (UID: "4726ffcd-f407-4cd4-a6e4-c422801f2617"). InnerVolumeSpecName "local-storage04-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Feb 27 08:09:43 crc kubenswrapper[4612]: I0227 08:09:43.067530 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4726ffcd-f407-4cd4-a6e4-c422801f2617-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4726ffcd-f407-4cd4-a6e4-c422801f2617" (UID: "4726ffcd-f407-4cd4-a6e4-c422801f2617"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:09:43 crc kubenswrapper[4612]: I0227 08:09:43.103003 4612 scope.go:117] "RemoveContainer" containerID="a661b93501b59c43ff694deb74a3b3bb3649f7a43b08aa70241c6752498b2d38" Feb 27 08:09:43 crc kubenswrapper[4612]: I0227 08:09:43.106507 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4726ffcd-f407-4cd4-a6e4-c422801f2617-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "4726ffcd-f407-4cd4-a6e4-c422801f2617" (UID: "4726ffcd-f407-4cd4-a6e4-c422801f2617"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:09:43 crc kubenswrapper[4612]: I0227 08:09:43.123790 4612 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4726ffcd-f407-4cd4-a6e4-c422801f2617-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 08:09:43 crc kubenswrapper[4612]: I0227 08:09:43.123820 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-645t5\" (UniqueName: \"kubernetes.io/projected/4726ffcd-f407-4cd4-a6e4-c422801f2617-kube-api-access-645t5\") on node \"crc\" DevicePath \"\"" Feb 27 08:09:43 crc kubenswrapper[4612]: I0227 08:09:43.123830 4612 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4726ffcd-f407-4cd4-a6e4-c422801f2617-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 27 08:09:43 crc kubenswrapper[4612]: I0227 08:09:43.123838 4612 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/4726ffcd-f407-4cd4-a6e4-c422801f2617-httpd-run\") on node \"crc\" DevicePath \"\"" Feb 27 08:09:43 crc kubenswrapper[4612]: I0227 08:09:43.123846 4612 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4726ffcd-f407-4cd4-a6e4-c422801f2617-logs\") on node \"crc\" DevicePath \"\"" Feb 27 08:09:43 crc kubenswrapper[4612]: I0227 08:09:43.123871 4612 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" " Feb 27 08:09:43 crc kubenswrapper[4612]: I0227 08:09:43.123881 4612 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4726ffcd-f407-4cd4-a6e4-c422801f2617-scripts\") on node \"crc\" DevicePath \"\"" Feb 27 08:09:43 crc kubenswrapper[4612]: I0227 08:09:43.152039 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4726ffcd-f407-4cd4-a6e4-c422801f2617-config-data" (OuterVolumeSpecName: "config-data") pod "4726ffcd-f407-4cd4-a6e4-c422801f2617" (UID: "4726ffcd-f407-4cd4-a6e4-c422801f2617"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:09:43 crc kubenswrapper[4612]: I0227 08:09:43.155908 4612 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage04-crc" (UniqueName: "kubernetes.io/local-volume/local-storage04-crc") on node "crc" Feb 27 08:09:43 crc kubenswrapper[4612]: I0227 08:09:43.225732 4612 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4726ffcd-f407-4cd4-a6e4-c422801f2617-config-data\") on node \"crc\" DevicePath \"\"" Feb 27 08:09:43 crc kubenswrapper[4612]: I0227 08:09:43.225765 4612 reconciler_common.go:293] "Volume detached for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" DevicePath \"\"" Feb 27 08:09:43 crc kubenswrapper[4612]: I0227 08:09:43.362830 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 27 08:09:43 crc kubenswrapper[4612]: I0227 08:09:43.374115 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 27 08:09:43 crc kubenswrapper[4612]: I0227 08:09:43.409370 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 27 08:09:43 crc kubenswrapper[4612]: E0227 08:09:43.409824 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4726ffcd-f407-4cd4-a6e4-c422801f2617" containerName="glance-log" Feb 27 08:09:43 crc kubenswrapper[4612]: I0227 08:09:43.409842 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="4726ffcd-f407-4cd4-a6e4-c422801f2617" containerName="glance-log" Feb 27 08:09:43 crc kubenswrapper[4612]: E0227 08:09:43.409853 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4726ffcd-f407-4cd4-a6e4-c422801f2617" containerName="glance-httpd" Feb 27 08:09:43 crc kubenswrapper[4612]: I0227 08:09:43.409859 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="4726ffcd-f407-4cd4-a6e4-c422801f2617" containerName="glance-httpd" Feb 27 08:09:43 crc kubenswrapper[4612]: I0227 08:09:43.410058 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="4726ffcd-f407-4cd4-a6e4-c422801f2617" containerName="glance-log" Feb 27 08:09:43 crc kubenswrapper[4612]: I0227 08:09:43.410075 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="4726ffcd-f407-4cd4-a6e4-c422801f2617" containerName="glance-httpd" Feb 27 08:09:43 crc kubenswrapper[4612]: I0227 08:09:43.411017 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 27 08:09:43 crc kubenswrapper[4612]: I0227 08:09:43.415261 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Feb 27 08:09:43 crc kubenswrapper[4612]: I0227 08:09:43.418733 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Feb 27 08:09:43 crc kubenswrapper[4612]: I0227 08:09:43.418749 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 27 08:09:43 crc kubenswrapper[4612]: I0227 08:09:43.433356 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"f3bb8dc2-0c27-4a7a-baa1-1bfd8916f920\") " pod="openstack/glance-default-internal-api-0" Feb 27 08:09:43 crc kubenswrapper[4612]: I0227 08:09:43.433410 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f3bb8dc2-0c27-4a7a-baa1-1bfd8916f920-logs\") pod \"glance-default-internal-api-0\" (UID: \"f3bb8dc2-0c27-4a7a-baa1-1bfd8916f920\") " pod="openstack/glance-default-internal-api-0" Feb 27 08:09:43 crc kubenswrapper[4612]: I0227 08:09:43.433435 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kqqmt\" (UniqueName: \"kubernetes.io/projected/f3bb8dc2-0c27-4a7a-baa1-1bfd8916f920-kube-api-access-kqqmt\") pod \"glance-default-internal-api-0\" (UID: \"f3bb8dc2-0c27-4a7a-baa1-1bfd8916f920\") " pod="openstack/glance-default-internal-api-0" Feb 27 08:09:43 crc kubenswrapper[4612]: I0227 08:09:43.433465 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f3bb8dc2-0c27-4a7a-baa1-1bfd8916f920-scripts\") pod \"glance-default-internal-api-0\" (UID: \"f3bb8dc2-0c27-4a7a-baa1-1bfd8916f920\") " pod="openstack/glance-default-internal-api-0" Feb 27 08:09:43 crc kubenswrapper[4612]: I0227 08:09:43.433499 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f3bb8dc2-0c27-4a7a-baa1-1bfd8916f920-config-data\") pod \"glance-default-internal-api-0\" (UID: \"f3bb8dc2-0c27-4a7a-baa1-1bfd8916f920\") " pod="openstack/glance-default-internal-api-0" Feb 27 08:09:43 crc kubenswrapper[4612]: I0227 08:09:43.433539 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f3bb8dc2-0c27-4a7a-baa1-1bfd8916f920-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"f3bb8dc2-0c27-4a7a-baa1-1bfd8916f920\") " pod="openstack/glance-default-internal-api-0" Feb 27 08:09:43 crc kubenswrapper[4612]: I0227 08:09:43.433611 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3bb8dc2-0c27-4a7a-baa1-1bfd8916f920-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"f3bb8dc2-0c27-4a7a-baa1-1bfd8916f920\") " pod="openstack/glance-default-internal-api-0" Feb 27 08:09:43 crc kubenswrapper[4612]: I0227 08:09:43.433635 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f3bb8dc2-0c27-4a7a-baa1-1bfd8916f920-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"f3bb8dc2-0c27-4a7a-baa1-1bfd8916f920\") " pod="openstack/glance-default-internal-api-0" Feb 27 08:09:43 crc kubenswrapper[4612]: I0227 08:09:43.536620 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f3bb8dc2-0c27-4a7a-baa1-1bfd8916f920-scripts\") pod \"glance-default-internal-api-0\" (UID: \"f3bb8dc2-0c27-4a7a-baa1-1bfd8916f920\") " pod="openstack/glance-default-internal-api-0" Feb 27 08:09:43 crc kubenswrapper[4612]: I0227 08:09:43.536703 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f3bb8dc2-0c27-4a7a-baa1-1bfd8916f920-config-data\") pod \"glance-default-internal-api-0\" (UID: \"f3bb8dc2-0c27-4a7a-baa1-1bfd8916f920\") " pod="openstack/glance-default-internal-api-0" Feb 27 08:09:43 crc kubenswrapper[4612]: I0227 08:09:43.536744 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f3bb8dc2-0c27-4a7a-baa1-1bfd8916f920-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"f3bb8dc2-0c27-4a7a-baa1-1bfd8916f920\") " pod="openstack/glance-default-internal-api-0" Feb 27 08:09:43 crc kubenswrapper[4612]: I0227 08:09:43.536797 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3bb8dc2-0c27-4a7a-baa1-1bfd8916f920-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"f3bb8dc2-0c27-4a7a-baa1-1bfd8916f920\") " pod="openstack/glance-default-internal-api-0" Feb 27 08:09:43 crc kubenswrapper[4612]: I0227 08:09:43.536815 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f3bb8dc2-0c27-4a7a-baa1-1bfd8916f920-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"f3bb8dc2-0c27-4a7a-baa1-1bfd8916f920\") " pod="openstack/glance-default-internal-api-0" Feb 27 08:09:43 crc kubenswrapper[4612]: I0227 08:09:43.536856 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"f3bb8dc2-0c27-4a7a-baa1-1bfd8916f920\") " pod="openstack/glance-default-internal-api-0" Feb 27 08:09:43 crc kubenswrapper[4612]: I0227 08:09:43.536883 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f3bb8dc2-0c27-4a7a-baa1-1bfd8916f920-logs\") pod \"glance-default-internal-api-0\" (UID: \"f3bb8dc2-0c27-4a7a-baa1-1bfd8916f920\") " pod="openstack/glance-default-internal-api-0" Feb 27 08:09:43 crc kubenswrapper[4612]: I0227 08:09:43.536904 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kqqmt\" (UniqueName: \"kubernetes.io/projected/f3bb8dc2-0c27-4a7a-baa1-1bfd8916f920-kube-api-access-kqqmt\") pod \"glance-default-internal-api-0\" (UID: \"f3bb8dc2-0c27-4a7a-baa1-1bfd8916f920\") " pod="openstack/glance-default-internal-api-0" Feb 27 08:09:43 crc kubenswrapper[4612]: I0227 08:09:43.537722 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f3bb8dc2-0c27-4a7a-baa1-1bfd8916f920-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"f3bb8dc2-0c27-4a7a-baa1-1bfd8916f920\") " pod="openstack/glance-default-internal-api-0" Feb 27 08:09:43 crc kubenswrapper[4612]: I0227 08:09:43.538007 4612 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"f3bb8dc2-0c27-4a7a-baa1-1bfd8916f920\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/glance-default-internal-api-0" Feb 27 08:09:43 crc kubenswrapper[4612]: I0227 08:09:43.549186 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f3bb8dc2-0c27-4a7a-baa1-1bfd8916f920-logs\") pod \"glance-default-internal-api-0\" (UID: \"f3bb8dc2-0c27-4a7a-baa1-1bfd8916f920\") " pod="openstack/glance-default-internal-api-0" Feb 27 08:09:43 crc kubenswrapper[4612]: I0227 08:09:43.564604 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3bb8dc2-0c27-4a7a-baa1-1bfd8916f920-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"f3bb8dc2-0c27-4a7a-baa1-1bfd8916f920\") " pod="openstack/glance-default-internal-api-0" Feb 27 08:09:43 crc kubenswrapper[4612]: I0227 08:09:43.574164 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f3bb8dc2-0c27-4a7a-baa1-1bfd8916f920-config-data\") pod \"glance-default-internal-api-0\" (UID: \"f3bb8dc2-0c27-4a7a-baa1-1bfd8916f920\") " pod="openstack/glance-default-internal-api-0" Feb 27 08:09:43 crc kubenswrapper[4612]: I0227 08:09:43.582135 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kqqmt\" (UniqueName: \"kubernetes.io/projected/f3bb8dc2-0c27-4a7a-baa1-1bfd8916f920-kube-api-access-kqqmt\") pod \"glance-default-internal-api-0\" (UID: \"f3bb8dc2-0c27-4a7a-baa1-1bfd8916f920\") " pod="openstack/glance-default-internal-api-0" Feb 27 08:09:43 crc kubenswrapper[4612]: I0227 08:09:43.582623 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f3bb8dc2-0c27-4a7a-baa1-1bfd8916f920-scripts\") pod \"glance-default-internal-api-0\" (UID: \"f3bb8dc2-0c27-4a7a-baa1-1bfd8916f920\") " pod="openstack/glance-default-internal-api-0" Feb 27 08:09:43 crc kubenswrapper[4612]: I0227 08:09:43.583369 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f3bb8dc2-0c27-4a7a-baa1-1bfd8916f920-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"f3bb8dc2-0c27-4a7a-baa1-1bfd8916f920\") " pod="openstack/glance-default-internal-api-0" Feb 27 08:09:43 crc kubenswrapper[4612]: I0227 08:09:43.633137 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"f3bb8dc2-0c27-4a7a-baa1-1bfd8916f920\") " pod="openstack/glance-default-internal-api-0" Feb 27 08:09:43 crc kubenswrapper[4612]: I0227 08:09:43.762375 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 27 08:09:44 crc kubenswrapper[4612]: I0227 08:09:44.046309 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"16f91699-dd9d-4e71-8ff2-2bb767b3655c","Type":"ContainerStarted","Data":"ddc49c3fd91b29cf3ba40b6f8e5717f9961e8e1840b755bb2cd2acf23ab56475"} Feb 27 08:09:44 crc kubenswrapper[4612]: I0227 08:09:44.553561 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=4.553538046 podStartE2EDuration="4.553538046s" podCreationTimestamp="2026-02-27 08:09:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 08:09:44.083207279 +0000 UTC m=+1241.937137277" watchObservedRunningTime="2026-02-27 08:09:44.553538046 +0000 UTC m=+1242.407468044" Feb 27 08:09:44 crc kubenswrapper[4612]: I0227 08:09:44.563392 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 27 08:09:44 crc kubenswrapper[4612]: I0227 08:09:44.864531 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4726ffcd-f407-4cd4-a6e4-c422801f2617" path="/var/lib/kubelet/pods/4726ffcd-f407-4cd4-a6e4-c422801f2617/volumes" Feb 27 08:09:45 crc kubenswrapper[4612]: I0227 08:09:45.088765 4612 generic.go:334] "Generic (PLEG): container finished" podID="36b549a3-c6e9-4f7e-8cff-85216c8e0e3d" containerID="a6d45103e694a574568b137e43fc7bcdcd0e18154a2c6b7ba6c90c509ea0da15" exitCode=0 Feb 27 08:09:45 crc kubenswrapper[4612]: I0227 08:09:45.088928 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-8zrcl" event={"ID":"36b549a3-c6e9-4f7e-8cff-85216c8e0e3d","Type":"ContainerDied","Data":"a6d45103e694a574568b137e43fc7bcdcd0e18154a2c6b7ba6c90c509ea0da15"} Feb 27 08:09:45 crc kubenswrapper[4612]: I0227 08:09:45.430952 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6b7b667979-pm5fl" Feb 27 08:09:45 crc kubenswrapper[4612]: I0227 08:09:45.507100 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-56df8fb6b7-9p274"] Feb 27 08:09:45 crc kubenswrapper[4612]: I0227 08:09:45.507338 4612 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-56df8fb6b7-9p274" podUID="fd3bd506-aad1-468a-b469-64ad78ec33fe" containerName="dnsmasq-dns" containerID="cri-o://10abcf416e4e5487e724027816bc29df51d2715c2f37c2ad093bd0b055e40712" gracePeriod=10 Feb 27 08:09:46 crc kubenswrapper[4612]: I0227 08:09:46.098976 4612 generic.go:334] "Generic (PLEG): container finished" podID="fd3bd506-aad1-468a-b469-64ad78ec33fe" containerID="10abcf416e4e5487e724027816bc29df51d2715c2f37c2ad093bd0b055e40712" exitCode=0 Feb 27 08:09:46 crc kubenswrapper[4612]: I0227 08:09:46.099069 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56df8fb6b7-9p274" event={"ID":"fd3bd506-aad1-468a-b469-64ad78ec33fe","Type":"ContainerDied","Data":"10abcf416e4e5487e724027816bc29df51d2715c2f37c2ad093bd0b055e40712"} Feb 27 08:09:48 crc kubenswrapper[4612]: I0227 08:09:48.124885 4612 generic.go:334] "Generic (PLEG): container finished" podID="87df2a96-7ae3-47af-92bb-97db038b957c" containerID="7a14e07a854b7c34d66faacc2713fc09fcde83a58ea6b802875545f39c503506" exitCode=0 Feb 27 08:09:48 crc kubenswrapper[4612]: I0227 08:09:48.124948 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-r8dpp" event={"ID":"87df2a96-7ae3-47af-92bb-97db038b957c","Type":"ContainerDied","Data":"7a14e07a854b7c34d66faacc2713fc09fcde83a58ea6b802875545f39c503506"} Feb 27 08:09:49 crc kubenswrapper[4612]: I0227 08:09:49.106984 4612 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-6b985547d4-mkfpv" Feb 27 08:09:49 crc kubenswrapper[4612]: I0227 08:09:49.107024 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-6b985547d4-mkfpv" Feb 27 08:09:49 crc kubenswrapper[4612]: I0227 08:09:49.134672 4612 generic.go:334] "Generic (PLEG): container finished" podID="1a05a7e5-8c46-441a-9faf-0fd3b41bbd98" containerID="d7fbb0369e2f86e02466cb49c5f6a3bba32f32d2e771e988cd25937fb4d8dd80" exitCode=0 Feb 27 08:09:49 crc kubenswrapper[4612]: I0227 08:09:49.134884 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-s5bjv" event={"ID":"1a05a7e5-8c46-441a-9faf-0fd3b41bbd98","Type":"ContainerDied","Data":"d7fbb0369e2f86e02466cb49c5f6a3bba32f32d2e771e988cd25937fb4d8dd80"} Feb 27 08:09:49 crc kubenswrapper[4612]: I0227 08:09:49.272048 4612 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-77948db5bb-97qh6" Feb 27 08:09:49 crc kubenswrapper[4612]: I0227 08:09:49.272089 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-77948db5bb-97qh6" Feb 27 08:09:50 crc kubenswrapper[4612]: I0227 08:09:50.263431 4612 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-56df8fb6b7-9p274" podUID="fd3bd506-aad1-468a-b469-64ad78ec33fe" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.143:5353: connect: connection refused" Feb 27 08:09:50 crc kubenswrapper[4612]: I0227 08:09:50.842531 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-r8dpp" Feb 27 08:09:50 crc kubenswrapper[4612]: I0227 08:09:50.848700 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-8zrcl" Feb 27 08:09:50 crc kubenswrapper[4612]: I0227 08:09:50.887296 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-s5bjv" Feb 27 08:09:50 crc kubenswrapper[4612]: I0227 08:09:50.953919 4612 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Feb 27 08:09:50 crc kubenswrapper[4612]: I0227 08:09:50.954062 4612 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Feb 27 08:09:50 crc kubenswrapper[4612]: I0227 08:09:50.996844 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36b549a3-c6e9-4f7e-8cff-85216c8e0e3d-combined-ca-bundle\") pod \"36b549a3-c6e9-4f7e-8cff-85216c8e0e3d\" (UID: \"36b549a3-c6e9-4f7e-8cff-85216c8e0e3d\") " Feb 27 08:09:50 crc kubenswrapper[4612]: I0227 08:09:50.996897 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1a05a7e5-8c46-441a-9faf-0fd3b41bbd98-config-data\") pod \"1a05a7e5-8c46-441a-9faf-0fd3b41bbd98\" (UID: \"1a05a7e5-8c46-441a-9faf-0fd3b41bbd98\") " Feb 27 08:09:50 crc kubenswrapper[4612]: I0227 08:09:50.996925 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b6hqv\" (UniqueName: \"kubernetes.io/projected/1a05a7e5-8c46-441a-9faf-0fd3b41bbd98-kube-api-access-b6hqv\") pod \"1a05a7e5-8c46-441a-9faf-0fd3b41bbd98\" (UID: \"1a05a7e5-8c46-441a-9faf-0fd3b41bbd98\") " Feb 27 08:09:50 crc kubenswrapper[4612]: I0227 08:09:50.997001 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/36b549a3-c6e9-4f7e-8cff-85216c8e0e3d-logs\") pod \"36b549a3-c6e9-4f7e-8cff-85216c8e0e3d\" (UID: \"36b549a3-c6e9-4f7e-8cff-85216c8e0e3d\") " Feb 27 08:09:50 crc kubenswrapper[4612]: I0227 08:09:50.997049 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qncrt\" (UniqueName: \"kubernetes.io/projected/36b549a3-c6e9-4f7e-8cff-85216c8e0e3d-kube-api-access-qncrt\") pod \"36b549a3-c6e9-4f7e-8cff-85216c8e0e3d\" (UID: \"36b549a3-c6e9-4f7e-8cff-85216c8e0e3d\") " Feb 27 08:09:50 crc kubenswrapper[4612]: I0227 08:09:50.997102 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/87df2a96-7ae3-47af-92bb-97db038b957c-db-sync-config-data\") pod \"87df2a96-7ae3-47af-92bb-97db038b957c\" (UID: \"87df2a96-7ae3-47af-92bb-97db038b957c\") " Feb 27 08:09:50 crc kubenswrapper[4612]: I0227 08:09:50.997163 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-625qb\" (UniqueName: \"kubernetes.io/projected/87df2a96-7ae3-47af-92bb-97db038b957c-kube-api-access-625qb\") pod \"87df2a96-7ae3-47af-92bb-97db038b957c\" (UID: \"87df2a96-7ae3-47af-92bb-97db038b957c\") " Feb 27 08:09:50 crc kubenswrapper[4612]: I0227 08:09:50.997192 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/1a05a7e5-8c46-441a-9faf-0fd3b41bbd98-credential-keys\") pod \"1a05a7e5-8c46-441a-9faf-0fd3b41bbd98\" (UID: \"1a05a7e5-8c46-441a-9faf-0fd3b41bbd98\") " Feb 27 08:09:50 crc kubenswrapper[4612]: I0227 08:09:50.997232 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/36b549a3-c6e9-4f7e-8cff-85216c8e0e3d-scripts\") pod \"36b549a3-c6e9-4f7e-8cff-85216c8e0e3d\" (UID: \"36b549a3-c6e9-4f7e-8cff-85216c8e0e3d\") " Feb 27 08:09:50 crc kubenswrapper[4612]: I0227 08:09:50.997278 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87df2a96-7ae3-47af-92bb-97db038b957c-combined-ca-bundle\") pod \"87df2a96-7ae3-47af-92bb-97db038b957c\" (UID: \"87df2a96-7ae3-47af-92bb-97db038b957c\") " Feb 27 08:09:50 crc kubenswrapper[4612]: I0227 08:09:50.997318 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1a05a7e5-8c46-441a-9faf-0fd3b41bbd98-scripts\") pod \"1a05a7e5-8c46-441a-9faf-0fd3b41bbd98\" (UID: \"1a05a7e5-8c46-441a-9faf-0fd3b41bbd98\") " Feb 27 08:09:50 crc kubenswrapper[4612]: I0227 08:09:50.997339 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a05a7e5-8c46-441a-9faf-0fd3b41bbd98-combined-ca-bundle\") pod \"1a05a7e5-8c46-441a-9faf-0fd3b41bbd98\" (UID: \"1a05a7e5-8c46-441a-9faf-0fd3b41bbd98\") " Feb 27 08:09:50 crc kubenswrapper[4612]: I0227 08:09:50.997395 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/1a05a7e5-8c46-441a-9faf-0fd3b41bbd98-fernet-keys\") pod \"1a05a7e5-8c46-441a-9faf-0fd3b41bbd98\" (UID: \"1a05a7e5-8c46-441a-9faf-0fd3b41bbd98\") " Feb 27 08:09:50 crc kubenswrapper[4612]: I0227 08:09:50.997416 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/36b549a3-c6e9-4f7e-8cff-85216c8e0e3d-config-data\") pod \"36b549a3-c6e9-4f7e-8cff-85216c8e0e3d\" (UID: \"36b549a3-c6e9-4f7e-8cff-85216c8e0e3d\") " Feb 27 08:09:51 crc kubenswrapper[4612]: I0227 08:09:50.998843 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/36b549a3-c6e9-4f7e-8cff-85216c8e0e3d-logs" (OuterVolumeSpecName: "logs") pod "36b549a3-c6e9-4f7e-8cff-85216c8e0e3d" (UID: "36b549a3-c6e9-4f7e-8cff-85216c8e0e3d"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 08:09:51 crc kubenswrapper[4612]: I0227 08:09:51.000627 4612 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/36b549a3-c6e9-4f7e-8cff-85216c8e0e3d-logs\") on node \"crc\" DevicePath \"\"" Feb 27 08:09:51 crc kubenswrapper[4612]: I0227 08:09:51.008437 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1a05a7e5-8c46-441a-9faf-0fd3b41bbd98-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "1a05a7e5-8c46-441a-9faf-0fd3b41bbd98" (UID: "1a05a7e5-8c46-441a-9faf-0fd3b41bbd98"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:09:51 crc kubenswrapper[4612]: I0227 08:09:51.008652 4612 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Feb 27 08:09:51 crc kubenswrapper[4612]: I0227 08:09:51.037532 4612 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Feb 27 08:09:51 crc kubenswrapper[4612]: I0227 08:09:51.047785 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87df2a96-7ae3-47af-92bb-97db038b957c-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "87df2a96-7ae3-47af-92bb-97db038b957c" (UID: "87df2a96-7ae3-47af-92bb-97db038b957c"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:09:51 crc kubenswrapper[4612]: I0227 08:09:51.047878 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/36b549a3-c6e9-4f7e-8cff-85216c8e0e3d-scripts" (OuterVolumeSpecName: "scripts") pod "36b549a3-c6e9-4f7e-8cff-85216c8e0e3d" (UID: "36b549a3-c6e9-4f7e-8cff-85216c8e0e3d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:09:51 crc kubenswrapper[4612]: I0227 08:09:51.051630 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87df2a96-7ae3-47af-92bb-97db038b957c-kube-api-access-625qb" (OuterVolumeSpecName: "kube-api-access-625qb") pod "87df2a96-7ae3-47af-92bb-97db038b957c" (UID: "87df2a96-7ae3-47af-92bb-97db038b957c"). InnerVolumeSpecName "kube-api-access-625qb". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 08:09:51 crc kubenswrapper[4612]: I0227 08:09:51.053537 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/36b549a3-c6e9-4f7e-8cff-85216c8e0e3d-kube-api-access-qncrt" (OuterVolumeSpecName: "kube-api-access-qncrt") pod "36b549a3-c6e9-4f7e-8cff-85216c8e0e3d" (UID: "36b549a3-c6e9-4f7e-8cff-85216c8e0e3d"). InnerVolumeSpecName "kube-api-access-qncrt". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 08:09:51 crc kubenswrapper[4612]: I0227 08:09:51.054311 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1a05a7e5-8c46-441a-9faf-0fd3b41bbd98-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "1a05a7e5-8c46-441a-9faf-0fd3b41bbd98" (UID: "1a05a7e5-8c46-441a-9faf-0fd3b41bbd98"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:09:51 crc kubenswrapper[4612]: I0227 08:09:51.060961 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1a05a7e5-8c46-441a-9faf-0fd3b41bbd98-scripts" (OuterVolumeSpecName: "scripts") pod "1a05a7e5-8c46-441a-9faf-0fd3b41bbd98" (UID: "1a05a7e5-8c46-441a-9faf-0fd3b41bbd98"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:09:51 crc kubenswrapper[4612]: I0227 08:09:51.081655 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1a05a7e5-8c46-441a-9faf-0fd3b41bbd98-kube-api-access-b6hqv" (OuterVolumeSpecName: "kube-api-access-b6hqv") pod "1a05a7e5-8c46-441a-9faf-0fd3b41bbd98" (UID: "1a05a7e5-8c46-441a-9faf-0fd3b41bbd98"). InnerVolumeSpecName "kube-api-access-b6hqv". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 08:09:51 crc kubenswrapper[4612]: I0227 08:09:51.103156 4612 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/87df2a96-7ae3-47af-92bb-97db038b957c-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Feb 27 08:09:51 crc kubenswrapper[4612]: I0227 08:09:51.103290 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-625qb\" (UniqueName: \"kubernetes.io/projected/87df2a96-7ae3-47af-92bb-97db038b957c-kube-api-access-625qb\") on node \"crc\" DevicePath \"\"" Feb 27 08:09:51 crc kubenswrapper[4612]: I0227 08:09:51.103461 4612 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/1a05a7e5-8c46-441a-9faf-0fd3b41bbd98-credential-keys\") on node \"crc\" DevicePath \"\"" Feb 27 08:09:51 crc kubenswrapper[4612]: I0227 08:09:51.103544 4612 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/36b549a3-c6e9-4f7e-8cff-85216c8e0e3d-scripts\") on node \"crc\" DevicePath \"\"" Feb 27 08:09:51 crc kubenswrapper[4612]: I0227 08:09:51.103616 4612 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1a05a7e5-8c46-441a-9faf-0fd3b41bbd98-scripts\") on node \"crc\" DevicePath \"\"" Feb 27 08:09:51 crc kubenswrapper[4612]: I0227 08:09:51.103998 4612 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/1a05a7e5-8c46-441a-9faf-0fd3b41bbd98-fernet-keys\") on node \"crc\" DevicePath \"\"" Feb 27 08:09:51 crc kubenswrapper[4612]: I0227 08:09:51.104090 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b6hqv\" (UniqueName: \"kubernetes.io/projected/1a05a7e5-8c46-441a-9faf-0fd3b41bbd98-kube-api-access-b6hqv\") on node \"crc\" DevicePath \"\"" Feb 27 08:09:51 crc kubenswrapper[4612]: I0227 08:09:51.104181 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qncrt\" (UniqueName: \"kubernetes.io/projected/36b549a3-c6e9-4f7e-8cff-85216c8e0e3d-kube-api-access-qncrt\") on node \"crc\" DevicePath \"\"" Feb 27 08:09:51 crc kubenswrapper[4612]: I0227 08:09:51.118707 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1a05a7e5-8c46-441a-9faf-0fd3b41bbd98-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1a05a7e5-8c46-441a-9faf-0fd3b41bbd98" (UID: "1a05a7e5-8c46-441a-9faf-0fd3b41bbd98"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:09:51 crc kubenswrapper[4612]: I0227 08:09:51.139790 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1a05a7e5-8c46-441a-9faf-0fd3b41bbd98-config-data" (OuterVolumeSpecName: "config-data") pod "1a05a7e5-8c46-441a-9faf-0fd3b41bbd98" (UID: "1a05a7e5-8c46-441a-9faf-0fd3b41bbd98"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:09:51 crc kubenswrapper[4612]: I0227 08:09:51.146152 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-56df8fb6b7-9p274" Feb 27 08:09:51 crc kubenswrapper[4612]: I0227 08:09:51.153540 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/36b549a3-c6e9-4f7e-8cff-85216c8e0e3d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "36b549a3-c6e9-4f7e-8cff-85216c8e0e3d" (UID: "36b549a3-c6e9-4f7e-8cff-85216c8e0e3d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:09:51 crc kubenswrapper[4612]: I0227 08:09:51.163390 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/36b549a3-c6e9-4f7e-8cff-85216c8e0e3d-config-data" (OuterVolumeSpecName: "config-data") pod "36b549a3-c6e9-4f7e-8cff-85216c8e0e3d" (UID: "36b549a3-c6e9-4f7e-8cff-85216c8e0e3d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:09:51 crc kubenswrapper[4612]: I0227 08:09:51.197329 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87df2a96-7ae3-47af-92bb-97db038b957c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "87df2a96-7ae3-47af-92bb-97db038b957c" (UID: "87df2a96-7ae3-47af-92bb-97db038b957c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:09:51 crc kubenswrapper[4612]: I0227 08:09:51.203776 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56df8fb6b7-9p274" event={"ID":"fd3bd506-aad1-468a-b469-64ad78ec33fe","Type":"ContainerDied","Data":"ed72968c7818e0306d78da7f79912f6d1d2741262a26a0eef8e6cfbaf13ad954"} Feb 27 08:09:51 crc kubenswrapper[4612]: I0227 08:09:51.203822 4612 scope.go:117] "RemoveContainer" containerID="10abcf416e4e5487e724027816bc29df51d2715c2f37c2ad093bd0b055e40712" Feb 27 08:09:51 crc kubenswrapper[4612]: I0227 08:09:51.203971 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-56df8fb6b7-9p274" Feb 27 08:09:51 crc kubenswrapper[4612]: I0227 08:09:51.206549 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fd3bd506-aad1-468a-b469-64ad78ec33fe-ovsdbserver-nb\") pod \"fd3bd506-aad1-468a-b469-64ad78ec33fe\" (UID: \"fd3bd506-aad1-468a-b469-64ad78ec33fe\") " Feb 27 08:09:51 crc kubenswrapper[4612]: I0227 08:09:51.206682 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/fd3bd506-aad1-468a-b469-64ad78ec33fe-dns-swift-storage-0\") pod \"fd3bd506-aad1-468a-b469-64ad78ec33fe\" (UID: \"fd3bd506-aad1-468a-b469-64ad78ec33fe\") " Feb 27 08:09:51 crc kubenswrapper[4612]: I0227 08:09:51.206747 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rm4pj\" (UniqueName: \"kubernetes.io/projected/fd3bd506-aad1-468a-b469-64ad78ec33fe-kube-api-access-rm4pj\") pod \"fd3bd506-aad1-468a-b469-64ad78ec33fe\" (UID: \"fd3bd506-aad1-468a-b469-64ad78ec33fe\") " Feb 27 08:09:51 crc kubenswrapper[4612]: I0227 08:09:51.211879 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fd3bd506-aad1-468a-b469-64ad78ec33fe-dns-svc\") pod \"fd3bd506-aad1-468a-b469-64ad78ec33fe\" (UID: \"fd3bd506-aad1-468a-b469-64ad78ec33fe\") " Feb 27 08:09:51 crc kubenswrapper[4612]: I0227 08:09:51.211946 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/fd3bd506-aad1-468a-b469-64ad78ec33fe-ovsdbserver-sb\") pod \"fd3bd506-aad1-468a-b469-64ad78ec33fe\" (UID: \"fd3bd506-aad1-468a-b469-64ad78ec33fe\") " Feb 27 08:09:51 crc kubenswrapper[4612]: I0227 08:09:51.211987 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fd3bd506-aad1-468a-b469-64ad78ec33fe-config\") pod \"fd3bd506-aad1-468a-b469-64ad78ec33fe\" (UID: \"fd3bd506-aad1-468a-b469-64ad78ec33fe\") " Feb 27 08:09:51 crc kubenswrapper[4612]: I0227 08:09:51.212742 4612 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87df2a96-7ae3-47af-92bb-97db038b957c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 08:09:51 crc kubenswrapper[4612]: I0227 08:09:51.212767 4612 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a05a7e5-8c46-441a-9faf-0fd3b41bbd98-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 08:09:51 crc kubenswrapper[4612]: I0227 08:09:51.212779 4612 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/36b549a3-c6e9-4f7e-8cff-85216c8e0e3d-config-data\") on node \"crc\" DevicePath \"\"" Feb 27 08:09:51 crc kubenswrapper[4612]: I0227 08:09:51.212789 4612 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36b549a3-c6e9-4f7e-8cff-85216c8e0e3d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 08:09:51 crc kubenswrapper[4612]: I0227 08:09:51.212801 4612 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1a05a7e5-8c46-441a-9faf-0fd3b41bbd98-config-data\") on node \"crc\" DevicePath \"\"" Feb 27 08:09:51 crc kubenswrapper[4612]: I0227 08:09:51.231006 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fd3bd506-aad1-468a-b469-64ad78ec33fe-kube-api-access-rm4pj" (OuterVolumeSpecName: "kube-api-access-rm4pj") pod "fd3bd506-aad1-468a-b469-64ad78ec33fe" (UID: "fd3bd506-aad1-468a-b469-64ad78ec33fe"). InnerVolumeSpecName "kube-api-access-rm4pj". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 08:09:51 crc kubenswrapper[4612]: I0227 08:09:51.232668 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-8zrcl" Feb 27 08:09:51 crc kubenswrapper[4612]: I0227 08:09:51.233645 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-8zrcl" event={"ID":"36b549a3-c6e9-4f7e-8cff-85216c8e0e3d","Type":"ContainerDied","Data":"bc5d4192eb9c649d7f8a8e0e4464574f124851dc9893ac58466888063eb6d9b4"} Feb 27 08:09:51 crc kubenswrapper[4612]: I0227 08:09:51.233707 4612 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bc5d4192eb9c649d7f8a8e0e4464574f124851dc9893ac58466888063eb6d9b4" Feb 27 08:09:51 crc kubenswrapper[4612]: I0227 08:09:51.237951 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-r8dpp" Feb 27 08:09:51 crc kubenswrapper[4612]: I0227 08:09:51.241741 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-r8dpp" event={"ID":"87df2a96-7ae3-47af-92bb-97db038b957c","Type":"ContainerDied","Data":"c19fec2225fbc4c95b2593b77eca8d9e14fa64d610e4c9180c77c7d4d60a5b8a"} Feb 27 08:09:51 crc kubenswrapper[4612]: I0227 08:09:51.241798 4612 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c19fec2225fbc4c95b2593b77eca8d9e14fa64d610e4c9180c77c7d4d60a5b8a" Feb 27 08:09:51 crc kubenswrapper[4612]: I0227 08:09:51.243630 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f3bb8dc2-0c27-4a7a-baa1-1bfd8916f920","Type":"ContainerStarted","Data":"061c95d3248b23c6fb1255c379f096d63ba92d34e7ca9cb254e6fffd9b2e9904"} Feb 27 08:09:51 crc kubenswrapper[4612]: I0227 08:09:51.249385 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-s5bjv" event={"ID":"1a05a7e5-8c46-441a-9faf-0fd3b41bbd98","Type":"ContainerDied","Data":"df93dc184990b247c92ef9f46224a41ef8a99e51093c8ececf390e03ff9d0660"} Feb 27 08:09:51 crc kubenswrapper[4612]: I0227 08:09:51.249442 4612 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="df93dc184990b247c92ef9f46224a41ef8a99e51093c8ececf390e03ff9d0660" Feb 27 08:09:51 crc kubenswrapper[4612]: I0227 08:09:51.249521 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-s5bjv" Feb 27 08:09:51 crc kubenswrapper[4612]: I0227 08:09:51.259300 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Feb 27 08:09:51 crc kubenswrapper[4612]: I0227 08:09:51.259735 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Feb 27 08:09:51 crc kubenswrapper[4612]: I0227 08:09:51.288223 4612 scope.go:117] "RemoveContainer" containerID="1d2b35afd6c9fee7411ec5debd6b49bc6753e4f253045d15c54fb9f8d61fd65e" Feb 27 08:09:51 crc kubenswrapper[4612]: I0227 08:09:51.356304 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rm4pj\" (UniqueName: \"kubernetes.io/projected/fd3bd506-aad1-468a-b469-64ad78ec33fe-kube-api-access-rm4pj\") on node \"crc\" DevicePath \"\"" Feb 27 08:09:51 crc kubenswrapper[4612]: I0227 08:09:51.412784 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-9c6fdd6-lvn8p"] Feb 27 08:09:51 crc kubenswrapper[4612]: E0227 08:09:51.413395 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a05a7e5-8c46-441a-9faf-0fd3b41bbd98" containerName="keystone-bootstrap" Feb 27 08:09:51 crc kubenswrapper[4612]: I0227 08:09:51.413473 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a05a7e5-8c46-441a-9faf-0fd3b41bbd98" containerName="keystone-bootstrap" Feb 27 08:09:51 crc kubenswrapper[4612]: E0227 08:09:51.413539 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fd3bd506-aad1-468a-b469-64ad78ec33fe" containerName="init" Feb 27 08:09:51 crc kubenswrapper[4612]: I0227 08:09:51.413595 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="fd3bd506-aad1-468a-b469-64ad78ec33fe" containerName="init" Feb 27 08:09:51 crc kubenswrapper[4612]: E0227 08:09:51.413659 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fd3bd506-aad1-468a-b469-64ad78ec33fe" containerName="dnsmasq-dns" Feb 27 08:09:51 crc kubenswrapper[4612]: I0227 08:09:51.413726 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="fd3bd506-aad1-468a-b469-64ad78ec33fe" containerName="dnsmasq-dns" Feb 27 08:09:51 crc kubenswrapper[4612]: E0227 08:09:51.413792 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="87df2a96-7ae3-47af-92bb-97db038b957c" containerName="barbican-db-sync" Feb 27 08:09:51 crc kubenswrapper[4612]: I0227 08:09:51.413850 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="87df2a96-7ae3-47af-92bb-97db038b957c" containerName="barbican-db-sync" Feb 27 08:09:51 crc kubenswrapper[4612]: E0227 08:09:51.413919 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="36b549a3-c6e9-4f7e-8cff-85216c8e0e3d" containerName="placement-db-sync" Feb 27 08:09:51 crc kubenswrapper[4612]: I0227 08:09:51.414026 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="36b549a3-c6e9-4f7e-8cff-85216c8e0e3d" containerName="placement-db-sync" Feb 27 08:09:51 crc kubenswrapper[4612]: I0227 08:09:51.414488 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="87df2a96-7ae3-47af-92bb-97db038b957c" containerName="barbican-db-sync" Feb 27 08:09:51 crc kubenswrapper[4612]: I0227 08:09:51.414818 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="fd3bd506-aad1-468a-b469-64ad78ec33fe" containerName="dnsmasq-dns" Feb 27 08:09:51 crc kubenswrapper[4612]: I0227 08:09:51.414894 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="1a05a7e5-8c46-441a-9faf-0fd3b41bbd98" containerName="keystone-bootstrap" Feb 27 08:09:51 crc kubenswrapper[4612]: I0227 08:09:51.421238 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="36b549a3-c6e9-4f7e-8cff-85216c8e0e3d" containerName="placement-db-sync" Feb 27 08:09:51 crc kubenswrapper[4612]: I0227 08:09:51.422110 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-9c6fdd6-lvn8p" Feb 27 08:09:51 crc kubenswrapper[4612]: I0227 08:09:51.427451 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Feb 27 08:09:51 crc kubenswrapper[4612]: I0227 08:09:51.428634 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-zdkwc" Feb 27 08:09:51 crc kubenswrapper[4612]: I0227 08:09:51.428957 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Feb 27 08:09:51 crc kubenswrapper[4612]: I0227 08:09:51.429141 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Feb 27 08:09:51 crc kubenswrapper[4612]: I0227 08:09:51.429331 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Feb 27 08:09:51 crc kubenswrapper[4612]: I0227 08:09:51.431874 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Feb 27 08:09:51 crc kubenswrapper[4612]: I0227 08:09:51.455719 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-9c6fdd6-lvn8p"] Feb 27 08:09:51 crc kubenswrapper[4612]: I0227 08:09:51.527449 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fd3bd506-aad1-468a-b469-64ad78ec33fe-config" (OuterVolumeSpecName: "config") pod "fd3bd506-aad1-468a-b469-64ad78ec33fe" (UID: "fd3bd506-aad1-468a-b469-64ad78ec33fe"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 08:09:51 crc kubenswrapper[4612]: I0227 08:09:51.564177 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/71940e05-ecfa-46de-aefe-c213add5362c-combined-ca-bundle\") pod \"keystone-9c6fdd6-lvn8p\" (UID: \"71940e05-ecfa-46de-aefe-c213add5362c\") " pod="openstack/keystone-9c6fdd6-lvn8p" Feb 27 08:09:51 crc kubenswrapper[4612]: I0227 08:09:51.564235 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/71940e05-ecfa-46de-aefe-c213add5362c-config-data\") pod \"keystone-9c6fdd6-lvn8p\" (UID: \"71940e05-ecfa-46de-aefe-c213add5362c\") " pod="openstack/keystone-9c6fdd6-lvn8p" Feb 27 08:09:51 crc kubenswrapper[4612]: I0227 08:09:51.564279 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/71940e05-ecfa-46de-aefe-c213add5362c-scripts\") pod \"keystone-9c6fdd6-lvn8p\" (UID: \"71940e05-ecfa-46de-aefe-c213add5362c\") " pod="openstack/keystone-9c6fdd6-lvn8p" Feb 27 08:09:51 crc kubenswrapper[4612]: I0227 08:09:51.564309 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/71940e05-ecfa-46de-aefe-c213add5362c-credential-keys\") pod \"keystone-9c6fdd6-lvn8p\" (UID: \"71940e05-ecfa-46de-aefe-c213add5362c\") " pod="openstack/keystone-9c6fdd6-lvn8p" Feb 27 08:09:51 crc kubenswrapper[4612]: I0227 08:09:51.564333 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/71940e05-ecfa-46de-aefe-c213add5362c-internal-tls-certs\") pod \"keystone-9c6fdd6-lvn8p\" (UID: \"71940e05-ecfa-46de-aefe-c213add5362c\") " pod="openstack/keystone-9c6fdd6-lvn8p" Feb 27 08:09:51 crc kubenswrapper[4612]: I0227 08:09:51.564355 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bhxjh\" (UniqueName: \"kubernetes.io/projected/71940e05-ecfa-46de-aefe-c213add5362c-kube-api-access-bhxjh\") pod \"keystone-9c6fdd6-lvn8p\" (UID: \"71940e05-ecfa-46de-aefe-c213add5362c\") " pod="openstack/keystone-9c6fdd6-lvn8p" Feb 27 08:09:51 crc kubenswrapper[4612]: I0227 08:09:51.564373 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/71940e05-ecfa-46de-aefe-c213add5362c-public-tls-certs\") pod \"keystone-9c6fdd6-lvn8p\" (UID: \"71940e05-ecfa-46de-aefe-c213add5362c\") " pod="openstack/keystone-9c6fdd6-lvn8p" Feb 27 08:09:51 crc kubenswrapper[4612]: I0227 08:09:51.564402 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/71940e05-ecfa-46de-aefe-c213add5362c-fernet-keys\") pod \"keystone-9c6fdd6-lvn8p\" (UID: \"71940e05-ecfa-46de-aefe-c213add5362c\") " pod="openstack/keystone-9c6fdd6-lvn8p" Feb 27 08:09:51 crc kubenswrapper[4612]: I0227 08:09:51.568967 4612 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fd3bd506-aad1-468a-b469-64ad78ec33fe-config\") on node \"crc\" DevicePath \"\"" Feb 27 08:09:51 crc kubenswrapper[4612]: I0227 08:09:51.575776 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fd3bd506-aad1-468a-b469-64ad78ec33fe-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "fd3bd506-aad1-468a-b469-64ad78ec33fe" (UID: "fd3bd506-aad1-468a-b469-64ad78ec33fe"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 08:09:51 crc kubenswrapper[4612]: I0227 08:09:51.604159 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fd3bd506-aad1-468a-b469-64ad78ec33fe-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "fd3bd506-aad1-468a-b469-64ad78ec33fe" (UID: "fd3bd506-aad1-468a-b469-64ad78ec33fe"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 08:09:51 crc kubenswrapper[4612]: I0227 08:09:51.610054 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fd3bd506-aad1-468a-b469-64ad78ec33fe-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "fd3bd506-aad1-468a-b469-64ad78ec33fe" (UID: "fd3bd506-aad1-468a-b469-64ad78ec33fe"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 08:09:51 crc kubenswrapper[4612]: I0227 08:09:51.611970 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fd3bd506-aad1-468a-b469-64ad78ec33fe-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "fd3bd506-aad1-468a-b469-64ad78ec33fe" (UID: "fd3bd506-aad1-468a-b469-64ad78ec33fe"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 08:09:51 crc kubenswrapper[4612]: I0227 08:09:51.671271 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/71940e05-ecfa-46de-aefe-c213add5362c-scripts\") pod \"keystone-9c6fdd6-lvn8p\" (UID: \"71940e05-ecfa-46de-aefe-c213add5362c\") " pod="openstack/keystone-9c6fdd6-lvn8p" Feb 27 08:09:51 crc kubenswrapper[4612]: I0227 08:09:51.671342 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/71940e05-ecfa-46de-aefe-c213add5362c-credential-keys\") pod \"keystone-9c6fdd6-lvn8p\" (UID: \"71940e05-ecfa-46de-aefe-c213add5362c\") " pod="openstack/keystone-9c6fdd6-lvn8p" Feb 27 08:09:51 crc kubenswrapper[4612]: I0227 08:09:51.671376 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/71940e05-ecfa-46de-aefe-c213add5362c-internal-tls-certs\") pod \"keystone-9c6fdd6-lvn8p\" (UID: \"71940e05-ecfa-46de-aefe-c213add5362c\") " pod="openstack/keystone-9c6fdd6-lvn8p" Feb 27 08:09:51 crc kubenswrapper[4612]: I0227 08:09:51.671417 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bhxjh\" (UniqueName: \"kubernetes.io/projected/71940e05-ecfa-46de-aefe-c213add5362c-kube-api-access-bhxjh\") pod \"keystone-9c6fdd6-lvn8p\" (UID: \"71940e05-ecfa-46de-aefe-c213add5362c\") " pod="openstack/keystone-9c6fdd6-lvn8p" Feb 27 08:09:51 crc kubenswrapper[4612]: I0227 08:09:51.671438 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/71940e05-ecfa-46de-aefe-c213add5362c-public-tls-certs\") pod \"keystone-9c6fdd6-lvn8p\" (UID: \"71940e05-ecfa-46de-aefe-c213add5362c\") " pod="openstack/keystone-9c6fdd6-lvn8p" Feb 27 08:09:51 crc kubenswrapper[4612]: I0227 08:09:51.671481 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/71940e05-ecfa-46de-aefe-c213add5362c-fernet-keys\") pod \"keystone-9c6fdd6-lvn8p\" (UID: \"71940e05-ecfa-46de-aefe-c213add5362c\") " pod="openstack/keystone-9c6fdd6-lvn8p" Feb 27 08:09:51 crc kubenswrapper[4612]: I0227 08:09:51.671520 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/71940e05-ecfa-46de-aefe-c213add5362c-combined-ca-bundle\") pod \"keystone-9c6fdd6-lvn8p\" (UID: \"71940e05-ecfa-46de-aefe-c213add5362c\") " pod="openstack/keystone-9c6fdd6-lvn8p" Feb 27 08:09:51 crc kubenswrapper[4612]: I0227 08:09:51.671566 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/71940e05-ecfa-46de-aefe-c213add5362c-config-data\") pod \"keystone-9c6fdd6-lvn8p\" (UID: \"71940e05-ecfa-46de-aefe-c213add5362c\") " pod="openstack/keystone-9c6fdd6-lvn8p" Feb 27 08:09:51 crc kubenswrapper[4612]: I0227 08:09:51.671644 4612 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/fd3bd506-aad1-468a-b469-64ad78ec33fe-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Feb 27 08:09:51 crc kubenswrapper[4612]: I0227 08:09:51.671656 4612 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fd3bd506-aad1-468a-b469-64ad78ec33fe-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 27 08:09:51 crc kubenswrapper[4612]: I0227 08:09:51.671664 4612 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/fd3bd506-aad1-468a-b469-64ad78ec33fe-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 27 08:09:51 crc kubenswrapper[4612]: I0227 08:09:51.671673 4612 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fd3bd506-aad1-468a-b469-64ad78ec33fe-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 27 08:09:51 crc kubenswrapper[4612]: I0227 08:09:51.677812 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/71940e05-ecfa-46de-aefe-c213add5362c-public-tls-certs\") pod \"keystone-9c6fdd6-lvn8p\" (UID: \"71940e05-ecfa-46de-aefe-c213add5362c\") " pod="openstack/keystone-9c6fdd6-lvn8p" Feb 27 08:09:51 crc kubenswrapper[4612]: I0227 08:09:51.678744 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/71940e05-ecfa-46de-aefe-c213add5362c-credential-keys\") pod \"keystone-9c6fdd6-lvn8p\" (UID: \"71940e05-ecfa-46de-aefe-c213add5362c\") " pod="openstack/keystone-9c6fdd6-lvn8p" Feb 27 08:09:51 crc kubenswrapper[4612]: I0227 08:09:51.680110 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/71940e05-ecfa-46de-aefe-c213add5362c-scripts\") pod \"keystone-9c6fdd6-lvn8p\" (UID: \"71940e05-ecfa-46de-aefe-c213add5362c\") " pod="openstack/keystone-9c6fdd6-lvn8p" Feb 27 08:09:51 crc kubenswrapper[4612]: I0227 08:09:51.680551 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/71940e05-ecfa-46de-aefe-c213add5362c-combined-ca-bundle\") pod \"keystone-9c6fdd6-lvn8p\" (UID: \"71940e05-ecfa-46de-aefe-c213add5362c\") " pod="openstack/keystone-9c6fdd6-lvn8p" Feb 27 08:09:51 crc kubenswrapper[4612]: I0227 08:09:51.680927 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/71940e05-ecfa-46de-aefe-c213add5362c-fernet-keys\") pod \"keystone-9c6fdd6-lvn8p\" (UID: \"71940e05-ecfa-46de-aefe-c213add5362c\") " pod="openstack/keystone-9c6fdd6-lvn8p" Feb 27 08:09:51 crc kubenswrapper[4612]: I0227 08:09:51.681801 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/71940e05-ecfa-46de-aefe-c213add5362c-config-data\") pod \"keystone-9c6fdd6-lvn8p\" (UID: \"71940e05-ecfa-46de-aefe-c213add5362c\") " pod="openstack/keystone-9c6fdd6-lvn8p" Feb 27 08:09:51 crc kubenswrapper[4612]: I0227 08:09:51.684189 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/71940e05-ecfa-46de-aefe-c213add5362c-internal-tls-certs\") pod \"keystone-9c6fdd6-lvn8p\" (UID: \"71940e05-ecfa-46de-aefe-c213add5362c\") " pod="openstack/keystone-9c6fdd6-lvn8p" Feb 27 08:09:51 crc kubenswrapper[4612]: I0227 08:09:51.689474 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bhxjh\" (UniqueName: \"kubernetes.io/projected/71940e05-ecfa-46de-aefe-c213add5362c-kube-api-access-bhxjh\") pod \"keystone-9c6fdd6-lvn8p\" (UID: \"71940e05-ecfa-46de-aefe-c213add5362c\") " pod="openstack/keystone-9c6fdd6-lvn8p" Feb 27 08:09:51 crc kubenswrapper[4612]: I0227 08:09:51.745130 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-9c6fdd6-lvn8p" Feb 27 08:09:51 crc kubenswrapper[4612]: I0227 08:09:51.889773 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-56df8fb6b7-9p274"] Feb 27 08:09:51 crc kubenswrapper[4612]: I0227 08:09:51.913154 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-56df8fb6b7-9p274"] Feb 27 08:09:52 crc kubenswrapper[4612]: I0227 08:09:52.132185 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-7558bddf6c-j9f2z"] Feb 27 08:09:52 crc kubenswrapper[4612]: I0227 08:09:52.134012 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-7558bddf6c-j9f2z" Feb 27 08:09:52 crc kubenswrapper[4612]: I0227 08:09:52.139082 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Feb 27 08:09:52 crc kubenswrapper[4612]: I0227 08:09:52.139265 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-msh9n" Feb 27 08:09:52 crc kubenswrapper[4612]: I0227 08:09:52.139594 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Feb 27 08:09:52 crc kubenswrapper[4612]: I0227 08:09:52.148487 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-745b984886-tzpzt"] Feb 27 08:09:52 crc kubenswrapper[4612]: I0227 08:09:52.150029 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-745b984886-tzpzt" Feb 27 08:09:52 crc kubenswrapper[4612]: I0227 08:09:52.151765 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Feb 27 08:09:52 crc kubenswrapper[4612]: I0227 08:09:52.166452 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-7558bddf6c-j9f2z"] Feb 27 08:09:52 crc kubenswrapper[4612]: I0227 08:09:52.175594 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-745b984886-tzpzt"] Feb 27 08:09:52 crc kubenswrapper[4612]: I0227 08:09:52.270670 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-9c6fdd6-lvn8p"] Feb 27 08:09:52 crc kubenswrapper[4612]: I0227 08:09:52.306214 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/74bec848-d5b3-43d0-97d7-eda229642bb5-config-data-custom\") pod \"barbican-worker-7558bddf6c-j9f2z\" (UID: \"74bec848-d5b3-43d0-97d7-eda229642bb5\") " pod="openstack/barbican-worker-7558bddf6c-j9f2z" Feb 27 08:09:52 crc kubenswrapper[4612]: I0227 08:09:52.308132 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/74bec848-d5b3-43d0-97d7-eda229642bb5-config-data\") pod \"barbican-worker-7558bddf6c-j9f2z\" (UID: \"74bec848-d5b3-43d0-97d7-eda229642bb5\") " pod="openstack/barbican-worker-7558bddf6c-j9f2z" Feb 27 08:09:52 crc kubenswrapper[4612]: I0227 08:09:52.308346 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6tctm\" (UniqueName: \"kubernetes.io/projected/74bec848-d5b3-43d0-97d7-eda229642bb5-kube-api-access-6tctm\") pod \"barbican-worker-7558bddf6c-j9f2z\" (UID: \"74bec848-d5b3-43d0-97d7-eda229642bb5\") " pod="openstack/barbican-worker-7558bddf6c-j9f2z" Feb 27 08:09:52 crc kubenswrapper[4612]: I0227 08:09:52.308439 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e2579b13-034e-474a-bbe6-2062a710f380-combined-ca-bundle\") pod \"barbican-keystone-listener-745b984886-tzpzt\" (UID: \"e2579b13-034e-474a-bbe6-2062a710f380\") " pod="openstack/barbican-keystone-listener-745b984886-tzpzt" Feb 27 08:09:52 crc kubenswrapper[4612]: I0227 08:09:52.308543 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e2579b13-034e-474a-bbe6-2062a710f380-config-data-custom\") pod \"barbican-keystone-listener-745b984886-tzpzt\" (UID: \"e2579b13-034e-474a-bbe6-2062a710f380\") " pod="openstack/barbican-keystone-listener-745b984886-tzpzt" Feb 27 08:09:52 crc kubenswrapper[4612]: I0227 08:09:52.308619 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e2579b13-034e-474a-bbe6-2062a710f380-logs\") pod \"barbican-keystone-listener-745b984886-tzpzt\" (UID: \"e2579b13-034e-474a-bbe6-2062a710f380\") " pod="openstack/barbican-keystone-listener-745b984886-tzpzt" Feb 27 08:09:52 crc kubenswrapper[4612]: I0227 08:09:52.308695 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/74bec848-d5b3-43d0-97d7-eda229642bb5-combined-ca-bundle\") pod \"barbican-worker-7558bddf6c-j9f2z\" (UID: \"74bec848-d5b3-43d0-97d7-eda229642bb5\") " pod="openstack/barbican-worker-7558bddf6c-j9f2z" Feb 27 08:09:52 crc kubenswrapper[4612]: I0227 08:09:52.308815 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/74bec848-d5b3-43d0-97d7-eda229642bb5-logs\") pod \"barbican-worker-7558bddf6c-j9f2z\" (UID: \"74bec848-d5b3-43d0-97d7-eda229642bb5\") " pod="openstack/barbican-worker-7558bddf6c-j9f2z" Feb 27 08:09:52 crc kubenswrapper[4612]: I0227 08:09:52.308882 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e2579b13-034e-474a-bbe6-2062a710f380-config-data\") pod \"barbican-keystone-listener-745b984886-tzpzt\" (UID: \"e2579b13-034e-474a-bbe6-2062a710f380\") " pod="openstack/barbican-keystone-listener-745b984886-tzpzt" Feb 27 08:09:52 crc kubenswrapper[4612]: I0227 08:09:52.308949 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gxk6s\" (UniqueName: \"kubernetes.io/projected/e2579b13-034e-474a-bbe6-2062a710f380-kube-api-access-gxk6s\") pod \"barbican-keystone-listener-745b984886-tzpzt\" (UID: \"e2579b13-034e-474a-bbe6-2062a710f380\") " pod="openstack/barbican-keystone-listener-745b984886-tzpzt" Feb 27 08:09:52 crc kubenswrapper[4612]: I0227 08:09:52.306659 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-848cf88cfc-tz267"] Feb 27 08:09:52 crc kubenswrapper[4612]: I0227 08:09:52.319278 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-848cf88cfc-tz267" Feb 27 08:09:52 crc kubenswrapper[4612]: I0227 08:09:52.361754 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-848cf88cfc-tz267"] Feb 27 08:09:52 crc kubenswrapper[4612]: I0227 08:09:52.406512 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-7d477cc4b6-5qtkx"] Feb 27 08:09:52 crc kubenswrapper[4612]: I0227 08:09:52.408756 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-7d477cc4b6-5qtkx" Feb 27 08:09:52 crc kubenswrapper[4612]: I0227 08:09:52.410256 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6tctm\" (UniqueName: \"kubernetes.io/projected/74bec848-d5b3-43d0-97d7-eda229642bb5-kube-api-access-6tctm\") pod \"barbican-worker-7558bddf6c-j9f2z\" (UID: \"74bec848-d5b3-43d0-97d7-eda229642bb5\") " pod="openstack/barbican-worker-7558bddf6c-j9f2z" Feb 27 08:09:52 crc kubenswrapper[4612]: I0227 08:09:52.410536 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e2579b13-034e-474a-bbe6-2062a710f380-combined-ca-bundle\") pod \"barbican-keystone-listener-745b984886-tzpzt\" (UID: \"e2579b13-034e-474a-bbe6-2062a710f380\") " pod="openstack/barbican-keystone-listener-745b984886-tzpzt" Feb 27 08:09:52 crc kubenswrapper[4612]: I0227 08:09:52.411464 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/61899b16-fb85-4ba6-b057-65ad130ea015-dns-svc\") pod \"dnsmasq-dns-848cf88cfc-tz267\" (UID: \"61899b16-fb85-4ba6-b057-65ad130ea015\") " pod="openstack/dnsmasq-dns-848cf88cfc-tz267" Feb 27 08:09:52 crc kubenswrapper[4612]: I0227 08:09:52.411556 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/61899b16-fb85-4ba6-b057-65ad130ea015-ovsdbserver-sb\") pod \"dnsmasq-dns-848cf88cfc-tz267\" (UID: \"61899b16-fb85-4ba6-b057-65ad130ea015\") " pod="openstack/dnsmasq-dns-848cf88cfc-tz267" Feb 27 08:09:52 crc kubenswrapper[4612]: I0227 08:09:52.411640 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/61899b16-fb85-4ba6-b057-65ad130ea015-ovsdbserver-nb\") pod \"dnsmasq-dns-848cf88cfc-tz267\" (UID: \"61899b16-fb85-4ba6-b057-65ad130ea015\") " pod="openstack/dnsmasq-dns-848cf88cfc-tz267" Feb 27 08:09:52 crc kubenswrapper[4612]: I0227 08:09:52.411736 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e2579b13-034e-474a-bbe6-2062a710f380-config-data-custom\") pod \"barbican-keystone-listener-745b984886-tzpzt\" (UID: \"e2579b13-034e-474a-bbe6-2062a710f380\") " pod="openstack/barbican-keystone-listener-745b984886-tzpzt" Feb 27 08:09:52 crc kubenswrapper[4612]: I0227 08:09:52.412292 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e2579b13-034e-474a-bbe6-2062a710f380-logs\") pod \"barbican-keystone-listener-745b984886-tzpzt\" (UID: \"e2579b13-034e-474a-bbe6-2062a710f380\") " pod="openstack/barbican-keystone-listener-745b984886-tzpzt" Feb 27 08:09:52 crc kubenswrapper[4612]: I0227 08:09:52.412380 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/61899b16-fb85-4ba6-b057-65ad130ea015-dns-swift-storage-0\") pod \"dnsmasq-dns-848cf88cfc-tz267\" (UID: \"61899b16-fb85-4ba6-b057-65ad130ea015\") " pod="openstack/dnsmasq-dns-848cf88cfc-tz267" Feb 27 08:09:52 crc kubenswrapper[4612]: I0227 08:09:52.412449 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/74bec848-d5b3-43d0-97d7-eda229642bb5-combined-ca-bundle\") pod \"barbican-worker-7558bddf6c-j9f2z\" (UID: \"74bec848-d5b3-43d0-97d7-eda229642bb5\") " pod="openstack/barbican-worker-7558bddf6c-j9f2z" Feb 27 08:09:52 crc kubenswrapper[4612]: I0227 08:09:52.412532 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/74bec848-d5b3-43d0-97d7-eda229642bb5-logs\") pod \"barbican-worker-7558bddf6c-j9f2z\" (UID: \"74bec848-d5b3-43d0-97d7-eda229642bb5\") " pod="openstack/barbican-worker-7558bddf6c-j9f2z" Feb 27 08:09:52 crc kubenswrapper[4612]: I0227 08:09:52.412595 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e2579b13-034e-474a-bbe6-2062a710f380-config-data\") pod \"barbican-keystone-listener-745b984886-tzpzt\" (UID: \"e2579b13-034e-474a-bbe6-2062a710f380\") " pod="openstack/barbican-keystone-listener-745b984886-tzpzt" Feb 27 08:09:52 crc kubenswrapper[4612]: I0227 08:09:52.412665 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gxk6s\" (UniqueName: \"kubernetes.io/projected/e2579b13-034e-474a-bbe6-2062a710f380-kube-api-access-gxk6s\") pod \"barbican-keystone-listener-745b984886-tzpzt\" (UID: \"e2579b13-034e-474a-bbe6-2062a710f380\") " pod="openstack/barbican-keystone-listener-745b984886-tzpzt" Feb 27 08:09:52 crc kubenswrapper[4612]: I0227 08:09:52.412801 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/61899b16-fb85-4ba6-b057-65ad130ea015-config\") pod \"dnsmasq-dns-848cf88cfc-tz267\" (UID: \"61899b16-fb85-4ba6-b057-65ad130ea015\") " pod="openstack/dnsmasq-dns-848cf88cfc-tz267" Feb 27 08:09:52 crc kubenswrapper[4612]: I0227 08:09:52.412891 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qp9jw\" (UniqueName: \"kubernetes.io/projected/61899b16-fb85-4ba6-b057-65ad130ea015-kube-api-access-qp9jw\") pod \"dnsmasq-dns-848cf88cfc-tz267\" (UID: \"61899b16-fb85-4ba6-b057-65ad130ea015\") " pod="openstack/dnsmasq-dns-848cf88cfc-tz267" Feb 27 08:09:52 crc kubenswrapper[4612]: I0227 08:09:52.413223 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/74bec848-d5b3-43d0-97d7-eda229642bb5-config-data-custom\") pod \"barbican-worker-7558bddf6c-j9f2z\" (UID: \"74bec848-d5b3-43d0-97d7-eda229642bb5\") " pod="openstack/barbican-worker-7558bddf6c-j9f2z" Feb 27 08:09:52 crc kubenswrapper[4612]: I0227 08:09:52.413308 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/74bec848-d5b3-43d0-97d7-eda229642bb5-config-data\") pod \"barbican-worker-7558bddf6c-j9f2z\" (UID: \"74bec848-d5b3-43d0-97d7-eda229642bb5\") " pod="openstack/barbican-worker-7558bddf6c-j9f2z" Feb 27 08:09:52 crc kubenswrapper[4612]: I0227 08:09:52.419552 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Feb 27 08:09:52 crc kubenswrapper[4612]: I0227 08:09:52.419783 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Feb 27 08:09:52 crc kubenswrapper[4612]: I0227 08:09:52.419962 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Feb 27 08:09:52 crc kubenswrapper[4612]: I0227 08:09:52.420460 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Feb 27 08:09:52 crc kubenswrapper[4612]: I0227 08:09:52.425245 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-n7qd9" Feb 27 08:09:52 crc kubenswrapper[4612]: I0227 08:09:52.425899 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e2579b13-034e-474a-bbe6-2062a710f380-logs\") pod \"barbican-keystone-listener-745b984886-tzpzt\" (UID: \"e2579b13-034e-474a-bbe6-2062a710f380\") " pod="openstack/barbican-keystone-listener-745b984886-tzpzt" Feb 27 08:09:52 crc kubenswrapper[4612]: I0227 08:09:52.426121 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/74bec848-d5b3-43d0-97d7-eda229642bb5-logs\") pod \"barbican-worker-7558bddf6c-j9f2z\" (UID: \"74bec848-d5b3-43d0-97d7-eda229642bb5\") " pod="openstack/barbican-worker-7558bddf6c-j9f2z" Feb 27 08:09:52 crc kubenswrapper[4612]: I0227 08:09:52.427349 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/74bec848-d5b3-43d0-97d7-eda229642bb5-config-data\") pod \"barbican-worker-7558bddf6c-j9f2z\" (UID: \"74bec848-d5b3-43d0-97d7-eda229642bb5\") " pod="openstack/barbican-worker-7558bddf6c-j9f2z" Feb 27 08:09:52 crc kubenswrapper[4612]: I0227 08:09:52.432251 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e2579b13-034e-474a-bbe6-2062a710f380-config-data-custom\") pod \"barbican-keystone-listener-745b984886-tzpzt\" (UID: \"e2579b13-034e-474a-bbe6-2062a710f380\") " pod="openstack/barbican-keystone-listener-745b984886-tzpzt" Feb 27 08:09:52 crc kubenswrapper[4612]: I0227 08:09:52.433032 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/74bec848-d5b3-43d0-97d7-eda229642bb5-config-data-custom\") pod \"barbican-worker-7558bddf6c-j9f2z\" (UID: \"74bec848-d5b3-43d0-97d7-eda229642bb5\") " pod="openstack/barbican-worker-7558bddf6c-j9f2z" Feb 27 08:09:52 crc kubenswrapper[4612]: I0227 08:09:52.440445 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e2579b13-034e-474a-bbe6-2062a710f380-config-data\") pod \"barbican-keystone-listener-745b984886-tzpzt\" (UID: \"e2579b13-034e-474a-bbe6-2062a710f380\") " pod="openstack/barbican-keystone-listener-745b984886-tzpzt" Feb 27 08:09:52 crc kubenswrapper[4612]: I0227 08:09:52.447276 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e2579b13-034e-474a-bbe6-2062a710f380-combined-ca-bundle\") pod \"barbican-keystone-listener-745b984886-tzpzt\" (UID: \"e2579b13-034e-474a-bbe6-2062a710f380\") " pod="openstack/barbican-keystone-listener-745b984886-tzpzt" Feb 27 08:09:52 crc kubenswrapper[4612]: I0227 08:09:52.447393 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9ab32d5a-a797-47a8-961a-a534f55fdfa7","Type":"ContainerStarted","Data":"e77b56aa24d4104786d48b89f361df289c7f02205d97f075d23ee055ec70f239"} Feb 27 08:09:52 crc kubenswrapper[4612]: I0227 08:09:52.466858 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/74bec848-d5b3-43d0-97d7-eda229642bb5-combined-ca-bundle\") pod \"barbican-worker-7558bddf6c-j9f2z\" (UID: \"74bec848-d5b3-43d0-97d7-eda229642bb5\") " pod="openstack/barbican-worker-7558bddf6c-j9f2z" Feb 27 08:09:52 crc kubenswrapper[4612]: I0227 08:09:52.467172 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6tctm\" (UniqueName: \"kubernetes.io/projected/74bec848-d5b3-43d0-97d7-eda229642bb5-kube-api-access-6tctm\") pod \"barbican-worker-7558bddf6c-j9f2z\" (UID: \"74bec848-d5b3-43d0-97d7-eda229642bb5\") " pod="openstack/barbican-worker-7558bddf6c-j9f2z" Feb 27 08:09:52 crc kubenswrapper[4612]: I0227 08:09:52.493563 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gxk6s\" (UniqueName: \"kubernetes.io/projected/e2579b13-034e-474a-bbe6-2062a710f380-kube-api-access-gxk6s\") pod \"barbican-keystone-listener-745b984886-tzpzt\" (UID: \"e2579b13-034e-474a-bbe6-2062a710f380\") " pod="openstack/barbican-keystone-listener-745b984886-tzpzt" Feb 27 08:09:52 crc kubenswrapper[4612]: I0227 08:09:52.504992 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-7d477cc4b6-5qtkx"] Feb 27 08:09:52 crc kubenswrapper[4612]: I0227 08:09:52.505234 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f3bb8dc2-0c27-4a7a-baa1-1bfd8916f920","Type":"ContainerStarted","Data":"5238c7525d8aa4bdbb238a8f27a1c89df731af309999f71d7cc75733c681e871"} Feb 27 08:09:52 crc kubenswrapper[4612]: I0227 08:09:52.514240 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5fb13c78-0dab-4785-90bc-21d5dabe81f6-combined-ca-bundle\") pod \"placement-7d477cc4b6-5qtkx\" (UID: \"5fb13c78-0dab-4785-90bc-21d5dabe81f6\") " pod="openstack/placement-7d477cc4b6-5qtkx" Feb 27 08:09:52 crc kubenswrapper[4612]: I0227 08:09:52.521192 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5fb13c78-0dab-4785-90bc-21d5dabe81f6-public-tls-certs\") pod \"placement-7d477cc4b6-5qtkx\" (UID: \"5fb13c78-0dab-4785-90bc-21d5dabe81f6\") " pod="openstack/placement-7d477cc4b6-5qtkx" Feb 27 08:09:52 crc kubenswrapper[4612]: I0227 08:09:52.521349 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/61899b16-fb85-4ba6-b057-65ad130ea015-dns-svc\") pod \"dnsmasq-dns-848cf88cfc-tz267\" (UID: \"61899b16-fb85-4ba6-b057-65ad130ea015\") " pod="openstack/dnsmasq-dns-848cf88cfc-tz267" Feb 27 08:09:52 crc kubenswrapper[4612]: I0227 08:09:52.521430 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/61899b16-fb85-4ba6-b057-65ad130ea015-ovsdbserver-sb\") pod \"dnsmasq-dns-848cf88cfc-tz267\" (UID: \"61899b16-fb85-4ba6-b057-65ad130ea015\") " pod="openstack/dnsmasq-dns-848cf88cfc-tz267" Feb 27 08:09:52 crc kubenswrapper[4612]: I0227 08:09:52.521535 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/61899b16-fb85-4ba6-b057-65ad130ea015-ovsdbserver-nb\") pod \"dnsmasq-dns-848cf88cfc-tz267\" (UID: \"61899b16-fb85-4ba6-b057-65ad130ea015\") " pod="openstack/dnsmasq-dns-848cf88cfc-tz267" Feb 27 08:09:52 crc kubenswrapper[4612]: I0227 08:09:52.521624 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/61899b16-fb85-4ba6-b057-65ad130ea015-dns-swift-storage-0\") pod \"dnsmasq-dns-848cf88cfc-tz267\" (UID: \"61899b16-fb85-4ba6-b057-65ad130ea015\") " pod="openstack/dnsmasq-dns-848cf88cfc-tz267" Feb 27 08:09:52 crc kubenswrapper[4612]: I0227 08:09:52.521785 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/61899b16-fb85-4ba6-b057-65ad130ea015-config\") pod \"dnsmasq-dns-848cf88cfc-tz267\" (UID: \"61899b16-fb85-4ba6-b057-65ad130ea015\") " pod="openstack/dnsmasq-dns-848cf88cfc-tz267" Feb 27 08:09:52 crc kubenswrapper[4612]: I0227 08:09:52.552211 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qp9jw\" (UniqueName: \"kubernetes.io/projected/61899b16-fb85-4ba6-b057-65ad130ea015-kube-api-access-qp9jw\") pod \"dnsmasq-dns-848cf88cfc-tz267\" (UID: \"61899b16-fb85-4ba6-b057-65ad130ea015\") " pod="openstack/dnsmasq-dns-848cf88cfc-tz267" Feb 27 08:09:52 crc kubenswrapper[4612]: I0227 08:09:52.553509 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5fb13c78-0dab-4785-90bc-21d5dabe81f6-scripts\") pod \"placement-7d477cc4b6-5qtkx\" (UID: \"5fb13c78-0dab-4785-90bc-21d5dabe81f6\") " pod="openstack/placement-7d477cc4b6-5qtkx" Feb 27 08:09:52 crc kubenswrapper[4612]: I0227 08:09:52.553656 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5fb13c78-0dab-4785-90bc-21d5dabe81f6-logs\") pod \"placement-7d477cc4b6-5qtkx\" (UID: \"5fb13c78-0dab-4785-90bc-21d5dabe81f6\") " pod="openstack/placement-7d477cc4b6-5qtkx" Feb 27 08:09:52 crc kubenswrapper[4612]: I0227 08:09:52.553855 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-htnt9\" (UniqueName: \"kubernetes.io/projected/5fb13c78-0dab-4785-90bc-21d5dabe81f6-kube-api-access-htnt9\") pod \"placement-7d477cc4b6-5qtkx\" (UID: \"5fb13c78-0dab-4785-90bc-21d5dabe81f6\") " pod="openstack/placement-7d477cc4b6-5qtkx" Feb 27 08:09:52 crc kubenswrapper[4612]: I0227 08:09:52.553983 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5fb13c78-0dab-4785-90bc-21d5dabe81f6-config-data\") pod \"placement-7d477cc4b6-5qtkx\" (UID: \"5fb13c78-0dab-4785-90bc-21d5dabe81f6\") " pod="openstack/placement-7d477cc4b6-5qtkx" Feb 27 08:09:52 crc kubenswrapper[4612]: I0227 08:09:52.555305 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5fb13c78-0dab-4785-90bc-21d5dabe81f6-internal-tls-certs\") pod \"placement-7d477cc4b6-5qtkx\" (UID: \"5fb13c78-0dab-4785-90bc-21d5dabe81f6\") " pod="openstack/placement-7d477cc4b6-5qtkx" Feb 27 08:09:52 crc kubenswrapper[4612]: I0227 08:09:52.536213 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/61899b16-fb85-4ba6-b057-65ad130ea015-ovsdbserver-sb\") pod \"dnsmasq-dns-848cf88cfc-tz267\" (UID: \"61899b16-fb85-4ba6-b057-65ad130ea015\") " pod="openstack/dnsmasq-dns-848cf88cfc-tz267" Feb 27 08:09:52 crc kubenswrapper[4612]: I0227 08:09:52.543528 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/61899b16-fb85-4ba6-b057-65ad130ea015-ovsdbserver-nb\") pod \"dnsmasq-dns-848cf88cfc-tz267\" (UID: \"61899b16-fb85-4ba6-b057-65ad130ea015\") " pod="openstack/dnsmasq-dns-848cf88cfc-tz267" Feb 27 08:09:52 crc kubenswrapper[4612]: I0227 08:09:52.544054 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/61899b16-fb85-4ba6-b057-65ad130ea015-dns-swift-storage-0\") pod \"dnsmasq-dns-848cf88cfc-tz267\" (UID: \"61899b16-fb85-4ba6-b057-65ad130ea015\") " pod="openstack/dnsmasq-dns-848cf88cfc-tz267" Feb 27 08:09:52 crc kubenswrapper[4612]: I0227 08:09:52.544538 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/61899b16-fb85-4ba6-b057-65ad130ea015-config\") pod \"dnsmasq-dns-848cf88cfc-tz267\" (UID: \"61899b16-fb85-4ba6-b057-65ad130ea015\") " pod="openstack/dnsmasq-dns-848cf88cfc-tz267" Feb 27 08:09:52 crc kubenswrapper[4612]: I0227 08:09:52.557150 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/61899b16-fb85-4ba6-b057-65ad130ea015-dns-svc\") pod \"dnsmasq-dns-848cf88cfc-tz267\" (UID: \"61899b16-fb85-4ba6-b057-65ad130ea015\") " pod="openstack/dnsmasq-dns-848cf88cfc-tz267" Feb 27 08:09:52 crc kubenswrapper[4612]: I0227 08:09:52.576792 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qp9jw\" (UniqueName: \"kubernetes.io/projected/61899b16-fb85-4ba6-b057-65ad130ea015-kube-api-access-qp9jw\") pod \"dnsmasq-dns-848cf88cfc-tz267\" (UID: \"61899b16-fb85-4ba6-b057-65ad130ea015\") " pod="openstack/dnsmasq-dns-848cf88cfc-tz267" Feb 27 08:09:52 crc kubenswrapper[4612]: I0227 08:09:52.577019 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-55cd89c446-4c5hx"] Feb 27 08:09:52 crc kubenswrapper[4612]: I0227 08:09:52.585746 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-55cd89c446-4c5hx" Feb 27 08:09:52 crc kubenswrapper[4612]: I0227 08:09:52.599678 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Feb 27 08:09:52 crc kubenswrapper[4612]: I0227 08:09:52.628001 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-55cd89c446-4c5hx"] Feb 27 08:09:52 crc kubenswrapper[4612]: I0227 08:09:52.659093 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-htnt9\" (UniqueName: \"kubernetes.io/projected/5fb13c78-0dab-4785-90bc-21d5dabe81f6-kube-api-access-htnt9\") pod \"placement-7d477cc4b6-5qtkx\" (UID: \"5fb13c78-0dab-4785-90bc-21d5dabe81f6\") " pod="openstack/placement-7d477cc4b6-5qtkx" Feb 27 08:09:52 crc kubenswrapper[4612]: I0227 08:09:52.662913 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5fb13c78-0dab-4785-90bc-21d5dabe81f6-config-data\") pod \"placement-7d477cc4b6-5qtkx\" (UID: \"5fb13c78-0dab-4785-90bc-21d5dabe81f6\") " pod="openstack/placement-7d477cc4b6-5qtkx" Feb 27 08:09:52 crc kubenswrapper[4612]: I0227 08:09:52.664732 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5fb13c78-0dab-4785-90bc-21d5dabe81f6-internal-tls-certs\") pod \"placement-7d477cc4b6-5qtkx\" (UID: \"5fb13c78-0dab-4785-90bc-21d5dabe81f6\") " pod="openstack/placement-7d477cc4b6-5qtkx" Feb 27 08:09:52 crc kubenswrapper[4612]: I0227 08:09:52.664983 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5fb13c78-0dab-4785-90bc-21d5dabe81f6-combined-ca-bundle\") pod \"placement-7d477cc4b6-5qtkx\" (UID: \"5fb13c78-0dab-4785-90bc-21d5dabe81f6\") " pod="openstack/placement-7d477cc4b6-5qtkx" Feb 27 08:09:52 crc kubenswrapper[4612]: I0227 08:09:52.665226 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5fb13c78-0dab-4785-90bc-21d5dabe81f6-public-tls-certs\") pod \"placement-7d477cc4b6-5qtkx\" (UID: \"5fb13c78-0dab-4785-90bc-21d5dabe81f6\") " pod="openstack/placement-7d477cc4b6-5qtkx" Feb 27 08:09:52 crc kubenswrapper[4612]: I0227 08:09:52.677052 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5fb13c78-0dab-4785-90bc-21d5dabe81f6-scripts\") pod \"placement-7d477cc4b6-5qtkx\" (UID: \"5fb13c78-0dab-4785-90bc-21d5dabe81f6\") " pod="openstack/placement-7d477cc4b6-5qtkx" Feb 27 08:09:52 crc kubenswrapper[4612]: I0227 08:09:52.721819 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5fb13c78-0dab-4785-90bc-21d5dabe81f6-logs\") pod \"placement-7d477cc4b6-5qtkx\" (UID: \"5fb13c78-0dab-4785-90bc-21d5dabe81f6\") " pod="openstack/placement-7d477cc4b6-5qtkx" Feb 27 08:09:52 crc kubenswrapper[4612]: I0227 08:09:52.695632 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5fb13c78-0dab-4785-90bc-21d5dabe81f6-combined-ca-bundle\") pod \"placement-7d477cc4b6-5qtkx\" (UID: \"5fb13c78-0dab-4785-90bc-21d5dabe81f6\") " pod="openstack/placement-7d477cc4b6-5qtkx" Feb 27 08:09:52 crc kubenswrapper[4612]: I0227 08:09:52.696078 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-htnt9\" (UniqueName: \"kubernetes.io/projected/5fb13c78-0dab-4785-90bc-21d5dabe81f6-kube-api-access-htnt9\") pod \"placement-7d477cc4b6-5qtkx\" (UID: \"5fb13c78-0dab-4785-90bc-21d5dabe81f6\") " pod="openstack/placement-7d477cc4b6-5qtkx" Feb 27 08:09:52 crc kubenswrapper[4612]: I0227 08:09:52.697310 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5fb13c78-0dab-4785-90bc-21d5dabe81f6-public-tls-certs\") pod \"placement-7d477cc4b6-5qtkx\" (UID: \"5fb13c78-0dab-4785-90bc-21d5dabe81f6\") " pod="openstack/placement-7d477cc4b6-5qtkx" Feb 27 08:09:52 crc kubenswrapper[4612]: I0227 08:09:52.718228 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5fb13c78-0dab-4785-90bc-21d5dabe81f6-scripts\") pod \"placement-7d477cc4b6-5qtkx\" (UID: \"5fb13c78-0dab-4785-90bc-21d5dabe81f6\") " pod="openstack/placement-7d477cc4b6-5qtkx" Feb 27 08:09:52 crc kubenswrapper[4612]: I0227 08:09:52.681904 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5fb13c78-0dab-4785-90bc-21d5dabe81f6-config-data\") pod \"placement-7d477cc4b6-5qtkx\" (UID: \"5fb13c78-0dab-4785-90bc-21d5dabe81f6\") " pod="openstack/placement-7d477cc4b6-5qtkx" Feb 27 08:09:52 crc kubenswrapper[4612]: I0227 08:09:52.722326 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5fb13c78-0dab-4785-90bc-21d5dabe81f6-logs\") pod \"placement-7d477cc4b6-5qtkx\" (UID: \"5fb13c78-0dab-4785-90bc-21d5dabe81f6\") " pod="openstack/placement-7d477cc4b6-5qtkx" Feb 27 08:09:52 crc kubenswrapper[4612]: I0227 08:09:52.750732 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5fb13c78-0dab-4785-90bc-21d5dabe81f6-internal-tls-certs\") pod \"placement-7d477cc4b6-5qtkx\" (UID: \"5fb13c78-0dab-4785-90bc-21d5dabe81f6\") " pod="openstack/placement-7d477cc4b6-5qtkx" Feb 27 08:09:52 crc kubenswrapper[4612]: I0227 08:09:52.780177 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-7558bddf6c-j9f2z" Feb 27 08:09:52 crc kubenswrapper[4612]: I0227 08:09:52.798525 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-745b984886-tzpzt" Feb 27 08:09:52 crc kubenswrapper[4612]: I0227 08:09:52.811538 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-848cf88cfc-tz267" Feb 27 08:09:52 crc kubenswrapper[4612]: I0227 08:09:52.824894 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb25e0d6-95cd-42a8-a182-f714a14d0084-combined-ca-bundle\") pod \"barbican-api-55cd89c446-4c5hx\" (UID: \"bb25e0d6-95cd-42a8-a182-f714a14d0084\") " pod="openstack/barbican-api-55cd89c446-4c5hx" Feb 27 08:09:52 crc kubenswrapper[4612]: I0227 08:09:52.825009 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bb25e0d6-95cd-42a8-a182-f714a14d0084-logs\") pod \"barbican-api-55cd89c446-4c5hx\" (UID: \"bb25e0d6-95cd-42a8-a182-f714a14d0084\") " pod="openstack/barbican-api-55cd89c446-4c5hx" Feb 27 08:09:52 crc kubenswrapper[4612]: I0227 08:09:52.825040 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bb25e0d6-95cd-42a8-a182-f714a14d0084-config-data\") pod \"barbican-api-55cd89c446-4c5hx\" (UID: \"bb25e0d6-95cd-42a8-a182-f714a14d0084\") " pod="openstack/barbican-api-55cd89c446-4c5hx" Feb 27 08:09:52 crc kubenswrapper[4612]: I0227 08:09:52.825058 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bb25e0d6-95cd-42a8-a182-f714a14d0084-config-data-custom\") pod \"barbican-api-55cd89c446-4c5hx\" (UID: \"bb25e0d6-95cd-42a8-a182-f714a14d0084\") " pod="openstack/barbican-api-55cd89c446-4c5hx" Feb 27 08:09:52 crc kubenswrapper[4612]: I0227 08:09:52.825265 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mkkdq\" (UniqueName: \"kubernetes.io/projected/bb25e0d6-95cd-42a8-a182-f714a14d0084-kube-api-access-mkkdq\") pod \"barbican-api-55cd89c446-4c5hx\" (UID: \"bb25e0d6-95cd-42a8-a182-f714a14d0084\") " pod="openstack/barbican-api-55cd89c446-4c5hx" Feb 27 08:09:52 crc kubenswrapper[4612]: I0227 08:09:52.872852 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-7d477cc4b6-5qtkx" Feb 27 08:09:52 crc kubenswrapper[4612]: I0227 08:09:52.889222 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fd3bd506-aad1-468a-b469-64ad78ec33fe" path="/var/lib/kubelet/pods/fd3bd506-aad1-468a-b469-64ad78ec33fe/volumes" Feb 27 08:09:52 crc kubenswrapper[4612]: I0227 08:09:52.926932 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bb25e0d6-95cd-42a8-a182-f714a14d0084-logs\") pod \"barbican-api-55cd89c446-4c5hx\" (UID: \"bb25e0d6-95cd-42a8-a182-f714a14d0084\") " pod="openstack/barbican-api-55cd89c446-4c5hx" Feb 27 08:09:52 crc kubenswrapper[4612]: I0227 08:09:52.927290 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bb25e0d6-95cd-42a8-a182-f714a14d0084-config-data\") pod \"barbican-api-55cd89c446-4c5hx\" (UID: \"bb25e0d6-95cd-42a8-a182-f714a14d0084\") " pod="openstack/barbican-api-55cd89c446-4c5hx" Feb 27 08:09:52 crc kubenswrapper[4612]: I0227 08:09:52.927410 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bb25e0d6-95cd-42a8-a182-f714a14d0084-config-data-custom\") pod \"barbican-api-55cd89c446-4c5hx\" (UID: \"bb25e0d6-95cd-42a8-a182-f714a14d0084\") " pod="openstack/barbican-api-55cd89c446-4c5hx" Feb 27 08:09:52 crc kubenswrapper[4612]: I0227 08:09:52.927513 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mkkdq\" (UniqueName: \"kubernetes.io/projected/bb25e0d6-95cd-42a8-a182-f714a14d0084-kube-api-access-mkkdq\") pod \"barbican-api-55cd89c446-4c5hx\" (UID: \"bb25e0d6-95cd-42a8-a182-f714a14d0084\") " pod="openstack/barbican-api-55cd89c446-4c5hx" Feb 27 08:09:52 crc kubenswrapper[4612]: I0227 08:09:52.927631 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb25e0d6-95cd-42a8-a182-f714a14d0084-combined-ca-bundle\") pod \"barbican-api-55cd89c446-4c5hx\" (UID: \"bb25e0d6-95cd-42a8-a182-f714a14d0084\") " pod="openstack/barbican-api-55cd89c446-4c5hx" Feb 27 08:09:52 crc kubenswrapper[4612]: I0227 08:09:52.944750 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bb25e0d6-95cd-42a8-a182-f714a14d0084-logs\") pod \"barbican-api-55cd89c446-4c5hx\" (UID: \"bb25e0d6-95cd-42a8-a182-f714a14d0084\") " pod="openstack/barbican-api-55cd89c446-4c5hx" Feb 27 08:09:53 crc kubenswrapper[4612]: I0227 08:09:52.973276 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mkkdq\" (UniqueName: \"kubernetes.io/projected/bb25e0d6-95cd-42a8-a182-f714a14d0084-kube-api-access-mkkdq\") pod \"barbican-api-55cd89c446-4c5hx\" (UID: \"bb25e0d6-95cd-42a8-a182-f714a14d0084\") " pod="openstack/barbican-api-55cd89c446-4c5hx" Feb 27 08:09:53 crc kubenswrapper[4612]: I0227 08:09:52.979315 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb25e0d6-95cd-42a8-a182-f714a14d0084-combined-ca-bundle\") pod \"barbican-api-55cd89c446-4c5hx\" (UID: \"bb25e0d6-95cd-42a8-a182-f714a14d0084\") " pod="openstack/barbican-api-55cd89c446-4c5hx" Feb 27 08:09:53 crc kubenswrapper[4612]: I0227 08:09:52.980556 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bb25e0d6-95cd-42a8-a182-f714a14d0084-config-data\") pod \"barbican-api-55cd89c446-4c5hx\" (UID: \"bb25e0d6-95cd-42a8-a182-f714a14d0084\") " pod="openstack/barbican-api-55cd89c446-4c5hx" Feb 27 08:09:53 crc kubenswrapper[4612]: I0227 08:09:52.981074 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bb25e0d6-95cd-42a8-a182-f714a14d0084-config-data-custom\") pod \"barbican-api-55cd89c446-4c5hx\" (UID: \"bb25e0d6-95cd-42a8-a182-f714a14d0084\") " pod="openstack/barbican-api-55cd89c446-4c5hx" Feb 27 08:09:53 crc kubenswrapper[4612]: I0227 08:09:53.232277 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-55cd89c446-4c5hx" Feb 27 08:09:53 crc kubenswrapper[4612]: I0227 08:09:53.553955 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f3bb8dc2-0c27-4a7a-baa1-1bfd8916f920","Type":"ContainerStarted","Data":"c698bb30368f5b284e472e8eeb1597a14d477233b727140ada2c89a5d8044522"} Feb 27 08:09:53 crc kubenswrapper[4612]: I0227 08:09:53.556521 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-9c6fdd6-lvn8p" event={"ID":"71940e05-ecfa-46de-aefe-c213add5362c","Type":"ContainerStarted","Data":"476747bd14c0f94563df02317b5e7cfd56d7e647bdd132965fcc9452742a4a8a"} Feb 27 08:09:53 crc kubenswrapper[4612]: I0227 08:09:53.556540 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-9c6fdd6-lvn8p" event={"ID":"71940e05-ecfa-46de-aefe-c213add5362c","Type":"ContainerStarted","Data":"96039e9fb1057bde110fe3f8907d771c76e869305dbdda0ac848fadd0c73b890"} Feb 27 08:09:53 crc kubenswrapper[4612]: I0227 08:09:53.557233 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-9c6fdd6-lvn8p" Feb 27 08:09:53 crc kubenswrapper[4612]: I0227 08:09:53.580613 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=10.580595216 podStartE2EDuration="10.580595216s" podCreationTimestamp="2026-02-27 08:09:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 08:09:53.5800394 +0000 UTC m=+1251.433969398" watchObservedRunningTime="2026-02-27 08:09:53.580595216 +0000 UTC m=+1251.434525214" Feb 27 08:09:53 crc kubenswrapper[4612]: I0227 08:09:53.616122 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-9c6fdd6-lvn8p" podStartSLOduration=2.616100223 podStartE2EDuration="2.616100223s" podCreationTimestamp="2026-02-27 08:09:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 08:09:53.601636619 +0000 UTC m=+1251.455566617" watchObservedRunningTime="2026-02-27 08:09:53.616100223 +0000 UTC m=+1251.470030221" Feb 27 08:09:53 crc kubenswrapper[4612]: I0227 08:09:53.667778 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-7558bddf6c-j9f2z"] Feb 27 08:09:53 crc kubenswrapper[4612]: I0227 08:09:53.782370 4612 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Feb 27 08:09:53 crc kubenswrapper[4612]: I0227 08:09:53.782491 4612 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Feb 27 08:09:53 crc kubenswrapper[4612]: I0227 08:09:53.842763 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-745b984886-tzpzt"] Feb 27 08:09:53 crc kubenswrapper[4612]: I0227 08:09:53.850615 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-7d477cc4b6-5qtkx"] Feb 27 08:09:53 crc kubenswrapper[4612]: W0227 08:09:53.877807 4612 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode2579b13_034e_474a_bbe6_2062a710f380.slice/crio-f9e4856dd7de79d452f6de21f8364191a8ce8ac023ea035cc269da1a852f978e WatchSource:0}: Error finding container f9e4856dd7de79d452f6de21f8364191a8ce8ac023ea035cc269da1a852f978e: Status 404 returned error can't find the container with id f9e4856dd7de79d452f6de21f8364191a8ce8ac023ea035cc269da1a852f978e Feb 27 08:09:53 crc kubenswrapper[4612]: I0227 08:09:53.938919 4612 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Feb 27 08:09:54 crc kubenswrapper[4612]: I0227 08:09:54.081800 4612 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Feb 27 08:09:54 crc kubenswrapper[4612]: I0227 08:09:54.124198 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-848cf88cfc-tz267"] Feb 27 08:09:54 crc kubenswrapper[4612]: I0227 08:09:54.296815 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-55cd89c446-4c5hx"] Feb 27 08:09:54 crc kubenswrapper[4612]: I0227 08:09:54.605300 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-745b984886-tzpzt" event={"ID":"e2579b13-034e-474a-bbe6-2062a710f380","Type":"ContainerStarted","Data":"f9e4856dd7de79d452f6de21f8364191a8ce8ac023ea035cc269da1a852f978e"} Feb 27 08:09:54 crc kubenswrapper[4612]: I0227 08:09:54.613388 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-55cd89c446-4c5hx" event={"ID":"bb25e0d6-95cd-42a8-a182-f714a14d0084","Type":"ContainerStarted","Data":"80abde5b490b842dde142302c01ef4b837956aa9c3244ea384215f0463274da5"} Feb 27 08:09:54 crc kubenswrapper[4612]: I0227 08:09:54.616512 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-7558bddf6c-j9f2z" event={"ID":"74bec848-d5b3-43d0-97d7-eda229642bb5","Type":"ContainerStarted","Data":"0a6461d10ca55d949f02386eadb5641242e43dace29277e3949cb5603f5c1cc8"} Feb 27 08:09:54 crc kubenswrapper[4612]: I0227 08:09:54.621131 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-848cf88cfc-tz267" event={"ID":"61899b16-fb85-4ba6-b057-65ad130ea015","Type":"ContainerStarted","Data":"67dbb6b7b932909ae4d2afaae2e16ad844e84d4fa7023cf89cf20bf38cc72f6a"} Feb 27 08:09:54 crc kubenswrapper[4612]: I0227 08:09:54.634008 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-7d477cc4b6-5qtkx" event={"ID":"5fb13c78-0dab-4785-90bc-21d5dabe81f6","Type":"ContainerStarted","Data":"733ae42ff1eca6902bba2de5e97073b8073d88406f9a965bddd2479f96df6cf0"} Feb 27 08:09:54 crc kubenswrapper[4612]: I0227 08:09:54.634076 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-7d477cc4b6-5qtkx" event={"ID":"5fb13c78-0dab-4785-90bc-21d5dabe81f6","Type":"ContainerStarted","Data":"e89273ed73ffedae95676df852754b419cd8932f33cda2ca62ed3409ed5ace4d"} Feb 27 08:09:54 crc kubenswrapper[4612]: I0227 08:09:54.634514 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Feb 27 08:09:54 crc kubenswrapper[4612]: I0227 08:09:54.634577 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Feb 27 08:09:55 crc kubenswrapper[4612]: I0227 08:09:55.655419 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-55cd89c446-4c5hx" event={"ID":"bb25e0d6-95cd-42a8-a182-f714a14d0084","Type":"ContainerStarted","Data":"fe5438c31c8c7aa2197f6896cae35a810753162b8b67a1c965e8e2d4cc8a9dfa"} Feb 27 08:09:55 crc kubenswrapper[4612]: I0227 08:09:55.657117 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-55cd89c446-4c5hx" Feb 27 08:09:55 crc kubenswrapper[4612]: I0227 08:09:55.657213 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-55cd89c446-4c5hx" event={"ID":"bb25e0d6-95cd-42a8-a182-f714a14d0084","Type":"ContainerStarted","Data":"093ddc25c2613ae5cc7e84e5832820695b3bdc3586636199cfdfdf0a520691c0"} Feb 27 08:09:55 crc kubenswrapper[4612]: I0227 08:09:55.657284 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-55cd89c446-4c5hx" Feb 27 08:09:55 crc kubenswrapper[4612]: I0227 08:09:55.675003 4612 generic.go:334] "Generic (PLEG): container finished" podID="61899b16-fb85-4ba6-b057-65ad130ea015" containerID="d825d98443adaf0ad96e63a1bf5e5533113dab4ebe6f67e171d94b86c28fd84f" exitCode=0 Feb 27 08:09:55 crc kubenswrapper[4612]: I0227 08:09:55.675100 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-848cf88cfc-tz267" event={"ID":"61899b16-fb85-4ba6-b057-65ad130ea015","Type":"ContainerDied","Data":"d825d98443adaf0ad96e63a1bf5e5533113dab4ebe6f67e171d94b86c28fd84f"} Feb 27 08:09:55 crc kubenswrapper[4612]: I0227 08:09:55.679509 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-55cd89c446-4c5hx" podStartSLOduration=3.679493908 podStartE2EDuration="3.679493908s" podCreationTimestamp="2026-02-27 08:09:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 08:09:55.676242755 +0000 UTC m=+1253.530172753" watchObservedRunningTime="2026-02-27 08:09:55.679493908 +0000 UTC m=+1253.533423906" Feb 27 08:09:55 crc kubenswrapper[4612]: I0227 08:09:55.687913 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-7d477cc4b6-5qtkx" event={"ID":"5fb13c78-0dab-4785-90bc-21d5dabe81f6","Type":"ContainerStarted","Data":"62dd8ae080d2b8abda6ce8d5de55c8a0534a0f2871da41818e711c1f54bca8c4"} Feb 27 08:09:55 crc kubenswrapper[4612]: I0227 08:09:55.688727 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-7d477cc4b6-5qtkx" Feb 27 08:09:55 crc kubenswrapper[4612]: I0227 08:09:55.688747 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-7d477cc4b6-5qtkx" Feb 27 08:09:55 crc kubenswrapper[4612]: I0227 08:09:55.695498 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-l5z5s" event={"ID":"0c9482e8-9eb1-4c30-95e9-bdaa5ccd239b","Type":"ContainerStarted","Data":"a35ff58cdedc1d36f8345528a64eca1b073d88b1dbf4edfd2f202894c24e07b0"} Feb 27 08:09:55 crc kubenswrapper[4612]: I0227 08:09:55.765653 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-7d477cc4b6-5qtkx" podStartSLOduration=3.765634616 podStartE2EDuration="3.765634616s" podCreationTimestamp="2026-02-27 08:09:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 08:09:55.734317029 +0000 UTC m=+1253.588247037" watchObservedRunningTime="2026-02-27 08:09:55.765634616 +0000 UTC m=+1253.619564614" Feb 27 08:09:55 crc kubenswrapper[4612]: I0227 08:09:55.765999 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-l5z5s" podStartSLOduration=4.504808048 podStartE2EDuration="56.765994247s" podCreationTimestamp="2026-02-27 08:08:59 +0000 UTC" firstStartedPulling="2026-02-27 08:09:01.333622684 +0000 UTC m=+1199.187552682" lastFinishedPulling="2026-02-27 08:09:53.594808883 +0000 UTC m=+1251.448738881" observedRunningTime="2026-02-27 08:09:55.751865982 +0000 UTC m=+1253.605795990" watchObservedRunningTime="2026-02-27 08:09:55.765994247 +0000 UTC m=+1253.619924245" Feb 27 08:09:55 crc kubenswrapper[4612]: I0227 08:09:55.950267 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-7db7f7dcc8-f95qr"] Feb 27 08:09:55 crc kubenswrapper[4612]: I0227 08:09:55.954293 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-7db7f7dcc8-f95qr" Feb 27 08:09:55 crc kubenswrapper[4612]: I0227 08:09:55.960213 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Feb 27 08:09:55 crc kubenswrapper[4612]: I0227 08:09:55.960432 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Feb 27 08:09:55 crc kubenswrapper[4612]: I0227 08:09:55.987046 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-7db7f7dcc8-f95qr"] Feb 27 08:09:56 crc kubenswrapper[4612]: I0227 08:09:56.018610 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/16621861-feac-4972-ad1c-1e365692c23e-combined-ca-bundle\") pod \"barbican-api-7db7f7dcc8-f95qr\" (UID: \"16621861-feac-4972-ad1c-1e365692c23e\") " pod="openstack/barbican-api-7db7f7dcc8-f95qr" Feb 27 08:09:56 crc kubenswrapper[4612]: I0227 08:09:56.018675 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/16621861-feac-4972-ad1c-1e365692c23e-internal-tls-certs\") pod \"barbican-api-7db7f7dcc8-f95qr\" (UID: \"16621861-feac-4972-ad1c-1e365692c23e\") " pod="openstack/barbican-api-7db7f7dcc8-f95qr" Feb 27 08:09:56 crc kubenswrapper[4612]: I0227 08:09:56.018782 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wwlvz\" (UniqueName: \"kubernetes.io/projected/16621861-feac-4972-ad1c-1e365692c23e-kube-api-access-wwlvz\") pod \"barbican-api-7db7f7dcc8-f95qr\" (UID: \"16621861-feac-4972-ad1c-1e365692c23e\") " pod="openstack/barbican-api-7db7f7dcc8-f95qr" Feb 27 08:09:56 crc kubenswrapper[4612]: I0227 08:09:56.018833 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/16621861-feac-4972-ad1c-1e365692c23e-logs\") pod \"barbican-api-7db7f7dcc8-f95qr\" (UID: \"16621861-feac-4972-ad1c-1e365692c23e\") " pod="openstack/barbican-api-7db7f7dcc8-f95qr" Feb 27 08:09:56 crc kubenswrapper[4612]: I0227 08:09:56.018853 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/16621861-feac-4972-ad1c-1e365692c23e-config-data-custom\") pod \"barbican-api-7db7f7dcc8-f95qr\" (UID: \"16621861-feac-4972-ad1c-1e365692c23e\") " pod="openstack/barbican-api-7db7f7dcc8-f95qr" Feb 27 08:09:56 crc kubenswrapper[4612]: I0227 08:09:56.018882 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/16621861-feac-4972-ad1c-1e365692c23e-config-data\") pod \"barbican-api-7db7f7dcc8-f95qr\" (UID: \"16621861-feac-4972-ad1c-1e365692c23e\") " pod="openstack/barbican-api-7db7f7dcc8-f95qr" Feb 27 08:09:56 crc kubenswrapper[4612]: I0227 08:09:56.018913 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/16621861-feac-4972-ad1c-1e365692c23e-public-tls-certs\") pod \"barbican-api-7db7f7dcc8-f95qr\" (UID: \"16621861-feac-4972-ad1c-1e365692c23e\") " pod="openstack/barbican-api-7db7f7dcc8-f95qr" Feb 27 08:09:56 crc kubenswrapper[4612]: I0227 08:09:56.120790 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/16621861-feac-4972-ad1c-1e365692c23e-combined-ca-bundle\") pod \"barbican-api-7db7f7dcc8-f95qr\" (UID: \"16621861-feac-4972-ad1c-1e365692c23e\") " pod="openstack/barbican-api-7db7f7dcc8-f95qr" Feb 27 08:09:56 crc kubenswrapper[4612]: I0227 08:09:56.121168 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/16621861-feac-4972-ad1c-1e365692c23e-internal-tls-certs\") pod \"barbican-api-7db7f7dcc8-f95qr\" (UID: \"16621861-feac-4972-ad1c-1e365692c23e\") " pod="openstack/barbican-api-7db7f7dcc8-f95qr" Feb 27 08:09:56 crc kubenswrapper[4612]: I0227 08:09:56.121242 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wwlvz\" (UniqueName: \"kubernetes.io/projected/16621861-feac-4972-ad1c-1e365692c23e-kube-api-access-wwlvz\") pod \"barbican-api-7db7f7dcc8-f95qr\" (UID: \"16621861-feac-4972-ad1c-1e365692c23e\") " pod="openstack/barbican-api-7db7f7dcc8-f95qr" Feb 27 08:09:56 crc kubenswrapper[4612]: I0227 08:09:56.121281 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/16621861-feac-4972-ad1c-1e365692c23e-logs\") pod \"barbican-api-7db7f7dcc8-f95qr\" (UID: \"16621861-feac-4972-ad1c-1e365692c23e\") " pod="openstack/barbican-api-7db7f7dcc8-f95qr" Feb 27 08:09:56 crc kubenswrapper[4612]: I0227 08:09:56.121302 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/16621861-feac-4972-ad1c-1e365692c23e-config-data-custom\") pod \"barbican-api-7db7f7dcc8-f95qr\" (UID: \"16621861-feac-4972-ad1c-1e365692c23e\") " pod="openstack/barbican-api-7db7f7dcc8-f95qr" Feb 27 08:09:56 crc kubenswrapper[4612]: I0227 08:09:56.121328 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/16621861-feac-4972-ad1c-1e365692c23e-config-data\") pod \"barbican-api-7db7f7dcc8-f95qr\" (UID: \"16621861-feac-4972-ad1c-1e365692c23e\") " pod="openstack/barbican-api-7db7f7dcc8-f95qr" Feb 27 08:09:56 crc kubenswrapper[4612]: I0227 08:09:56.121351 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/16621861-feac-4972-ad1c-1e365692c23e-public-tls-certs\") pod \"barbican-api-7db7f7dcc8-f95qr\" (UID: \"16621861-feac-4972-ad1c-1e365692c23e\") " pod="openstack/barbican-api-7db7f7dcc8-f95qr" Feb 27 08:09:56 crc kubenswrapper[4612]: I0227 08:09:56.121820 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/16621861-feac-4972-ad1c-1e365692c23e-logs\") pod \"barbican-api-7db7f7dcc8-f95qr\" (UID: \"16621861-feac-4972-ad1c-1e365692c23e\") " pod="openstack/barbican-api-7db7f7dcc8-f95qr" Feb 27 08:09:56 crc kubenswrapper[4612]: I0227 08:09:56.129479 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/16621861-feac-4972-ad1c-1e365692c23e-config-data\") pod \"barbican-api-7db7f7dcc8-f95qr\" (UID: \"16621861-feac-4972-ad1c-1e365692c23e\") " pod="openstack/barbican-api-7db7f7dcc8-f95qr" Feb 27 08:09:56 crc kubenswrapper[4612]: I0227 08:09:56.132158 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/16621861-feac-4972-ad1c-1e365692c23e-config-data-custom\") pod \"barbican-api-7db7f7dcc8-f95qr\" (UID: \"16621861-feac-4972-ad1c-1e365692c23e\") " pod="openstack/barbican-api-7db7f7dcc8-f95qr" Feb 27 08:09:56 crc kubenswrapper[4612]: I0227 08:09:56.135416 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/16621861-feac-4972-ad1c-1e365692c23e-public-tls-certs\") pod \"barbican-api-7db7f7dcc8-f95qr\" (UID: \"16621861-feac-4972-ad1c-1e365692c23e\") " pod="openstack/barbican-api-7db7f7dcc8-f95qr" Feb 27 08:09:56 crc kubenswrapper[4612]: I0227 08:09:56.149748 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wwlvz\" (UniqueName: \"kubernetes.io/projected/16621861-feac-4972-ad1c-1e365692c23e-kube-api-access-wwlvz\") pod \"barbican-api-7db7f7dcc8-f95qr\" (UID: \"16621861-feac-4972-ad1c-1e365692c23e\") " pod="openstack/barbican-api-7db7f7dcc8-f95qr" Feb 27 08:09:56 crc kubenswrapper[4612]: I0227 08:09:56.149856 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/16621861-feac-4972-ad1c-1e365692c23e-internal-tls-certs\") pod \"barbican-api-7db7f7dcc8-f95qr\" (UID: \"16621861-feac-4972-ad1c-1e365692c23e\") " pod="openstack/barbican-api-7db7f7dcc8-f95qr" Feb 27 08:09:56 crc kubenswrapper[4612]: I0227 08:09:56.151173 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/16621861-feac-4972-ad1c-1e365692c23e-combined-ca-bundle\") pod \"barbican-api-7db7f7dcc8-f95qr\" (UID: \"16621861-feac-4972-ad1c-1e365692c23e\") " pod="openstack/barbican-api-7db7f7dcc8-f95qr" Feb 27 08:09:56 crc kubenswrapper[4612]: I0227 08:09:56.294474 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-7db7f7dcc8-f95qr" Feb 27 08:09:56 crc kubenswrapper[4612]: I0227 08:09:56.715157 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-848cf88cfc-tz267" event={"ID":"61899b16-fb85-4ba6-b057-65ad130ea015","Type":"ContainerStarted","Data":"452de9407de7ea11ac17676dc5978b7bcc066fe94ddc3253fac1679600573f08"} Feb 27 08:09:56 crc kubenswrapper[4612]: I0227 08:09:56.715727 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-848cf88cfc-tz267" Feb 27 08:09:56 crc kubenswrapper[4612]: I0227 08:09:56.756408 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-848cf88cfc-tz267" podStartSLOduration=4.756388246 podStartE2EDuration="4.756388246s" podCreationTimestamp="2026-02-27 08:09:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 08:09:56.737804043 +0000 UTC m=+1254.591734041" watchObservedRunningTime="2026-02-27 08:09:56.756388246 +0000 UTC m=+1254.610318244" Feb 27 08:09:59 crc kubenswrapper[4612]: I0227 08:09:59.099590 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-7db7f7dcc8-f95qr"] Feb 27 08:09:59 crc kubenswrapper[4612]: I0227 08:09:59.109912 4612 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-6b985547d4-mkfpv" podUID="76009243-16b3-4652-abfe-a6cdba363724" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.152:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.152:8443: connect: connection refused" Feb 27 08:09:59 crc kubenswrapper[4612]: I0227 08:09:59.275062 4612 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-77948db5bb-97qh6" podUID="b72b541d-4661-44a7-a121-c2a8aef6db11" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.153:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.153:8443: connect: connection refused" Feb 27 08:09:59 crc kubenswrapper[4612]: I0227 08:09:59.666839 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Feb 27 08:09:59 crc kubenswrapper[4612]: I0227 08:09:59.666969 4612 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 27 08:09:59 crc kubenswrapper[4612]: I0227 08:09:59.746141 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Feb 27 08:09:59 crc kubenswrapper[4612]: I0227 08:09:59.827416 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7db7f7dcc8-f95qr" event={"ID":"16621861-feac-4972-ad1c-1e365692c23e","Type":"ContainerStarted","Data":"30372e568bf8a0d10173191c5f34840f082ca08e39631ab17bfdfb1b8ffeb206"} Feb 27 08:09:59 crc kubenswrapper[4612]: I0227 08:09:59.827459 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7db7f7dcc8-f95qr" event={"ID":"16621861-feac-4972-ad1c-1e365692c23e","Type":"ContainerStarted","Data":"ce485b5024245abd356026b5e03cafc2e3160f5aaf26a1e2e841bd963cb808bb"} Feb 27 08:09:59 crc kubenswrapper[4612]: I0227 08:09:59.831067 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-745b984886-tzpzt" event={"ID":"e2579b13-034e-474a-bbe6-2062a710f380","Type":"ContainerStarted","Data":"5a421e0b77a79855045a49cf9e6704d009c86579a099245df52b57460e425f43"} Feb 27 08:09:59 crc kubenswrapper[4612]: I0227 08:09:59.839837 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-7558bddf6c-j9f2z" event={"ID":"74bec848-d5b3-43d0-97d7-eda229642bb5","Type":"ContainerStarted","Data":"bdb9f347334b0974b8361e4368b047b21d9ea07584cd3662758be865453c05af"} Feb 27 08:09:59 crc kubenswrapper[4612]: I0227 08:09:59.839890 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-7558bddf6c-j9f2z" event={"ID":"74bec848-d5b3-43d0-97d7-eda229642bb5","Type":"ContainerStarted","Data":"46b2478005f05a68148245f2d1a9f8e57f2bca767e6246c85c4ac87be87f4c3b"} Feb 27 08:09:59 crc kubenswrapper[4612]: I0227 08:09:59.856412 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-7558bddf6c-j9f2z" podStartSLOduration=3.259955327 podStartE2EDuration="7.856394784s" podCreationTimestamp="2026-02-27 08:09:52 +0000 UTC" firstStartedPulling="2026-02-27 08:09:53.688065445 +0000 UTC m=+1251.541995443" lastFinishedPulling="2026-02-27 08:09:58.284504902 +0000 UTC m=+1256.138434900" observedRunningTime="2026-02-27 08:09:59.855013664 +0000 UTC m=+1257.708943662" watchObservedRunningTime="2026-02-27 08:09:59.856394784 +0000 UTC m=+1257.710324772" Feb 27 08:10:00 crc kubenswrapper[4612]: I0227 08:10:00.166872 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29536330-nck2q"] Feb 27 08:10:00 crc kubenswrapper[4612]: I0227 08:10:00.168523 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536330-nck2q" Feb 27 08:10:00 crc kubenswrapper[4612]: I0227 08:10:00.175516 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-5zwfx" Feb 27 08:10:00 crc kubenswrapper[4612]: I0227 08:10:00.175916 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 27 08:10:00 crc kubenswrapper[4612]: I0227 08:10:00.176809 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 27 08:10:00 crc kubenswrapper[4612]: I0227 08:10:00.180796 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536330-nck2q"] Feb 27 08:10:00 crc kubenswrapper[4612]: I0227 08:10:00.313188 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fzgl5\" (UniqueName: \"kubernetes.io/projected/073232e9-6f4e-409e-bf29-989b354fcc39-kube-api-access-fzgl5\") pod \"auto-csr-approver-29536330-nck2q\" (UID: \"073232e9-6f4e-409e-bf29-989b354fcc39\") " pod="openshift-infra/auto-csr-approver-29536330-nck2q" Feb 27 08:10:00 crc kubenswrapper[4612]: I0227 08:10:00.415321 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fzgl5\" (UniqueName: \"kubernetes.io/projected/073232e9-6f4e-409e-bf29-989b354fcc39-kube-api-access-fzgl5\") pod \"auto-csr-approver-29536330-nck2q\" (UID: \"073232e9-6f4e-409e-bf29-989b354fcc39\") " pod="openshift-infra/auto-csr-approver-29536330-nck2q" Feb 27 08:10:00 crc kubenswrapper[4612]: I0227 08:10:00.453740 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fzgl5\" (UniqueName: \"kubernetes.io/projected/073232e9-6f4e-409e-bf29-989b354fcc39-kube-api-access-fzgl5\") pod \"auto-csr-approver-29536330-nck2q\" (UID: \"073232e9-6f4e-409e-bf29-989b354fcc39\") " pod="openshift-infra/auto-csr-approver-29536330-nck2q" Feb 27 08:10:00 crc kubenswrapper[4612]: I0227 08:10:00.493323 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536330-nck2q" Feb 27 08:10:00 crc kubenswrapper[4612]: I0227 08:10:00.873261 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-745b984886-tzpzt" event={"ID":"e2579b13-034e-474a-bbe6-2062a710f380","Type":"ContainerStarted","Data":"744311f0310bc8915140f36f0d62f92ae95ee319a62863c7ae430083fb2baadb"} Feb 27 08:10:00 crc kubenswrapper[4612]: I0227 08:10:00.884976 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7db7f7dcc8-f95qr" event={"ID":"16621861-feac-4972-ad1c-1e365692c23e","Type":"ContainerStarted","Data":"6614e85efec336766b03c3c3e6492ef0bab1511f15927eb94bf42d2cf615d94e"} Feb 27 08:10:00 crc kubenswrapper[4612]: I0227 08:10:00.885038 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-7db7f7dcc8-f95qr" Feb 27 08:10:00 crc kubenswrapper[4612]: I0227 08:10:00.885052 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-7db7f7dcc8-f95qr" Feb 27 08:10:00 crc kubenswrapper[4612]: I0227 08:10:00.894339 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-745b984886-tzpzt" podStartSLOduration=4.529245247 podStartE2EDuration="8.894326184s" podCreationTimestamp="2026-02-27 08:09:52 +0000 UTC" firstStartedPulling="2026-02-27 08:09:53.902572462 +0000 UTC m=+1251.756502460" lastFinishedPulling="2026-02-27 08:09:58.267653399 +0000 UTC m=+1256.121583397" observedRunningTime="2026-02-27 08:10:00.892825171 +0000 UTC m=+1258.746755169" watchObservedRunningTime="2026-02-27 08:10:00.894326184 +0000 UTC m=+1258.748256182" Feb 27 08:10:00 crc kubenswrapper[4612]: I0227 08:10:00.938126 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-7db7f7dcc8-f95qr" podStartSLOduration=5.938108649 podStartE2EDuration="5.938108649s" podCreationTimestamp="2026-02-27 08:09:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 08:10:00.934773073 +0000 UTC m=+1258.788703071" watchObservedRunningTime="2026-02-27 08:10:00.938108649 +0000 UTC m=+1258.792038647" Feb 27 08:10:01 crc kubenswrapper[4612]: I0227 08:10:01.097162 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536330-nck2q"] Feb 27 08:10:01 crc kubenswrapper[4612]: I0227 08:10:01.923477 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536330-nck2q" event={"ID":"073232e9-6f4e-409e-bf29-989b354fcc39","Type":"ContainerStarted","Data":"1755fd2ce8358fd0df3e900e64f6e14e513b1b51402750e05b2e3cd2e246327c"} Feb 27 08:10:02 crc kubenswrapper[4612]: I0227 08:10:02.820844 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-848cf88cfc-tz267" Feb 27 08:10:03 crc kubenswrapper[4612]: I0227 08:10:03.079939 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6b7b667979-pm5fl"] Feb 27 08:10:03 crc kubenswrapper[4612]: I0227 08:10:03.080546 4612 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6b7b667979-pm5fl" podUID="bf325571-0b74-447f-bd04-213d54a8437e" containerName="dnsmasq-dns" containerID="cri-o://cd920bf9cdd1f21efee933e8fb8dc84fd5aac93ceafa5fc01745e88c2e9aaf86" gracePeriod=10 Feb 27 08:10:03 crc kubenswrapper[4612]: I0227 08:10:03.975372 4612 generic.go:334] "Generic (PLEG): container finished" podID="bf325571-0b74-447f-bd04-213d54a8437e" containerID="cd920bf9cdd1f21efee933e8fb8dc84fd5aac93ceafa5fc01745e88c2e9aaf86" exitCode=0 Feb 27 08:10:03 crc kubenswrapper[4612]: I0227 08:10:03.975513 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b7b667979-pm5fl" event={"ID":"bf325571-0b74-447f-bd04-213d54a8437e","Type":"ContainerDied","Data":"cd920bf9cdd1f21efee933e8fb8dc84fd5aac93ceafa5fc01745e88c2e9aaf86"} Feb 27 08:10:05 crc kubenswrapper[4612]: I0227 08:10:05.430026 4612 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-6b7b667979-pm5fl" podUID="bf325571-0b74-447f-bd04-213d54a8437e" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.155:5353: connect: connection refused" Feb 27 08:10:05 crc kubenswrapper[4612]: I0227 08:10:05.601838 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-77854b564-8nkbp" Feb 27 08:10:05 crc kubenswrapper[4612]: I0227 08:10:05.931405 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-5ddbbb8689-bm5cp"] Feb 27 08:10:05 crc kubenswrapper[4612]: I0227 08:10:05.931751 4612 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-5ddbbb8689-bm5cp" podUID="649e0df0-b37c-42f5-8af0-c804d4f467d4" containerName="neutron-api" containerID="cri-o://c8e24f4c991021623da34d4abac610956069b675ba545401121c750c3d4a1070" gracePeriod=30 Feb 27 08:10:05 crc kubenswrapper[4612]: I0227 08:10:05.932122 4612 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-5ddbbb8689-bm5cp" podUID="649e0df0-b37c-42f5-8af0-c804d4f467d4" containerName="neutron-httpd" containerID="cri-o://399439f0cfe95ee5a2d4886d8c116dcf58ccc56b31529381b60a5bb7558643db" gracePeriod=30 Feb 27 08:10:05 crc kubenswrapper[4612]: I0227 08:10:05.962039 4612 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/neutron-5ddbbb8689-bm5cp" podUID="649e0df0-b37c-42f5-8af0-c804d4f467d4" containerName="neutron-httpd" probeResult="failure" output="Get \"https://10.217.0.157:9696/\": EOF" Feb 27 08:10:05 crc kubenswrapper[4612]: I0227 08:10:05.964797 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-cc6d746d7-b8wbb"] Feb 27 08:10:05 crc kubenswrapper[4612]: I0227 08:10:05.979931 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-cc6d746d7-b8wbb" Feb 27 08:10:05 crc kubenswrapper[4612]: I0227 08:10:05.983033 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/41eb38cf-80fa-408e-b16c-686ef58a143f-httpd-config\") pod \"neutron-cc6d746d7-b8wbb\" (UID: \"41eb38cf-80fa-408e-b16c-686ef58a143f\") " pod="openstack/neutron-cc6d746d7-b8wbb" Feb 27 08:10:05 crc kubenswrapper[4612]: I0227 08:10:05.990866 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/41eb38cf-80fa-408e-b16c-686ef58a143f-public-tls-certs\") pod \"neutron-cc6d746d7-b8wbb\" (UID: \"41eb38cf-80fa-408e-b16c-686ef58a143f\") " pod="openstack/neutron-cc6d746d7-b8wbb" Feb 27 08:10:05 crc kubenswrapper[4612]: I0227 08:10:05.991038 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/41eb38cf-80fa-408e-b16c-686ef58a143f-internal-tls-certs\") pod \"neutron-cc6d746d7-b8wbb\" (UID: \"41eb38cf-80fa-408e-b16c-686ef58a143f\") " pod="openstack/neutron-cc6d746d7-b8wbb" Feb 27 08:10:05 crc kubenswrapper[4612]: I0227 08:10:05.991093 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41eb38cf-80fa-408e-b16c-686ef58a143f-combined-ca-bundle\") pod \"neutron-cc6d746d7-b8wbb\" (UID: \"41eb38cf-80fa-408e-b16c-686ef58a143f\") " pod="openstack/neutron-cc6d746d7-b8wbb" Feb 27 08:10:05 crc kubenswrapper[4612]: I0227 08:10:05.991240 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/41eb38cf-80fa-408e-b16c-686ef58a143f-config\") pod \"neutron-cc6d746d7-b8wbb\" (UID: \"41eb38cf-80fa-408e-b16c-686ef58a143f\") " pod="openstack/neutron-cc6d746d7-b8wbb" Feb 27 08:10:05 crc kubenswrapper[4612]: I0227 08:10:05.991273 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/41eb38cf-80fa-408e-b16c-686ef58a143f-ovndb-tls-certs\") pod \"neutron-cc6d746d7-b8wbb\" (UID: \"41eb38cf-80fa-408e-b16c-686ef58a143f\") " pod="openstack/neutron-cc6d746d7-b8wbb" Feb 27 08:10:05 crc kubenswrapper[4612]: I0227 08:10:05.991342 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k49zg\" (UniqueName: \"kubernetes.io/projected/41eb38cf-80fa-408e-b16c-686ef58a143f-kube-api-access-k49zg\") pod \"neutron-cc6d746d7-b8wbb\" (UID: \"41eb38cf-80fa-408e-b16c-686ef58a143f\") " pod="openstack/neutron-cc6d746d7-b8wbb" Feb 27 08:10:06 crc kubenswrapper[4612]: I0227 08:10:06.003836 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-cc6d746d7-b8wbb"] Feb 27 08:10:06 crc kubenswrapper[4612]: I0227 08:10:06.093642 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/41eb38cf-80fa-408e-b16c-686ef58a143f-httpd-config\") pod \"neutron-cc6d746d7-b8wbb\" (UID: \"41eb38cf-80fa-408e-b16c-686ef58a143f\") " pod="openstack/neutron-cc6d746d7-b8wbb" Feb 27 08:10:06 crc kubenswrapper[4612]: I0227 08:10:06.093684 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/41eb38cf-80fa-408e-b16c-686ef58a143f-public-tls-certs\") pod \"neutron-cc6d746d7-b8wbb\" (UID: \"41eb38cf-80fa-408e-b16c-686ef58a143f\") " pod="openstack/neutron-cc6d746d7-b8wbb" Feb 27 08:10:06 crc kubenswrapper[4612]: I0227 08:10:06.093750 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/41eb38cf-80fa-408e-b16c-686ef58a143f-internal-tls-certs\") pod \"neutron-cc6d746d7-b8wbb\" (UID: \"41eb38cf-80fa-408e-b16c-686ef58a143f\") " pod="openstack/neutron-cc6d746d7-b8wbb" Feb 27 08:10:06 crc kubenswrapper[4612]: I0227 08:10:06.093777 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41eb38cf-80fa-408e-b16c-686ef58a143f-combined-ca-bundle\") pod \"neutron-cc6d746d7-b8wbb\" (UID: \"41eb38cf-80fa-408e-b16c-686ef58a143f\") " pod="openstack/neutron-cc6d746d7-b8wbb" Feb 27 08:10:06 crc kubenswrapper[4612]: I0227 08:10:06.093839 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/41eb38cf-80fa-408e-b16c-686ef58a143f-config\") pod \"neutron-cc6d746d7-b8wbb\" (UID: \"41eb38cf-80fa-408e-b16c-686ef58a143f\") " pod="openstack/neutron-cc6d746d7-b8wbb" Feb 27 08:10:06 crc kubenswrapper[4612]: I0227 08:10:06.093860 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/41eb38cf-80fa-408e-b16c-686ef58a143f-ovndb-tls-certs\") pod \"neutron-cc6d746d7-b8wbb\" (UID: \"41eb38cf-80fa-408e-b16c-686ef58a143f\") " pod="openstack/neutron-cc6d746d7-b8wbb" Feb 27 08:10:06 crc kubenswrapper[4612]: I0227 08:10:06.093889 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k49zg\" (UniqueName: \"kubernetes.io/projected/41eb38cf-80fa-408e-b16c-686ef58a143f-kube-api-access-k49zg\") pod \"neutron-cc6d746d7-b8wbb\" (UID: \"41eb38cf-80fa-408e-b16c-686ef58a143f\") " pod="openstack/neutron-cc6d746d7-b8wbb" Feb 27 08:10:06 crc kubenswrapper[4612]: I0227 08:10:06.100520 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/41eb38cf-80fa-408e-b16c-686ef58a143f-httpd-config\") pod \"neutron-cc6d746d7-b8wbb\" (UID: \"41eb38cf-80fa-408e-b16c-686ef58a143f\") " pod="openstack/neutron-cc6d746d7-b8wbb" Feb 27 08:10:06 crc kubenswrapper[4612]: I0227 08:10:06.100525 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41eb38cf-80fa-408e-b16c-686ef58a143f-combined-ca-bundle\") pod \"neutron-cc6d746d7-b8wbb\" (UID: \"41eb38cf-80fa-408e-b16c-686ef58a143f\") " pod="openstack/neutron-cc6d746d7-b8wbb" Feb 27 08:10:06 crc kubenswrapper[4612]: I0227 08:10:06.107220 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/41eb38cf-80fa-408e-b16c-686ef58a143f-config\") pod \"neutron-cc6d746d7-b8wbb\" (UID: \"41eb38cf-80fa-408e-b16c-686ef58a143f\") " pod="openstack/neutron-cc6d746d7-b8wbb" Feb 27 08:10:06 crc kubenswrapper[4612]: I0227 08:10:06.111313 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/41eb38cf-80fa-408e-b16c-686ef58a143f-public-tls-certs\") pod \"neutron-cc6d746d7-b8wbb\" (UID: \"41eb38cf-80fa-408e-b16c-686ef58a143f\") " pod="openstack/neutron-cc6d746d7-b8wbb" Feb 27 08:10:06 crc kubenswrapper[4612]: I0227 08:10:06.111860 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/41eb38cf-80fa-408e-b16c-686ef58a143f-ovndb-tls-certs\") pod \"neutron-cc6d746d7-b8wbb\" (UID: \"41eb38cf-80fa-408e-b16c-686ef58a143f\") " pod="openstack/neutron-cc6d746d7-b8wbb" Feb 27 08:10:06 crc kubenswrapper[4612]: I0227 08:10:06.149435 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/41eb38cf-80fa-408e-b16c-686ef58a143f-internal-tls-certs\") pod \"neutron-cc6d746d7-b8wbb\" (UID: \"41eb38cf-80fa-408e-b16c-686ef58a143f\") " pod="openstack/neutron-cc6d746d7-b8wbb" Feb 27 08:10:06 crc kubenswrapper[4612]: I0227 08:10:06.149511 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k49zg\" (UniqueName: \"kubernetes.io/projected/41eb38cf-80fa-408e-b16c-686ef58a143f-kube-api-access-k49zg\") pod \"neutron-cc6d746d7-b8wbb\" (UID: \"41eb38cf-80fa-408e-b16c-686ef58a143f\") " pod="openstack/neutron-cc6d746d7-b8wbb" Feb 27 08:10:06 crc kubenswrapper[4612]: I0227 08:10:06.337493 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-cc6d746d7-b8wbb" Feb 27 08:10:06 crc kubenswrapper[4612]: I0227 08:10:06.788191 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-55cd89c446-4c5hx" Feb 27 08:10:07 crc kubenswrapper[4612]: I0227 08:10:07.026115 4612 generic.go:334] "Generic (PLEG): container finished" podID="0c9482e8-9eb1-4c30-95e9-bdaa5ccd239b" containerID="a35ff58cdedc1d36f8345528a64eca1b073d88b1dbf4edfd2f202894c24e07b0" exitCode=0 Feb 27 08:10:07 crc kubenswrapper[4612]: I0227 08:10:07.026175 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-l5z5s" event={"ID":"0c9482e8-9eb1-4c30-95e9-bdaa5ccd239b","Type":"ContainerDied","Data":"a35ff58cdedc1d36f8345528a64eca1b073d88b1dbf4edfd2f202894c24e07b0"} Feb 27 08:10:07 crc kubenswrapper[4612]: I0227 08:10:07.030466 4612 generic.go:334] "Generic (PLEG): container finished" podID="649e0df0-b37c-42f5-8af0-c804d4f467d4" containerID="399439f0cfe95ee5a2d4886d8c116dcf58ccc56b31529381b60a5bb7558643db" exitCode=0 Feb 27 08:10:07 crc kubenswrapper[4612]: I0227 08:10:07.030516 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5ddbbb8689-bm5cp" event={"ID":"649e0df0-b37c-42f5-8af0-c804d4f467d4","Type":"ContainerDied","Data":"399439f0cfe95ee5a2d4886d8c116dcf58ccc56b31529381b60a5bb7558643db"} Feb 27 08:10:07 crc kubenswrapper[4612]: I0227 08:10:07.584125 4612 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/neutron-5ddbbb8689-bm5cp" podUID="649e0df0-b37c-42f5-8af0-c804d4f467d4" containerName="neutron-httpd" probeResult="failure" output="Get \"https://10.217.0.157:9696/\": dial tcp 10.217.0.157:9696: connect: connection refused" Feb 27 08:10:08 crc kubenswrapper[4612]: I0227 08:10:08.295091 4612 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-55cd89c446-4c5hx" podUID="bb25e0d6-95cd-42a8-a182-f714a14d0084" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.165:9311/healthcheck\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Feb 27 08:10:09 crc kubenswrapper[4612]: I0227 08:10:09.109240 4612 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-6b985547d4-mkfpv" podUID="76009243-16b3-4652-abfe-a6cdba363724" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.152:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.152:8443: connect: connection refused" Feb 27 08:10:09 crc kubenswrapper[4612]: I0227 08:10:09.272498 4612 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-77948db5bb-97qh6" podUID="b72b541d-4661-44a7-a121-c2a8aef6db11" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.153:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.153:8443: connect: connection refused" Feb 27 08:10:09 crc kubenswrapper[4612]: I0227 08:10:09.824152 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-7db7f7dcc8-f95qr" Feb 27 08:10:10 crc kubenswrapper[4612]: I0227 08:10:10.433115 4612 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-6b7b667979-pm5fl" podUID="bf325571-0b74-447f-bd04-213d54a8437e" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.155:5353: connect: connection refused" Feb 27 08:10:11 crc kubenswrapper[4612]: I0227 08:10:11.089907 4612 generic.go:334] "Generic (PLEG): container finished" podID="649e0df0-b37c-42f5-8af0-c804d4f467d4" containerID="c8e24f4c991021623da34d4abac610956069b675ba545401121c750c3d4a1070" exitCode=0 Feb 27 08:10:11 crc kubenswrapper[4612]: I0227 08:10:11.089945 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5ddbbb8689-bm5cp" event={"ID":"649e0df0-b37c-42f5-8af0-c804d4f467d4","Type":"ContainerDied","Data":"c8e24f4c991021623da34d4abac610956069b675ba545401121c750c3d4a1070"} Feb 27 08:10:11 crc kubenswrapper[4612]: I0227 08:10:11.300949 4612 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-7db7f7dcc8-f95qr" podUID="16621861-feac-4972-ad1c-1e365692c23e" containerName="barbican-api-log" probeResult="failure" output="Get \"https://10.217.0.166:9311/healthcheck\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Feb 27 08:10:11 crc kubenswrapper[4612]: I0227 08:10:11.957851 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-55cd89c446-4c5hx" Feb 27 08:10:11 crc kubenswrapper[4612]: I0227 08:10:11.985915 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-l5z5s" Feb 27 08:10:12 crc kubenswrapper[4612]: I0227 08:10:12.054213 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0c9482e8-9eb1-4c30-95e9-bdaa5ccd239b-etc-machine-id\") pod \"0c9482e8-9eb1-4c30-95e9-bdaa5ccd239b\" (UID: \"0c9482e8-9eb1-4c30-95e9-bdaa5ccd239b\") " Feb 27 08:10:12 crc kubenswrapper[4612]: I0227 08:10:12.054357 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0c9482e8-9eb1-4c30-95e9-bdaa5ccd239b-config-data\") pod \"0c9482e8-9eb1-4c30-95e9-bdaa5ccd239b\" (UID: \"0c9482e8-9eb1-4c30-95e9-bdaa5ccd239b\") " Feb 27 08:10:12 crc kubenswrapper[4612]: I0227 08:10:12.054436 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fhfxv\" (UniqueName: \"kubernetes.io/projected/0c9482e8-9eb1-4c30-95e9-bdaa5ccd239b-kube-api-access-fhfxv\") pod \"0c9482e8-9eb1-4c30-95e9-bdaa5ccd239b\" (UID: \"0c9482e8-9eb1-4c30-95e9-bdaa5ccd239b\") " Feb 27 08:10:12 crc kubenswrapper[4612]: I0227 08:10:12.054476 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c9482e8-9eb1-4c30-95e9-bdaa5ccd239b-combined-ca-bundle\") pod \"0c9482e8-9eb1-4c30-95e9-bdaa5ccd239b\" (UID: \"0c9482e8-9eb1-4c30-95e9-bdaa5ccd239b\") " Feb 27 08:10:12 crc kubenswrapper[4612]: I0227 08:10:12.054501 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/0c9482e8-9eb1-4c30-95e9-bdaa5ccd239b-db-sync-config-data\") pod \"0c9482e8-9eb1-4c30-95e9-bdaa5ccd239b\" (UID: \"0c9482e8-9eb1-4c30-95e9-bdaa5ccd239b\") " Feb 27 08:10:12 crc kubenswrapper[4612]: I0227 08:10:12.054558 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0c9482e8-9eb1-4c30-95e9-bdaa5ccd239b-scripts\") pod \"0c9482e8-9eb1-4c30-95e9-bdaa5ccd239b\" (UID: \"0c9482e8-9eb1-4c30-95e9-bdaa5ccd239b\") " Feb 27 08:10:12 crc kubenswrapper[4612]: I0227 08:10:12.055953 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0c9482e8-9eb1-4c30-95e9-bdaa5ccd239b-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "0c9482e8-9eb1-4c30-95e9-bdaa5ccd239b" (UID: "0c9482e8-9eb1-4c30-95e9-bdaa5ccd239b"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 27 08:10:12 crc kubenswrapper[4612]: I0227 08:10:12.072868 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0c9482e8-9eb1-4c30-95e9-bdaa5ccd239b-scripts" (OuterVolumeSpecName: "scripts") pod "0c9482e8-9eb1-4c30-95e9-bdaa5ccd239b" (UID: "0c9482e8-9eb1-4c30-95e9-bdaa5ccd239b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:10:12 crc kubenswrapper[4612]: I0227 08:10:12.090369 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0c9482e8-9eb1-4c30-95e9-bdaa5ccd239b-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "0c9482e8-9eb1-4c30-95e9-bdaa5ccd239b" (UID: "0c9482e8-9eb1-4c30-95e9-bdaa5ccd239b"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:10:12 crc kubenswrapper[4612]: I0227 08:10:12.100902 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-l5z5s" event={"ID":"0c9482e8-9eb1-4c30-95e9-bdaa5ccd239b","Type":"ContainerDied","Data":"90344fc51bffd173e570a5b1aa79b9bda0567598d62ff08e647ad6b0b87af68a"} Feb 27 08:10:12 crc kubenswrapper[4612]: I0227 08:10:12.101197 4612 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="90344fc51bffd173e570a5b1aa79b9bda0567598d62ff08e647ad6b0b87af68a" Feb 27 08:10:12 crc kubenswrapper[4612]: I0227 08:10:12.101364 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-l5z5s" Feb 27 08:10:12 crc kubenswrapper[4612]: I0227 08:10:12.113449 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0c9482e8-9eb1-4c30-95e9-bdaa5ccd239b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0c9482e8-9eb1-4c30-95e9-bdaa5ccd239b" (UID: "0c9482e8-9eb1-4c30-95e9-bdaa5ccd239b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:10:12 crc kubenswrapper[4612]: I0227 08:10:12.120330 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0c9482e8-9eb1-4c30-95e9-bdaa5ccd239b-kube-api-access-fhfxv" (OuterVolumeSpecName: "kube-api-access-fhfxv") pod "0c9482e8-9eb1-4c30-95e9-bdaa5ccd239b" (UID: "0c9482e8-9eb1-4c30-95e9-bdaa5ccd239b"). InnerVolumeSpecName "kube-api-access-fhfxv". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 08:10:12 crc kubenswrapper[4612]: I0227 08:10:12.159506 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fhfxv\" (UniqueName: \"kubernetes.io/projected/0c9482e8-9eb1-4c30-95e9-bdaa5ccd239b-kube-api-access-fhfxv\") on node \"crc\" DevicePath \"\"" Feb 27 08:10:12 crc kubenswrapper[4612]: I0227 08:10:12.159551 4612 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c9482e8-9eb1-4c30-95e9-bdaa5ccd239b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 08:10:12 crc kubenswrapper[4612]: I0227 08:10:12.159564 4612 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/0c9482e8-9eb1-4c30-95e9-bdaa5ccd239b-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Feb 27 08:10:12 crc kubenswrapper[4612]: I0227 08:10:12.159576 4612 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0c9482e8-9eb1-4c30-95e9-bdaa5ccd239b-scripts\") on node \"crc\" DevicePath \"\"" Feb 27 08:10:12 crc kubenswrapper[4612]: I0227 08:10:12.159588 4612 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0c9482e8-9eb1-4c30-95e9-bdaa5ccd239b-etc-machine-id\") on node \"crc\" DevicePath \"\"" Feb 27 08:10:12 crc kubenswrapper[4612]: I0227 08:10:12.195932 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0c9482e8-9eb1-4c30-95e9-bdaa5ccd239b-config-data" (OuterVolumeSpecName: "config-data") pod "0c9482e8-9eb1-4c30-95e9-bdaa5ccd239b" (UID: "0c9482e8-9eb1-4c30-95e9-bdaa5ccd239b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:10:12 crc kubenswrapper[4612]: I0227 08:10:12.270144 4612 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0c9482e8-9eb1-4c30-95e9-bdaa5ccd239b-config-data\") on node \"crc\" DevicePath \"\"" Feb 27 08:10:13 crc kubenswrapper[4612]: E0227 08:10:13.007288 4612 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/ubi9/httpd-24:latest" Feb 27 08:10:13 crc kubenswrapper[4612]: E0227 08:10:13.007462 4612 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:proxy-httpd,Image:registry.redhat.io/ubi9/httpd-24:latest,Command:[/usr/sbin/httpd],Args:[-DFOREGROUND],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:proxy-httpd,HostPort:0,ContainerPort:3000,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/httpd/conf/httpd.conf,SubPath:httpd.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/httpd/conf.d/ssl.conf,SubPath:ssl.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:run-httpd,ReadOnly:false,MountPath:/run/httpd,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:log-httpd,ReadOnly:false,MountPath:/var/log/httpd,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-sh9th,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/,Port:{0 3000 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:300,TimeoutSeconds:30,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/,Port:{0 3000 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:10,TimeoutSeconds:30,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ceilometer-0_openstack(9ab32d5a-a797-47a8-961a-a534f55fdfa7): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Feb 27 08:10:13 crc kubenswrapper[4612]: E0227 08:10:13.009564 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"ceilometer-central-agent\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"proxy-httpd\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"]" pod="openstack/ceilometer-0" podUID="9ab32d5a-a797-47a8-961a-a534f55fdfa7" Feb 27 08:10:13 crc kubenswrapper[4612]: I0227 08:10:13.101300 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5ddbbb8689-bm5cp" Feb 27 08:10:13 crc kubenswrapper[4612]: I0227 08:10:13.113971 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b7b667979-pm5fl" Feb 27 08:10:13 crc kubenswrapper[4612]: I0227 08:10:13.115496 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b7b667979-pm5fl" event={"ID":"bf325571-0b74-447f-bd04-213d54a8437e","Type":"ContainerDied","Data":"26d204ca3654b872870f1ae7b2046125ba43c32433373e33bbf690ab1087c66c"} Feb 27 08:10:13 crc kubenswrapper[4612]: I0227 08:10:13.115563 4612 scope.go:117] "RemoveContainer" containerID="cd920bf9cdd1f21efee933e8fb8dc84fd5aac93ceafa5fc01745e88c2e9aaf86" Feb 27 08:10:13 crc kubenswrapper[4612]: I0227 08:10:13.122309 4612 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="9ab32d5a-a797-47a8-961a-a534f55fdfa7" containerName="ceilometer-notification-agent" containerID="cri-o://0b87c11e9c758edb6f988ef0ee90e9d6d19fa4c1d46725a5fdb737090d12c6be" gracePeriod=30 Feb 27 08:10:13 crc kubenswrapper[4612]: I0227 08:10:13.122981 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5ddbbb8689-bm5cp" Feb 27 08:10:13 crc kubenswrapper[4612]: I0227 08:10:13.123804 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5ddbbb8689-bm5cp" event={"ID":"649e0df0-b37c-42f5-8af0-c804d4f467d4","Type":"ContainerDied","Data":"0974569712b15ea4dcc9022105f7deb4a6621d88d67b162ab1e1ec41c914fe97"} Feb 27 08:10:13 crc kubenswrapper[4612]: I0227 08:10:13.123928 4612 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="9ab32d5a-a797-47a8-961a-a534f55fdfa7" containerName="sg-core" containerID="cri-o://e77b56aa24d4104786d48b89f361df289c7f02205d97f075d23ee055ec70f239" gracePeriod=30 Feb 27 08:10:13 crc kubenswrapper[4612]: I0227 08:10:13.193245 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/649e0df0-b37c-42f5-8af0-c804d4f467d4-httpd-config\") pod \"649e0df0-b37c-42f5-8af0-c804d4f467d4\" (UID: \"649e0df0-b37c-42f5-8af0-c804d4f467d4\") " Feb 27 08:10:13 crc kubenswrapper[4612]: I0227 08:10:13.193516 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bf325571-0b74-447f-bd04-213d54a8437e-config\") pod \"bf325571-0b74-447f-bd04-213d54a8437e\" (UID: \"bf325571-0b74-447f-bd04-213d54a8437e\") " Feb 27 08:10:13 crc kubenswrapper[4612]: I0227 08:10:13.193538 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8zzrm\" (UniqueName: \"kubernetes.io/projected/bf325571-0b74-447f-bd04-213d54a8437e-kube-api-access-8zzrm\") pod \"bf325571-0b74-447f-bd04-213d54a8437e\" (UID: \"bf325571-0b74-447f-bd04-213d54a8437e\") " Feb 27 08:10:13 crc kubenswrapper[4612]: I0227 08:10:13.193565 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/649e0df0-b37c-42f5-8af0-c804d4f467d4-ovndb-tls-certs\") pod \"649e0df0-b37c-42f5-8af0-c804d4f467d4\" (UID: \"649e0df0-b37c-42f5-8af0-c804d4f467d4\") " Feb 27 08:10:13 crc kubenswrapper[4612]: I0227 08:10:13.193635 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bf325571-0b74-447f-bd04-213d54a8437e-ovsdbserver-nb\") pod \"bf325571-0b74-447f-bd04-213d54a8437e\" (UID: \"bf325571-0b74-447f-bd04-213d54a8437e\") " Feb 27 08:10:13 crc kubenswrapper[4612]: I0227 08:10:13.193676 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/649e0df0-b37c-42f5-8af0-c804d4f467d4-combined-ca-bundle\") pod \"649e0df0-b37c-42f5-8af0-c804d4f467d4\" (UID: \"649e0df0-b37c-42f5-8af0-c804d4f467d4\") " Feb 27 08:10:13 crc kubenswrapper[4612]: I0227 08:10:13.193800 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bf325571-0b74-447f-bd04-213d54a8437e-dns-svc\") pod \"bf325571-0b74-447f-bd04-213d54a8437e\" (UID: \"bf325571-0b74-447f-bd04-213d54a8437e\") " Feb 27 08:10:13 crc kubenswrapper[4612]: I0227 08:10:13.193827 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/bf325571-0b74-447f-bd04-213d54a8437e-dns-swift-storage-0\") pod \"bf325571-0b74-447f-bd04-213d54a8437e\" (UID: \"bf325571-0b74-447f-bd04-213d54a8437e\") " Feb 27 08:10:13 crc kubenswrapper[4612]: I0227 08:10:13.193842 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/649e0df0-b37c-42f5-8af0-c804d4f467d4-public-tls-certs\") pod \"649e0df0-b37c-42f5-8af0-c804d4f467d4\" (UID: \"649e0df0-b37c-42f5-8af0-c804d4f467d4\") " Feb 27 08:10:13 crc kubenswrapper[4612]: I0227 08:10:13.193864 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kl8ws\" (UniqueName: \"kubernetes.io/projected/649e0df0-b37c-42f5-8af0-c804d4f467d4-kube-api-access-kl8ws\") pod \"649e0df0-b37c-42f5-8af0-c804d4f467d4\" (UID: \"649e0df0-b37c-42f5-8af0-c804d4f467d4\") " Feb 27 08:10:13 crc kubenswrapper[4612]: I0227 08:10:13.193901 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/649e0df0-b37c-42f5-8af0-c804d4f467d4-internal-tls-certs\") pod \"649e0df0-b37c-42f5-8af0-c804d4f467d4\" (UID: \"649e0df0-b37c-42f5-8af0-c804d4f467d4\") " Feb 27 08:10:13 crc kubenswrapper[4612]: I0227 08:10:13.193925 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bf325571-0b74-447f-bd04-213d54a8437e-ovsdbserver-sb\") pod \"bf325571-0b74-447f-bd04-213d54a8437e\" (UID: \"bf325571-0b74-447f-bd04-213d54a8437e\") " Feb 27 08:10:13 crc kubenswrapper[4612]: I0227 08:10:13.193979 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/649e0df0-b37c-42f5-8af0-c804d4f467d4-config\") pod \"649e0df0-b37c-42f5-8af0-c804d4f467d4\" (UID: \"649e0df0-b37c-42f5-8af0-c804d4f467d4\") " Feb 27 08:10:13 crc kubenswrapper[4612]: I0227 08:10:13.228269 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf325571-0b74-447f-bd04-213d54a8437e-kube-api-access-8zzrm" (OuterVolumeSpecName: "kube-api-access-8zzrm") pod "bf325571-0b74-447f-bd04-213d54a8437e" (UID: "bf325571-0b74-447f-bd04-213d54a8437e"). InnerVolumeSpecName "kube-api-access-8zzrm". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 08:10:13 crc kubenswrapper[4612]: I0227 08:10:13.252302 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/649e0df0-b37c-42f5-8af0-c804d4f467d4-kube-api-access-kl8ws" (OuterVolumeSpecName: "kube-api-access-kl8ws") pod "649e0df0-b37c-42f5-8af0-c804d4f467d4" (UID: "649e0df0-b37c-42f5-8af0-c804d4f467d4"). InnerVolumeSpecName "kube-api-access-kl8ws". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 08:10:13 crc kubenswrapper[4612]: I0227 08:10:13.297892 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kl8ws\" (UniqueName: \"kubernetes.io/projected/649e0df0-b37c-42f5-8af0-c804d4f467d4-kube-api-access-kl8ws\") on node \"crc\" DevicePath \"\"" Feb 27 08:10:13 crc kubenswrapper[4612]: I0227 08:10:13.297916 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8zzrm\" (UniqueName: \"kubernetes.io/projected/bf325571-0b74-447f-bd04-213d54a8437e-kube-api-access-8zzrm\") on node \"crc\" DevicePath \"\"" Feb 27 08:10:13 crc kubenswrapper[4612]: I0227 08:10:13.313900 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/649e0df0-b37c-42f5-8af0-c804d4f467d4-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "649e0df0-b37c-42f5-8af0-c804d4f467d4" (UID: "649e0df0-b37c-42f5-8af0-c804d4f467d4"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:10:13 crc kubenswrapper[4612]: I0227 08:10:13.321138 4612 scope.go:117] "RemoveContainer" containerID="273c781089e6725f6eb0a109249acabcde54e129dd2aef517f776687ba0f1701" Feb 27 08:10:13 crc kubenswrapper[4612]: I0227 08:10:13.418678 4612 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/649e0df0-b37c-42f5-8af0-c804d4f467d4-httpd-config\") on node \"crc\" DevicePath \"\"" Feb 27 08:10:13 crc kubenswrapper[4612]: I0227 08:10:13.512964 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/649e0df0-b37c-42f5-8af0-c804d4f467d4-config" (OuterVolumeSpecName: "config") pod "649e0df0-b37c-42f5-8af0-c804d4f467d4" (UID: "649e0df0-b37c-42f5-8af0-c804d4f467d4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:10:13 crc kubenswrapper[4612]: I0227 08:10:13.518777 4612 scope.go:117] "RemoveContainer" containerID="399439f0cfe95ee5a2d4886d8c116dcf58ccc56b31529381b60a5bb7558643db" Feb 27 08:10:13 crc kubenswrapper[4612]: I0227 08:10:13.520101 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Feb 27 08:10:13 crc kubenswrapper[4612]: E0227 08:10:13.539537 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf325571-0b74-447f-bd04-213d54a8437e" containerName="init" Feb 27 08:10:13 crc kubenswrapper[4612]: I0227 08:10:13.539568 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf325571-0b74-447f-bd04-213d54a8437e" containerName="init" Feb 27 08:10:13 crc kubenswrapper[4612]: E0227 08:10:13.539607 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="649e0df0-b37c-42f5-8af0-c804d4f467d4" containerName="neutron-api" Feb 27 08:10:13 crc kubenswrapper[4612]: I0227 08:10:13.539613 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="649e0df0-b37c-42f5-8af0-c804d4f467d4" containerName="neutron-api" Feb 27 08:10:13 crc kubenswrapper[4612]: E0227 08:10:13.539641 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf325571-0b74-447f-bd04-213d54a8437e" containerName="dnsmasq-dns" Feb 27 08:10:13 crc kubenswrapper[4612]: I0227 08:10:13.539646 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf325571-0b74-447f-bd04-213d54a8437e" containerName="dnsmasq-dns" Feb 27 08:10:13 crc kubenswrapper[4612]: E0227 08:10:13.539658 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c9482e8-9eb1-4c30-95e9-bdaa5ccd239b" containerName="cinder-db-sync" Feb 27 08:10:13 crc kubenswrapper[4612]: I0227 08:10:13.539664 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c9482e8-9eb1-4c30-95e9-bdaa5ccd239b" containerName="cinder-db-sync" Feb 27 08:10:13 crc kubenswrapper[4612]: E0227 08:10:13.539703 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="649e0df0-b37c-42f5-8af0-c804d4f467d4" containerName="neutron-httpd" Feb 27 08:10:13 crc kubenswrapper[4612]: I0227 08:10:13.539709 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="649e0df0-b37c-42f5-8af0-c804d4f467d4" containerName="neutron-httpd" Feb 27 08:10:13 crc kubenswrapper[4612]: I0227 08:10:13.539920 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="649e0df0-b37c-42f5-8af0-c804d4f467d4" containerName="neutron-httpd" Feb 27 08:10:13 crc kubenswrapper[4612]: I0227 08:10:13.539938 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="0c9482e8-9eb1-4c30-95e9-bdaa5ccd239b" containerName="cinder-db-sync" Feb 27 08:10:13 crc kubenswrapper[4612]: I0227 08:10:13.539949 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="bf325571-0b74-447f-bd04-213d54a8437e" containerName="dnsmasq-dns" Feb 27 08:10:13 crc kubenswrapper[4612]: I0227 08:10:13.539961 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="649e0df0-b37c-42f5-8af0-c804d4f467d4" containerName="neutron-api" Feb 27 08:10:13 crc kubenswrapper[4612]: I0227 08:10:13.540327 4612 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/649e0df0-b37c-42f5-8af0-c804d4f467d4-config\") on node \"crc\" DevicePath \"\"" Feb 27 08:10:13 crc kubenswrapper[4612]: I0227 08:10:13.541185 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Feb 27 08:10:13 crc kubenswrapper[4612]: I0227 08:10:13.553116 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-5v76w" Feb 27 08:10:13 crc kubenswrapper[4612]: I0227 08:10:13.556149 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Feb 27 08:10:13 crc kubenswrapper[4612]: I0227 08:10:13.561235 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Feb 27 08:10:13 crc kubenswrapper[4612]: I0227 08:10:13.561268 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Feb 27 08:10:13 crc kubenswrapper[4612]: I0227 08:10:13.590968 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Feb 27 08:10:13 crc kubenswrapper[4612]: I0227 08:10:13.605708 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf325571-0b74-447f-bd04-213d54a8437e-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "bf325571-0b74-447f-bd04-213d54a8437e" (UID: "bf325571-0b74-447f-bd04-213d54a8437e"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 08:10:13 crc kubenswrapper[4612]: I0227 08:10:13.644040 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-spf6v\" (UniqueName: \"kubernetes.io/projected/62b23ba9-5e68-4665-86cd-df78c13de661-kube-api-access-spf6v\") pod \"cinder-scheduler-0\" (UID: \"62b23ba9-5e68-4665-86cd-df78c13de661\") " pod="openstack/cinder-scheduler-0" Feb 27 08:10:13 crc kubenswrapper[4612]: I0227 08:10:13.644093 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/62b23ba9-5e68-4665-86cd-df78c13de661-scripts\") pod \"cinder-scheduler-0\" (UID: \"62b23ba9-5e68-4665-86cd-df78c13de661\") " pod="openstack/cinder-scheduler-0" Feb 27 08:10:13 crc kubenswrapper[4612]: I0227 08:10:13.644170 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/62b23ba9-5e68-4665-86cd-df78c13de661-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"62b23ba9-5e68-4665-86cd-df78c13de661\") " pod="openstack/cinder-scheduler-0" Feb 27 08:10:13 crc kubenswrapper[4612]: I0227 08:10:13.644190 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62b23ba9-5e68-4665-86cd-df78c13de661-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"62b23ba9-5e68-4665-86cd-df78c13de661\") " pod="openstack/cinder-scheduler-0" Feb 27 08:10:13 crc kubenswrapper[4612]: I0227 08:10:13.644213 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/62b23ba9-5e68-4665-86cd-df78c13de661-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"62b23ba9-5e68-4665-86cd-df78c13de661\") " pod="openstack/cinder-scheduler-0" Feb 27 08:10:13 crc kubenswrapper[4612]: I0227 08:10:13.644244 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/62b23ba9-5e68-4665-86cd-df78c13de661-config-data\") pod \"cinder-scheduler-0\" (UID: \"62b23ba9-5e68-4665-86cd-df78c13de661\") " pod="openstack/cinder-scheduler-0" Feb 27 08:10:13 crc kubenswrapper[4612]: I0227 08:10:13.644286 4612 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/bf325571-0b74-447f-bd04-213d54a8437e-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Feb 27 08:10:13 crc kubenswrapper[4612]: I0227 08:10:13.662892 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf325571-0b74-447f-bd04-213d54a8437e-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "bf325571-0b74-447f-bd04-213d54a8437e" (UID: "bf325571-0b74-447f-bd04-213d54a8437e"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 08:10:13 crc kubenswrapper[4612]: I0227 08:10:13.663156 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6578955fd5-ds7jf"] Feb 27 08:10:13 crc kubenswrapper[4612]: I0227 08:10:13.664551 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6578955fd5-ds7jf" Feb 27 08:10:13 crc kubenswrapper[4612]: I0227 08:10:13.672122 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/649e0df0-b37c-42f5-8af0-c804d4f467d4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "649e0df0-b37c-42f5-8af0-c804d4f467d4" (UID: "649e0df0-b37c-42f5-8af0-c804d4f467d4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:10:13 crc kubenswrapper[4612]: I0227 08:10:13.686579 4612 scope.go:117] "RemoveContainer" containerID="c8e24f4c991021623da34d4abac610956069b675ba545401121c750c3d4a1070" Feb 27 08:10:13 crc kubenswrapper[4612]: I0227 08:10:13.709865 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/649e0df0-b37c-42f5-8af0-c804d4f467d4-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "649e0df0-b37c-42f5-8af0-c804d4f467d4" (UID: "649e0df0-b37c-42f5-8af0-c804d4f467d4"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:10:13 crc kubenswrapper[4612]: I0227 08:10:13.745979 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/62b23ba9-5e68-4665-86cd-df78c13de661-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"62b23ba9-5e68-4665-86cd-df78c13de661\") " pod="openstack/cinder-scheduler-0" Feb 27 08:10:13 crc kubenswrapper[4612]: I0227 08:10:13.746028 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/aedb2ee7-269c-49e6-9bf7-8ac34d6c1d48-dns-svc\") pod \"dnsmasq-dns-6578955fd5-ds7jf\" (UID: \"aedb2ee7-269c-49e6-9bf7-8ac34d6c1d48\") " pod="openstack/dnsmasq-dns-6578955fd5-ds7jf" Feb 27 08:10:13 crc kubenswrapper[4612]: I0227 08:10:13.746058 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62b23ba9-5e68-4665-86cd-df78c13de661-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"62b23ba9-5e68-4665-86cd-df78c13de661\") " pod="openstack/cinder-scheduler-0" Feb 27 08:10:13 crc kubenswrapper[4612]: I0227 08:10:13.746076 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/aedb2ee7-269c-49e6-9bf7-8ac34d6c1d48-ovsdbserver-nb\") pod \"dnsmasq-dns-6578955fd5-ds7jf\" (UID: \"aedb2ee7-269c-49e6-9bf7-8ac34d6c1d48\") " pod="openstack/dnsmasq-dns-6578955fd5-ds7jf" Feb 27 08:10:13 crc kubenswrapper[4612]: I0227 08:10:13.746106 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/aedb2ee7-269c-49e6-9bf7-8ac34d6c1d48-ovsdbserver-sb\") pod \"dnsmasq-dns-6578955fd5-ds7jf\" (UID: \"aedb2ee7-269c-49e6-9bf7-8ac34d6c1d48\") " pod="openstack/dnsmasq-dns-6578955fd5-ds7jf" Feb 27 08:10:13 crc kubenswrapper[4612]: I0227 08:10:13.746130 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/62b23ba9-5e68-4665-86cd-df78c13de661-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"62b23ba9-5e68-4665-86cd-df78c13de661\") " pod="openstack/cinder-scheduler-0" Feb 27 08:10:13 crc kubenswrapper[4612]: I0227 08:10:13.746190 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aedb2ee7-269c-49e6-9bf7-8ac34d6c1d48-config\") pod \"dnsmasq-dns-6578955fd5-ds7jf\" (UID: \"aedb2ee7-269c-49e6-9bf7-8ac34d6c1d48\") " pod="openstack/dnsmasq-dns-6578955fd5-ds7jf" Feb 27 08:10:13 crc kubenswrapper[4612]: I0227 08:10:13.746217 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/62b23ba9-5e68-4665-86cd-df78c13de661-config-data\") pod \"cinder-scheduler-0\" (UID: \"62b23ba9-5e68-4665-86cd-df78c13de661\") " pod="openstack/cinder-scheduler-0" Feb 27 08:10:13 crc kubenswrapper[4612]: I0227 08:10:13.746341 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-spf6v\" (UniqueName: \"kubernetes.io/projected/62b23ba9-5e68-4665-86cd-df78c13de661-kube-api-access-spf6v\") pod \"cinder-scheduler-0\" (UID: \"62b23ba9-5e68-4665-86cd-df78c13de661\") " pod="openstack/cinder-scheduler-0" Feb 27 08:10:13 crc kubenswrapper[4612]: I0227 08:10:13.746390 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/aedb2ee7-269c-49e6-9bf7-8ac34d6c1d48-dns-swift-storage-0\") pod \"dnsmasq-dns-6578955fd5-ds7jf\" (UID: \"aedb2ee7-269c-49e6-9bf7-8ac34d6c1d48\") " pod="openstack/dnsmasq-dns-6578955fd5-ds7jf" Feb 27 08:10:13 crc kubenswrapper[4612]: I0227 08:10:13.746415 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/62b23ba9-5e68-4665-86cd-df78c13de661-scripts\") pod \"cinder-scheduler-0\" (UID: \"62b23ba9-5e68-4665-86cd-df78c13de661\") " pod="openstack/cinder-scheduler-0" Feb 27 08:10:13 crc kubenswrapper[4612]: I0227 08:10:13.746482 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nvldg\" (UniqueName: \"kubernetes.io/projected/aedb2ee7-269c-49e6-9bf7-8ac34d6c1d48-kube-api-access-nvldg\") pod \"dnsmasq-dns-6578955fd5-ds7jf\" (UID: \"aedb2ee7-269c-49e6-9bf7-8ac34d6c1d48\") " pod="openstack/dnsmasq-dns-6578955fd5-ds7jf" Feb 27 08:10:13 crc kubenswrapper[4612]: I0227 08:10:13.746620 4612 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/649e0df0-b37c-42f5-8af0-c804d4f467d4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 08:10:13 crc kubenswrapper[4612]: I0227 08:10:13.746639 4612 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bf325571-0b74-447f-bd04-213d54a8437e-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 27 08:10:13 crc kubenswrapper[4612]: I0227 08:10:13.746652 4612 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/649e0df0-b37c-42f5-8af0-c804d4f467d4-public-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 27 08:10:13 crc kubenswrapper[4612]: I0227 08:10:13.746729 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/62b23ba9-5e68-4665-86cd-df78c13de661-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"62b23ba9-5e68-4665-86cd-df78c13de661\") " pod="openstack/cinder-scheduler-0" Feb 27 08:10:13 crc kubenswrapper[4612]: I0227 08:10:13.758346 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/62b23ba9-5e68-4665-86cd-df78c13de661-config-data\") pod \"cinder-scheduler-0\" (UID: \"62b23ba9-5e68-4665-86cd-df78c13de661\") " pod="openstack/cinder-scheduler-0" Feb 27 08:10:13 crc kubenswrapper[4612]: I0227 08:10:13.760366 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf325571-0b74-447f-bd04-213d54a8437e-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "bf325571-0b74-447f-bd04-213d54a8437e" (UID: "bf325571-0b74-447f-bd04-213d54a8437e"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 08:10:13 crc kubenswrapper[4612]: I0227 08:10:13.778413 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/649e0df0-b37c-42f5-8af0-c804d4f467d4-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "649e0df0-b37c-42f5-8af0-c804d4f467d4" (UID: "649e0df0-b37c-42f5-8af0-c804d4f467d4"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:10:13 crc kubenswrapper[4612]: I0227 08:10:13.781898 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/62b23ba9-5e68-4665-86cd-df78c13de661-scripts\") pod \"cinder-scheduler-0\" (UID: \"62b23ba9-5e68-4665-86cd-df78c13de661\") " pod="openstack/cinder-scheduler-0" Feb 27 08:10:13 crc kubenswrapper[4612]: I0227 08:10:13.783142 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62b23ba9-5e68-4665-86cd-df78c13de661-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"62b23ba9-5e68-4665-86cd-df78c13de661\") " pod="openstack/cinder-scheduler-0" Feb 27 08:10:13 crc kubenswrapper[4612]: I0227 08:10:13.788329 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6578955fd5-ds7jf"] Feb 27 08:10:13 crc kubenswrapper[4612]: I0227 08:10:13.789027 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-spf6v\" (UniqueName: \"kubernetes.io/projected/62b23ba9-5e68-4665-86cd-df78c13de661-kube-api-access-spf6v\") pod \"cinder-scheduler-0\" (UID: \"62b23ba9-5e68-4665-86cd-df78c13de661\") " pod="openstack/cinder-scheduler-0" Feb 27 08:10:13 crc kubenswrapper[4612]: I0227 08:10:13.789169 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/62b23ba9-5e68-4665-86cd-df78c13de661-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"62b23ba9-5e68-4665-86cd-df78c13de661\") " pod="openstack/cinder-scheduler-0" Feb 27 08:10:13 crc kubenswrapper[4612]: I0227 08:10:13.848385 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/aedb2ee7-269c-49e6-9bf7-8ac34d6c1d48-dns-svc\") pod \"dnsmasq-dns-6578955fd5-ds7jf\" (UID: \"aedb2ee7-269c-49e6-9bf7-8ac34d6c1d48\") " pod="openstack/dnsmasq-dns-6578955fd5-ds7jf" Feb 27 08:10:13 crc kubenswrapper[4612]: I0227 08:10:13.848428 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/aedb2ee7-269c-49e6-9bf7-8ac34d6c1d48-ovsdbserver-nb\") pod \"dnsmasq-dns-6578955fd5-ds7jf\" (UID: \"aedb2ee7-269c-49e6-9bf7-8ac34d6c1d48\") " pod="openstack/dnsmasq-dns-6578955fd5-ds7jf" Feb 27 08:10:13 crc kubenswrapper[4612]: I0227 08:10:13.848449 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/aedb2ee7-269c-49e6-9bf7-8ac34d6c1d48-ovsdbserver-sb\") pod \"dnsmasq-dns-6578955fd5-ds7jf\" (UID: \"aedb2ee7-269c-49e6-9bf7-8ac34d6c1d48\") " pod="openstack/dnsmasq-dns-6578955fd5-ds7jf" Feb 27 08:10:13 crc kubenswrapper[4612]: I0227 08:10:13.848483 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aedb2ee7-269c-49e6-9bf7-8ac34d6c1d48-config\") pod \"dnsmasq-dns-6578955fd5-ds7jf\" (UID: \"aedb2ee7-269c-49e6-9bf7-8ac34d6c1d48\") " pod="openstack/dnsmasq-dns-6578955fd5-ds7jf" Feb 27 08:10:13 crc kubenswrapper[4612]: I0227 08:10:13.848550 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/aedb2ee7-269c-49e6-9bf7-8ac34d6c1d48-dns-swift-storage-0\") pod \"dnsmasq-dns-6578955fd5-ds7jf\" (UID: \"aedb2ee7-269c-49e6-9bf7-8ac34d6c1d48\") " pod="openstack/dnsmasq-dns-6578955fd5-ds7jf" Feb 27 08:10:13 crc kubenswrapper[4612]: I0227 08:10:13.848588 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nvldg\" (UniqueName: \"kubernetes.io/projected/aedb2ee7-269c-49e6-9bf7-8ac34d6c1d48-kube-api-access-nvldg\") pod \"dnsmasq-dns-6578955fd5-ds7jf\" (UID: \"aedb2ee7-269c-49e6-9bf7-8ac34d6c1d48\") " pod="openstack/dnsmasq-dns-6578955fd5-ds7jf" Feb 27 08:10:13 crc kubenswrapper[4612]: I0227 08:10:13.848634 4612 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/649e0df0-b37c-42f5-8af0-c804d4f467d4-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 27 08:10:13 crc kubenswrapper[4612]: I0227 08:10:13.848645 4612 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bf325571-0b74-447f-bd04-213d54a8437e-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 27 08:10:13 crc kubenswrapper[4612]: I0227 08:10:13.849579 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/aedb2ee7-269c-49e6-9bf7-8ac34d6c1d48-ovsdbserver-nb\") pod \"dnsmasq-dns-6578955fd5-ds7jf\" (UID: \"aedb2ee7-269c-49e6-9bf7-8ac34d6c1d48\") " pod="openstack/dnsmasq-dns-6578955fd5-ds7jf" Feb 27 08:10:13 crc kubenswrapper[4612]: I0227 08:10:13.849713 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/aedb2ee7-269c-49e6-9bf7-8ac34d6c1d48-dns-svc\") pod \"dnsmasq-dns-6578955fd5-ds7jf\" (UID: \"aedb2ee7-269c-49e6-9bf7-8ac34d6c1d48\") " pod="openstack/dnsmasq-dns-6578955fd5-ds7jf" Feb 27 08:10:13 crc kubenswrapper[4612]: I0227 08:10:13.850117 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/aedb2ee7-269c-49e6-9bf7-8ac34d6c1d48-ovsdbserver-sb\") pod \"dnsmasq-dns-6578955fd5-ds7jf\" (UID: \"aedb2ee7-269c-49e6-9bf7-8ac34d6c1d48\") " pod="openstack/dnsmasq-dns-6578955fd5-ds7jf" Feb 27 08:10:13 crc kubenswrapper[4612]: I0227 08:10:13.850600 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aedb2ee7-269c-49e6-9bf7-8ac34d6c1d48-config\") pod \"dnsmasq-dns-6578955fd5-ds7jf\" (UID: \"aedb2ee7-269c-49e6-9bf7-8ac34d6c1d48\") " pod="openstack/dnsmasq-dns-6578955fd5-ds7jf" Feb 27 08:10:13 crc kubenswrapper[4612]: I0227 08:10:13.850636 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/aedb2ee7-269c-49e6-9bf7-8ac34d6c1d48-dns-swift-storage-0\") pod \"dnsmasq-dns-6578955fd5-ds7jf\" (UID: \"aedb2ee7-269c-49e6-9bf7-8ac34d6c1d48\") " pod="openstack/dnsmasq-dns-6578955fd5-ds7jf" Feb 27 08:10:13 crc kubenswrapper[4612]: I0227 08:10:13.853562 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf325571-0b74-447f-bd04-213d54a8437e-config" (OuterVolumeSpecName: "config") pod "bf325571-0b74-447f-bd04-213d54a8437e" (UID: "bf325571-0b74-447f-bd04-213d54a8437e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 08:10:13 crc kubenswrapper[4612]: I0227 08:10:13.871255 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf325571-0b74-447f-bd04-213d54a8437e-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "bf325571-0b74-447f-bd04-213d54a8437e" (UID: "bf325571-0b74-447f-bd04-213d54a8437e"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 08:10:13 crc kubenswrapper[4612]: I0227 08:10:13.887512 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nvldg\" (UniqueName: \"kubernetes.io/projected/aedb2ee7-269c-49e6-9bf7-8ac34d6c1d48-kube-api-access-nvldg\") pod \"dnsmasq-dns-6578955fd5-ds7jf\" (UID: \"aedb2ee7-269c-49e6-9bf7-8ac34d6c1d48\") " pod="openstack/dnsmasq-dns-6578955fd5-ds7jf" Feb 27 08:10:13 crc kubenswrapper[4612]: I0227 08:10:13.889660 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/649e0df0-b37c-42f5-8af0-c804d4f467d4-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "649e0df0-b37c-42f5-8af0-c804d4f467d4" (UID: "649e0df0-b37c-42f5-8af0-c804d4f467d4"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:10:13 crc kubenswrapper[4612]: I0227 08:10:13.901504 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Feb 27 08:10:13 crc kubenswrapper[4612]: I0227 08:10:13.909657 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Feb 27 08:10:13 crc kubenswrapper[4612]: I0227 08:10:13.911767 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Feb 27 08:10:13 crc kubenswrapper[4612]: I0227 08:10:13.913999 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Feb 27 08:10:13 crc kubenswrapper[4612]: I0227 08:10:13.948914 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Feb 27 08:10:13 crc kubenswrapper[4612]: I0227 08:10:13.958402 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9a40110d-5675-40b6-87c6-63b1d2faba9d-scripts\") pod \"cinder-api-0\" (UID: \"9a40110d-5675-40b6-87c6-63b1d2faba9d\") " pod="openstack/cinder-api-0" Feb 27 08:10:13 crc kubenswrapper[4612]: I0227 08:10:13.958642 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/9a40110d-5675-40b6-87c6-63b1d2faba9d-etc-machine-id\") pod \"cinder-api-0\" (UID: \"9a40110d-5675-40b6-87c6-63b1d2faba9d\") " pod="openstack/cinder-api-0" Feb 27 08:10:13 crc kubenswrapper[4612]: I0227 08:10:13.958715 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9a40110d-5675-40b6-87c6-63b1d2faba9d-logs\") pod \"cinder-api-0\" (UID: \"9a40110d-5675-40b6-87c6-63b1d2faba9d\") " pod="openstack/cinder-api-0" Feb 27 08:10:13 crc kubenswrapper[4612]: I0227 08:10:13.958788 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9a40110d-5675-40b6-87c6-63b1d2faba9d-config-data\") pod \"cinder-api-0\" (UID: \"9a40110d-5675-40b6-87c6-63b1d2faba9d\") " pod="openstack/cinder-api-0" Feb 27 08:10:13 crc kubenswrapper[4612]: I0227 08:10:13.958900 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nkgbh\" (UniqueName: \"kubernetes.io/projected/9a40110d-5675-40b6-87c6-63b1d2faba9d-kube-api-access-nkgbh\") pod \"cinder-api-0\" (UID: \"9a40110d-5675-40b6-87c6-63b1d2faba9d\") " pod="openstack/cinder-api-0" Feb 27 08:10:13 crc kubenswrapper[4612]: I0227 08:10:13.958941 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9a40110d-5675-40b6-87c6-63b1d2faba9d-config-data-custom\") pod \"cinder-api-0\" (UID: \"9a40110d-5675-40b6-87c6-63b1d2faba9d\") " pod="openstack/cinder-api-0" Feb 27 08:10:13 crc kubenswrapper[4612]: I0227 08:10:13.958968 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9a40110d-5675-40b6-87c6-63b1d2faba9d-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"9a40110d-5675-40b6-87c6-63b1d2faba9d\") " pod="openstack/cinder-api-0" Feb 27 08:10:13 crc kubenswrapper[4612]: I0227 08:10:13.959104 4612 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bf325571-0b74-447f-bd04-213d54a8437e-config\") on node \"crc\" DevicePath \"\"" Feb 27 08:10:13 crc kubenswrapper[4612]: I0227 08:10:13.959116 4612 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/649e0df0-b37c-42f5-8af0-c804d4f467d4-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 27 08:10:13 crc kubenswrapper[4612]: I0227 08:10:13.959126 4612 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bf325571-0b74-447f-bd04-213d54a8437e-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 27 08:10:14 crc kubenswrapper[4612]: I0227 08:10:14.041363 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6578955fd5-ds7jf" Feb 27 08:10:14 crc kubenswrapper[4612]: I0227 08:10:14.045444 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-7db7f7dcc8-f95qr" Feb 27 08:10:14 crc kubenswrapper[4612]: I0227 08:10:14.063366 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nkgbh\" (UniqueName: \"kubernetes.io/projected/9a40110d-5675-40b6-87c6-63b1d2faba9d-kube-api-access-nkgbh\") pod \"cinder-api-0\" (UID: \"9a40110d-5675-40b6-87c6-63b1d2faba9d\") " pod="openstack/cinder-api-0" Feb 27 08:10:14 crc kubenswrapper[4612]: I0227 08:10:14.063411 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9a40110d-5675-40b6-87c6-63b1d2faba9d-config-data-custom\") pod \"cinder-api-0\" (UID: \"9a40110d-5675-40b6-87c6-63b1d2faba9d\") " pod="openstack/cinder-api-0" Feb 27 08:10:14 crc kubenswrapper[4612]: I0227 08:10:14.063436 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9a40110d-5675-40b6-87c6-63b1d2faba9d-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"9a40110d-5675-40b6-87c6-63b1d2faba9d\") " pod="openstack/cinder-api-0" Feb 27 08:10:14 crc kubenswrapper[4612]: I0227 08:10:14.063485 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9a40110d-5675-40b6-87c6-63b1d2faba9d-scripts\") pod \"cinder-api-0\" (UID: \"9a40110d-5675-40b6-87c6-63b1d2faba9d\") " pod="openstack/cinder-api-0" Feb 27 08:10:14 crc kubenswrapper[4612]: I0227 08:10:14.063627 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/9a40110d-5675-40b6-87c6-63b1d2faba9d-etc-machine-id\") pod \"cinder-api-0\" (UID: \"9a40110d-5675-40b6-87c6-63b1d2faba9d\") " pod="openstack/cinder-api-0" Feb 27 08:10:14 crc kubenswrapper[4612]: I0227 08:10:14.063651 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9a40110d-5675-40b6-87c6-63b1d2faba9d-logs\") pod \"cinder-api-0\" (UID: \"9a40110d-5675-40b6-87c6-63b1d2faba9d\") " pod="openstack/cinder-api-0" Feb 27 08:10:14 crc kubenswrapper[4612]: I0227 08:10:14.063716 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9a40110d-5675-40b6-87c6-63b1d2faba9d-config-data\") pod \"cinder-api-0\" (UID: \"9a40110d-5675-40b6-87c6-63b1d2faba9d\") " pod="openstack/cinder-api-0" Feb 27 08:10:14 crc kubenswrapper[4612]: I0227 08:10:14.107980 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9a40110d-5675-40b6-87c6-63b1d2faba9d-config-data-custom\") pod \"cinder-api-0\" (UID: \"9a40110d-5675-40b6-87c6-63b1d2faba9d\") " pod="openstack/cinder-api-0" Feb 27 08:10:14 crc kubenswrapper[4612]: I0227 08:10:14.110479 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-cc6d746d7-b8wbb"] Feb 27 08:10:14 crc kubenswrapper[4612]: I0227 08:10:14.111256 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9a40110d-5675-40b6-87c6-63b1d2faba9d-config-data\") pod \"cinder-api-0\" (UID: \"9a40110d-5675-40b6-87c6-63b1d2faba9d\") " pod="openstack/cinder-api-0" Feb 27 08:10:14 crc kubenswrapper[4612]: I0227 08:10:14.111278 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/9a40110d-5675-40b6-87c6-63b1d2faba9d-etc-machine-id\") pod \"cinder-api-0\" (UID: \"9a40110d-5675-40b6-87c6-63b1d2faba9d\") " pod="openstack/cinder-api-0" Feb 27 08:10:14 crc kubenswrapper[4612]: I0227 08:10:14.111838 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9a40110d-5675-40b6-87c6-63b1d2faba9d-logs\") pod \"cinder-api-0\" (UID: \"9a40110d-5675-40b6-87c6-63b1d2faba9d\") " pod="openstack/cinder-api-0" Feb 27 08:10:14 crc kubenswrapper[4612]: I0227 08:10:14.115268 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9a40110d-5675-40b6-87c6-63b1d2faba9d-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"9a40110d-5675-40b6-87c6-63b1d2faba9d\") " pod="openstack/cinder-api-0" Feb 27 08:10:14 crc kubenswrapper[4612]: I0227 08:10:14.126757 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9a40110d-5675-40b6-87c6-63b1d2faba9d-scripts\") pod \"cinder-api-0\" (UID: \"9a40110d-5675-40b6-87c6-63b1d2faba9d\") " pod="openstack/cinder-api-0" Feb 27 08:10:14 crc kubenswrapper[4612]: I0227 08:10:14.127355 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nkgbh\" (UniqueName: \"kubernetes.io/projected/9a40110d-5675-40b6-87c6-63b1d2faba9d-kube-api-access-nkgbh\") pod \"cinder-api-0\" (UID: \"9a40110d-5675-40b6-87c6-63b1d2faba9d\") " pod="openstack/cinder-api-0" Feb 27 08:10:14 crc kubenswrapper[4612]: I0227 08:10:14.263898 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-55cd89c446-4c5hx"] Feb 27 08:10:14 crc kubenswrapper[4612]: I0227 08:10:14.264387 4612 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-55cd89c446-4c5hx" podUID="bb25e0d6-95cd-42a8-a182-f714a14d0084" containerName="barbican-api-log" containerID="cri-o://093ddc25c2613ae5cc7e84e5832820695b3bdc3586636199cfdfdf0a520691c0" gracePeriod=30 Feb 27 08:10:14 crc kubenswrapper[4612]: I0227 08:10:14.264488 4612 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-55cd89c446-4c5hx" podUID="bb25e0d6-95cd-42a8-a182-f714a14d0084" containerName="barbican-api" containerID="cri-o://fe5438c31c8c7aa2197f6896cae35a810753162b8b67a1c965e8e2d4cc8a9dfa" gracePeriod=30 Feb 27 08:10:14 crc kubenswrapper[4612]: I0227 08:10:14.336885 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b7b667979-pm5fl" Feb 27 08:10:14 crc kubenswrapper[4612]: I0227 08:10:14.352993 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Feb 27 08:10:14 crc kubenswrapper[4612]: I0227 08:10:14.367796 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-5ddbbb8689-bm5cp"] Feb 27 08:10:14 crc kubenswrapper[4612]: I0227 08:10:14.372207 4612 generic.go:334] "Generic (PLEG): container finished" podID="9ab32d5a-a797-47a8-961a-a534f55fdfa7" containerID="e77b56aa24d4104786d48b89f361df289c7f02205d97f075d23ee055ec70f239" exitCode=2 Feb 27 08:10:14 crc kubenswrapper[4612]: I0227 08:10:14.372264 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9ab32d5a-a797-47a8-961a-a534f55fdfa7","Type":"ContainerDied","Data":"e77b56aa24d4104786d48b89f361df289c7f02205d97f075d23ee055ec70f239"} Feb 27 08:10:14 crc kubenswrapper[4612]: I0227 08:10:14.385558 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-cc6d746d7-b8wbb" event={"ID":"41eb38cf-80fa-408e-b16c-686ef58a143f","Type":"ContainerStarted","Data":"6a669817801e1fffa65ffeb133be855b87b6347565fa99dbe3c9ca5fe1c089f1"} Feb 27 08:10:14 crc kubenswrapper[4612]: I0227 08:10:14.406297 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-5ddbbb8689-bm5cp"] Feb 27 08:10:14 crc kubenswrapper[4612]: I0227 08:10:14.413392 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29536330-nck2q" podStartSLOduration=2.441281991 podStartE2EDuration="14.413374119s" podCreationTimestamp="2026-02-27 08:10:00 +0000 UTC" firstStartedPulling="2026-02-27 08:10:01.118440476 +0000 UTC m=+1258.972370474" lastFinishedPulling="2026-02-27 08:10:13.090532604 +0000 UTC m=+1270.944462602" observedRunningTime="2026-02-27 08:10:14.372545649 +0000 UTC m=+1272.226475677" watchObservedRunningTime="2026-02-27 08:10:14.413374119 +0000 UTC m=+1272.267304107" Feb 27 08:10:14 crc kubenswrapper[4612]: I0227 08:10:14.451292 4612 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/barbican-api-55cd89c446-4c5hx" podUID="bb25e0d6-95cd-42a8-a182-f714a14d0084" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.165:9311/healthcheck\": EOF" Feb 27 08:10:14 crc kubenswrapper[4612]: I0227 08:10:14.451587 4612 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/barbican-api-55cd89c446-4c5hx" podUID="bb25e0d6-95cd-42a8-a182-f714a14d0084" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.165:9311/healthcheck\": EOF" Feb 27 08:10:14 crc kubenswrapper[4612]: I0227 08:10:14.463734 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6b7b667979-pm5fl"] Feb 27 08:10:14 crc kubenswrapper[4612]: I0227 08:10:14.486855 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6b7b667979-pm5fl"] Feb 27 08:10:14 crc kubenswrapper[4612]: E0227 08:10:14.610294 4612 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod649e0df0_b37c_42f5_8af0_c804d4f467d4.slice/crio-0974569712b15ea4dcc9022105f7deb4a6621d88d67b162ab1e1ec41c914fe97\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod649e0df0_b37c_42f5_8af0_c804d4f467d4.slice\": RecentStats: unable to find data in memory cache]" Feb 27 08:10:14 crc kubenswrapper[4612]: I0227 08:10:14.879494 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="649e0df0-b37c-42f5-8af0-c804d4f467d4" path="/var/lib/kubelet/pods/649e0df0-b37c-42f5-8af0-c804d4f467d4/volumes" Feb 27 08:10:14 crc kubenswrapper[4612]: I0227 08:10:14.880807 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf325571-0b74-447f-bd04-213d54a8437e" path="/var/lib/kubelet/pods/bf325571-0b74-447f-bd04-213d54a8437e/volumes" Feb 27 08:10:14 crc kubenswrapper[4612]: I0227 08:10:14.939805 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Feb 27 08:10:15 crc kubenswrapper[4612]: I0227 08:10:15.223744 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6578955fd5-ds7jf"] Feb 27 08:10:15 crc kubenswrapper[4612]: I0227 08:10:15.283401 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Feb 27 08:10:15 crc kubenswrapper[4612]: W0227 08:10:15.303409 4612 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9a40110d_5675_40b6_87c6_63b1d2faba9d.slice/crio-dbbfaac3a7d15a84f4fdaecc71f20fb511c1ae04bfb068e83c0f5eab387bc2f6 WatchSource:0}: Error finding container dbbfaac3a7d15a84f4fdaecc71f20fb511c1ae04bfb068e83c0f5eab387bc2f6: Status 404 returned error can't find the container with id dbbfaac3a7d15a84f4fdaecc71f20fb511c1ae04bfb068e83c0f5eab387bc2f6 Feb 27 08:10:15 crc kubenswrapper[4612]: I0227 08:10:15.411929 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"9a40110d-5675-40b6-87c6-63b1d2faba9d","Type":"ContainerStarted","Data":"dbbfaac3a7d15a84f4fdaecc71f20fb511c1ae04bfb068e83c0f5eab387bc2f6"} Feb 27 08:10:15 crc kubenswrapper[4612]: I0227 08:10:15.419029 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536330-nck2q" event={"ID":"073232e9-6f4e-409e-bf29-989b354fcc39","Type":"ContainerStarted","Data":"dd37a24177c3f58809092c5a7984ecc67bec7440f6050f16ab9c5a4322a5fe8a"} Feb 27 08:10:15 crc kubenswrapper[4612]: I0227 08:10:15.432482 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"62b23ba9-5e68-4665-86cd-df78c13de661","Type":"ContainerStarted","Data":"e2c7ae14f9b40037b3a83bbdb57a7b74920b1025583ccd173903d55118596ce9"} Feb 27 08:10:15 crc kubenswrapper[4612]: I0227 08:10:15.449302 4612 generic.go:334] "Generic (PLEG): container finished" podID="bb25e0d6-95cd-42a8-a182-f714a14d0084" containerID="093ddc25c2613ae5cc7e84e5832820695b3bdc3586636199cfdfdf0a520691c0" exitCode=143 Feb 27 08:10:15 crc kubenswrapper[4612]: I0227 08:10:15.449361 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-55cd89c446-4c5hx" event={"ID":"bb25e0d6-95cd-42a8-a182-f714a14d0084","Type":"ContainerDied","Data":"093ddc25c2613ae5cc7e84e5832820695b3bdc3586636199cfdfdf0a520691c0"} Feb 27 08:10:15 crc kubenswrapper[4612]: I0227 08:10:15.455073 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6578955fd5-ds7jf" event={"ID":"aedb2ee7-269c-49e6-9bf7-8ac34d6c1d48","Type":"ContainerStarted","Data":"ff55bc6b8616236eae12d1bc0aae268a90ceceb39912082d2ae50274b6bd1b60"} Feb 27 08:10:15 crc kubenswrapper[4612]: I0227 08:10:15.458586 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-cc6d746d7-b8wbb" event={"ID":"41eb38cf-80fa-408e-b16c-686ef58a143f","Type":"ContainerStarted","Data":"ac50d75330f839b389fa8f2af081c421468825ac3cb02fb973c11150b790512a"} Feb 27 08:10:16 crc kubenswrapper[4612]: I0227 08:10:16.484049 4612 generic.go:334] "Generic (PLEG): container finished" podID="aedb2ee7-269c-49e6-9bf7-8ac34d6c1d48" containerID="26a9589d9d888d24be276369ce5e1ce8a8c8a9d4d5b6ff3df566a8f91e230e69" exitCode=0 Feb 27 08:10:16 crc kubenswrapper[4612]: I0227 08:10:16.484412 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6578955fd5-ds7jf" event={"ID":"aedb2ee7-269c-49e6-9bf7-8ac34d6c1d48","Type":"ContainerDied","Data":"26a9589d9d888d24be276369ce5e1ce8a8c8a9d4d5b6ff3df566a8f91e230e69"} Feb 27 08:10:16 crc kubenswrapper[4612]: I0227 08:10:16.490211 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-cc6d746d7-b8wbb" event={"ID":"41eb38cf-80fa-408e-b16c-686ef58a143f","Type":"ContainerStarted","Data":"ae401dbdef654d5b18de2da2e5e44063a4e19bcd4df5c5c6bf3ac7455da7e71a"} Feb 27 08:10:16 crc kubenswrapper[4612]: I0227 08:10:16.490520 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-cc6d746d7-b8wbb" Feb 27 08:10:16 crc kubenswrapper[4612]: I0227 08:10:16.537415 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-cc6d746d7-b8wbb" podStartSLOduration=11.53739863 podStartE2EDuration="11.53739863s" podCreationTimestamp="2026-02-27 08:10:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 08:10:16.53181854 +0000 UTC m=+1274.385748538" watchObservedRunningTime="2026-02-27 08:10:16.53739863 +0000 UTC m=+1274.391328628" Feb 27 08:10:16 crc kubenswrapper[4612]: I0227 08:10:16.753983 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Feb 27 08:10:17 crc kubenswrapper[4612]: I0227 08:10:17.509290 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"9a40110d-5675-40b6-87c6-63b1d2faba9d","Type":"ContainerStarted","Data":"c198f8b781fb884ae7ba10334f69026ead837cda8cfc728772b0e7b942cb069e"} Feb 27 08:10:17 crc kubenswrapper[4612]: I0227 08:10:17.532320 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"62b23ba9-5e68-4665-86cd-df78c13de661","Type":"ContainerStarted","Data":"2b830865135697d89e47f2d5e4890e88bf23bb1fe36455912bcdc5ea5f0f65c1"} Feb 27 08:10:17 crc kubenswrapper[4612]: I0227 08:10:17.548305 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6578955fd5-ds7jf" event={"ID":"aedb2ee7-269c-49e6-9bf7-8ac34d6c1d48","Type":"ContainerStarted","Data":"c4bbdaba9741e72544fb14edcdad80fe5dc614a3942c1d0557e3c416add35dd7"} Feb 27 08:10:17 crc kubenswrapper[4612]: I0227 08:10:17.548384 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6578955fd5-ds7jf" Feb 27 08:10:17 crc kubenswrapper[4612]: I0227 08:10:17.586615 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6578955fd5-ds7jf" podStartSLOduration=4.586595454 podStartE2EDuration="4.586595454s" podCreationTimestamp="2026-02-27 08:10:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 08:10:17.582477175 +0000 UTC m=+1275.436407163" watchObservedRunningTime="2026-02-27 08:10:17.586595454 +0000 UTC m=+1275.440525462" Feb 27 08:10:18 crc kubenswrapper[4612]: I0227 08:10:18.387356 4612 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-55cd89c446-4c5hx" podUID="bb25e0d6-95cd-42a8-a182-f714a14d0084" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.165:9311/healthcheck\": read tcp 10.217.0.2:50532->10.217.0.165:9311: read: connection reset by peer" Feb 27 08:10:18 crc kubenswrapper[4612]: I0227 08:10:18.387374 4612 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-55cd89c446-4c5hx" podUID="bb25e0d6-95cd-42a8-a182-f714a14d0084" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.165:9311/healthcheck\": read tcp 10.217.0.2:50546->10.217.0.165:9311: read: connection reset by peer" Feb 27 08:10:18 crc kubenswrapper[4612]: I0227 08:10:18.572343 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"62b23ba9-5e68-4665-86cd-df78c13de661","Type":"ContainerStarted","Data":"370921685f3a020371fa0b5634f3b819b4505e8822283b03a1a586a24f9229e8"} Feb 27 08:10:18 crc kubenswrapper[4612]: I0227 08:10:18.580362 4612 generic.go:334] "Generic (PLEG): container finished" podID="bb25e0d6-95cd-42a8-a182-f714a14d0084" containerID="fe5438c31c8c7aa2197f6896cae35a810753162b8b67a1c965e8e2d4cc8a9dfa" exitCode=0 Feb 27 08:10:18 crc kubenswrapper[4612]: I0227 08:10:18.580432 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-55cd89c446-4c5hx" event={"ID":"bb25e0d6-95cd-42a8-a182-f714a14d0084","Type":"ContainerDied","Data":"fe5438c31c8c7aa2197f6896cae35a810753162b8b67a1c965e8e2d4cc8a9dfa"} Feb 27 08:10:18 crc kubenswrapper[4612]: I0227 08:10:18.592904 4612 generic.go:334] "Generic (PLEG): container finished" podID="9ab32d5a-a797-47a8-961a-a534f55fdfa7" containerID="0b87c11e9c758edb6f988ef0ee90e9d6d19fa4c1d46725a5fdb737090d12c6be" exitCode=0 Feb 27 08:10:18 crc kubenswrapper[4612]: I0227 08:10:18.592969 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9ab32d5a-a797-47a8-961a-a534f55fdfa7","Type":"ContainerDied","Data":"0b87c11e9c758edb6f988ef0ee90e9d6d19fa4c1d46725a5fdb737090d12c6be"} Feb 27 08:10:18 crc kubenswrapper[4612]: I0227 08:10:18.606197 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=4.680227158 podStartE2EDuration="5.606180729s" podCreationTimestamp="2026-02-27 08:10:13 +0000 UTC" firstStartedPulling="2026-02-27 08:10:14.94562683 +0000 UTC m=+1272.799556828" lastFinishedPulling="2026-02-27 08:10:15.871580401 +0000 UTC m=+1273.725510399" observedRunningTime="2026-02-27 08:10:18.59680724 +0000 UTC m=+1276.450737238" watchObservedRunningTime="2026-02-27 08:10:18.606180729 +0000 UTC m=+1276.460110727" Feb 27 08:10:18 crc kubenswrapper[4612]: I0227 08:10:18.617072 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"9a40110d-5675-40b6-87c6-63b1d2faba9d","Type":"ContainerStarted","Data":"6a08fb66db51ba5788882f4d7cb293f89bccd46dba71bf71878951639068185f"} Feb 27 08:10:18 crc kubenswrapper[4612]: I0227 08:10:18.617234 4612 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="9a40110d-5675-40b6-87c6-63b1d2faba9d" containerName="cinder-api-log" containerID="cri-o://c198f8b781fb884ae7ba10334f69026ead837cda8cfc728772b0e7b942cb069e" gracePeriod=30 Feb 27 08:10:18 crc kubenswrapper[4612]: I0227 08:10:18.617509 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Feb 27 08:10:18 crc kubenswrapper[4612]: I0227 08:10:18.617761 4612 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="9a40110d-5675-40b6-87c6-63b1d2faba9d" containerName="cinder-api" containerID="cri-o://6a08fb66db51ba5788882f4d7cb293f89bccd46dba71bf71878951639068185f" gracePeriod=30 Feb 27 08:10:18 crc kubenswrapper[4612]: I0227 08:10:18.626100 4612 generic.go:334] "Generic (PLEG): container finished" podID="073232e9-6f4e-409e-bf29-989b354fcc39" containerID="dd37a24177c3f58809092c5a7984ecc67bec7440f6050f16ab9c5a4322a5fe8a" exitCode=0 Feb 27 08:10:18 crc kubenswrapper[4612]: I0227 08:10:18.627052 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536330-nck2q" event={"ID":"073232e9-6f4e-409e-bf29-989b354fcc39","Type":"ContainerDied","Data":"dd37a24177c3f58809092c5a7984ecc67bec7440f6050f16ab9c5a4322a5fe8a"} Feb 27 08:10:18 crc kubenswrapper[4612]: I0227 08:10:18.653414 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=5.65337388 podStartE2EDuration="5.65337388s" podCreationTimestamp="2026-02-27 08:10:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 08:10:18.640162752 +0000 UTC m=+1276.494092750" watchObservedRunningTime="2026-02-27 08:10:18.65337388 +0000 UTC m=+1276.507303878" Feb 27 08:10:18 crc kubenswrapper[4612]: I0227 08:10:18.915110 4612 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Feb 27 08:10:19 crc kubenswrapper[4612]: I0227 08:10:19.095026 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 27 08:10:19 crc kubenswrapper[4612]: I0227 08:10:19.107521 4612 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-6b985547d4-mkfpv" podUID="76009243-16b3-4652-abfe-a6cdba363724" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.152:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.152:8443: connect: connection refused" Feb 27 08:10:19 crc kubenswrapper[4612]: I0227 08:10:19.107594 4612 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-6b985547d4-mkfpv" Feb 27 08:10:19 crc kubenswrapper[4612]: I0227 08:10:19.108347 4612 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="horizon" containerStatusID={"Type":"cri-o","ID":"a0baf671a66706916e8b02ab97ccdead951de20e3f62e64d913a8590aa359cb3"} pod="openstack/horizon-6b985547d4-mkfpv" containerMessage="Container horizon failed startup probe, will be restarted" Feb 27 08:10:19 crc kubenswrapper[4612]: I0227 08:10:19.108378 4612 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-6b985547d4-mkfpv" podUID="76009243-16b3-4652-abfe-a6cdba363724" containerName="horizon" containerID="cri-o://a0baf671a66706916e8b02ab97ccdead951de20e3f62e64d913a8590aa359cb3" gracePeriod=30 Feb 27 08:10:19 crc kubenswrapper[4612]: I0227 08:10:19.132778 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-55cd89c446-4c5hx" Feb 27 08:10:19 crc kubenswrapper[4612]: I0227 08:10:19.157391 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bb25e0d6-95cd-42a8-a182-f714a14d0084-config-data-custom\") pod \"bb25e0d6-95cd-42a8-a182-f714a14d0084\" (UID: \"bb25e0d6-95cd-42a8-a182-f714a14d0084\") " Feb 27 08:10:19 crc kubenswrapper[4612]: I0227 08:10:19.157439 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bb25e0d6-95cd-42a8-a182-f714a14d0084-config-data\") pod \"bb25e0d6-95cd-42a8-a182-f714a14d0084\" (UID: \"bb25e0d6-95cd-42a8-a182-f714a14d0084\") " Feb 27 08:10:19 crc kubenswrapper[4612]: I0227 08:10:19.157478 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bb25e0d6-95cd-42a8-a182-f714a14d0084-logs\") pod \"bb25e0d6-95cd-42a8-a182-f714a14d0084\" (UID: \"bb25e0d6-95cd-42a8-a182-f714a14d0084\") " Feb 27 08:10:19 crc kubenswrapper[4612]: I0227 08:10:19.157519 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9ab32d5a-a797-47a8-961a-a534f55fdfa7-log-httpd\") pod \"9ab32d5a-a797-47a8-961a-a534f55fdfa7\" (UID: \"9ab32d5a-a797-47a8-961a-a534f55fdfa7\") " Feb 27 08:10:19 crc kubenswrapper[4612]: I0227 08:10:19.157539 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mkkdq\" (UniqueName: \"kubernetes.io/projected/bb25e0d6-95cd-42a8-a182-f714a14d0084-kube-api-access-mkkdq\") pod \"bb25e0d6-95cd-42a8-a182-f714a14d0084\" (UID: \"bb25e0d6-95cd-42a8-a182-f714a14d0084\") " Feb 27 08:10:19 crc kubenswrapper[4612]: I0227 08:10:19.157657 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9ab32d5a-a797-47a8-961a-a534f55fdfa7-sg-core-conf-yaml\") pod \"9ab32d5a-a797-47a8-961a-a534f55fdfa7\" (UID: \"9ab32d5a-a797-47a8-961a-a534f55fdfa7\") " Feb 27 08:10:19 crc kubenswrapper[4612]: I0227 08:10:19.157677 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9ab32d5a-a797-47a8-961a-a534f55fdfa7-config-data\") pod \"9ab32d5a-a797-47a8-961a-a534f55fdfa7\" (UID: \"9ab32d5a-a797-47a8-961a-a534f55fdfa7\") " Feb 27 08:10:19 crc kubenswrapper[4612]: I0227 08:10:19.157705 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb25e0d6-95cd-42a8-a182-f714a14d0084-combined-ca-bundle\") pod \"bb25e0d6-95cd-42a8-a182-f714a14d0084\" (UID: \"bb25e0d6-95cd-42a8-a182-f714a14d0084\") " Feb 27 08:10:19 crc kubenswrapper[4612]: I0227 08:10:19.157728 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sh9th\" (UniqueName: \"kubernetes.io/projected/9ab32d5a-a797-47a8-961a-a534f55fdfa7-kube-api-access-sh9th\") pod \"9ab32d5a-a797-47a8-961a-a534f55fdfa7\" (UID: \"9ab32d5a-a797-47a8-961a-a534f55fdfa7\") " Feb 27 08:10:19 crc kubenswrapper[4612]: I0227 08:10:19.157745 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9ab32d5a-a797-47a8-961a-a534f55fdfa7-combined-ca-bundle\") pod \"9ab32d5a-a797-47a8-961a-a534f55fdfa7\" (UID: \"9ab32d5a-a797-47a8-961a-a534f55fdfa7\") " Feb 27 08:10:19 crc kubenswrapper[4612]: I0227 08:10:19.157775 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9ab32d5a-a797-47a8-961a-a534f55fdfa7-run-httpd\") pod \"9ab32d5a-a797-47a8-961a-a534f55fdfa7\" (UID: \"9ab32d5a-a797-47a8-961a-a534f55fdfa7\") " Feb 27 08:10:19 crc kubenswrapper[4612]: I0227 08:10:19.157846 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9ab32d5a-a797-47a8-961a-a534f55fdfa7-scripts\") pod \"9ab32d5a-a797-47a8-961a-a534f55fdfa7\" (UID: \"9ab32d5a-a797-47a8-961a-a534f55fdfa7\") " Feb 27 08:10:19 crc kubenswrapper[4612]: I0227 08:10:19.162319 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bb25e0d6-95cd-42a8-a182-f714a14d0084-logs" (OuterVolumeSpecName: "logs") pod "bb25e0d6-95cd-42a8-a182-f714a14d0084" (UID: "bb25e0d6-95cd-42a8-a182-f714a14d0084"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 08:10:19 crc kubenswrapper[4612]: I0227 08:10:19.177058 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9ab32d5a-a797-47a8-961a-a534f55fdfa7-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "9ab32d5a-a797-47a8-961a-a534f55fdfa7" (UID: "9ab32d5a-a797-47a8-961a-a534f55fdfa7"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 08:10:19 crc kubenswrapper[4612]: I0227 08:10:19.177909 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9ab32d5a-a797-47a8-961a-a534f55fdfa7-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "9ab32d5a-a797-47a8-961a-a534f55fdfa7" (UID: "9ab32d5a-a797-47a8-961a-a534f55fdfa7"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 08:10:19 crc kubenswrapper[4612]: I0227 08:10:19.183936 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9ab32d5a-a797-47a8-961a-a534f55fdfa7-kube-api-access-sh9th" (OuterVolumeSpecName: "kube-api-access-sh9th") pod "9ab32d5a-a797-47a8-961a-a534f55fdfa7" (UID: "9ab32d5a-a797-47a8-961a-a534f55fdfa7"). InnerVolumeSpecName "kube-api-access-sh9th". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 08:10:19 crc kubenswrapper[4612]: I0227 08:10:19.197028 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9ab32d5a-a797-47a8-961a-a534f55fdfa7-scripts" (OuterVolumeSpecName: "scripts") pod "9ab32d5a-a797-47a8-961a-a534f55fdfa7" (UID: "9ab32d5a-a797-47a8-961a-a534f55fdfa7"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:10:19 crc kubenswrapper[4612]: I0227 08:10:19.197118 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bb25e0d6-95cd-42a8-a182-f714a14d0084-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "bb25e0d6-95cd-42a8-a182-f714a14d0084" (UID: "bb25e0d6-95cd-42a8-a182-f714a14d0084"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:10:19 crc kubenswrapper[4612]: I0227 08:10:19.197200 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bb25e0d6-95cd-42a8-a182-f714a14d0084-kube-api-access-mkkdq" (OuterVolumeSpecName: "kube-api-access-mkkdq") pod "bb25e0d6-95cd-42a8-a182-f714a14d0084" (UID: "bb25e0d6-95cd-42a8-a182-f714a14d0084"). InnerVolumeSpecName "kube-api-access-mkkdq". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 08:10:19 crc kubenswrapper[4612]: I0227 08:10:19.252737 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9ab32d5a-a797-47a8-961a-a534f55fdfa7-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "9ab32d5a-a797-47a8-961a-a534f55fdfa7" (UID: "9ab32d5a-a797-47a8-961a-a534f55fdfa7"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:10:19 crc kubenswrapper[4612]: I0227 08:10:19.257070 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bb25e0d6-95cd-42a8-a182-f714a14d0084-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bb25e0d6-95cd-42a8-a182-f714a14d0084" (UID: "bb25e0d6-95cd-42a8-a182-f714a14d0084"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:10:19 crc kubenswrapper[4612]: I0227 08:10:19.265819 4612 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bb25e0d6-95cd-42a8-a182-f714a14d0084-config-data-custom\") on node \"crc\" DevicePath \"\"" Feb 27 08:10:19 crc kubenswrapper[4612]: I0227 08:10:19.265852 4612 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bb25e0d6-95cd-42a8-a182-f714a14d0084-logs\") on node \"crc\" DevicePath \"\"" Feb 27 08:10:19 crc kubenswrapper[4612]: I0227 08:10:19.265863 4612 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9ab32d5a-a797-47a8-961a-a534f55fdfa7-log-httpd\") on node \"crc\" DevicePath \"\"" Feb 27 08:10:19 crc kubenswrapper[4612]: I0227 08:10:19.265872 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mkkdq\" (UniqueName: \"kubernetes.io/projected/bb25e0d6-95cd-42a8-a182-f714a14d0084-kube-api-access-mkkdq\") on node \"crc\" DevicePath \"\"" Feb 27 08:10:19 crc kubenswrapper[4612]: I0227 08:10:19.265883 4612 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9ab32d5a-a797-47a8-961a-a534f55fdfa7-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Feb 27 08:10:19 crc kubenswrapper[4612]: I0227 08:10:19.265893 4612 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb25e0d6-95cd-42a8-a182-f714a14d0084-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 08:10:19 crc kubenswrapper[4612]: I0227 08:10:19.265902 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sh9th\" (UniqueName: \"kubernetes.io/projected/9ab32d5a-a797-47a8-961a-a534f55fdfa7-kube-api-access-sh9th\") on node \"crc\" DevicePath \"\"" Feb 27 08:10:19 crc kubenswrapper[4612]: I0227 08:10:19.265909 4612 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9ab32d5a-a797-47a8-961a-a534f55fdfa7-run-httpd\") on node \"crc\" DevicePath \"\"" Feb 27 08:10:19 crc kubenswrapper[4612]: I0227 08:10:19.265917 4612 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9ab32d5a-a797-47a8-961a-a534f55fdfa7-scripts\") on node \"crc\" DevicePath \"\"" Feb 27 08:10:19 crc kubenswrapper[4612]: I0227 08:10:19.272151 4612 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-77948db5bb-97qh6" podUID="b72b541d-4661-44a7-a121-c2a8aef6db11" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.153:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.153:8443: connect: connection refused" Feb 27 08:10:19 crc kubenswrapper[4612]: I0227 08:10:19.272221 4612 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-77948db5bb-97qh6" Feb 27 08:10:19 crc kubenswrapper[4612]: I0227 08:10:19.272957 4612 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="horizon" containerStatusID={"Type":"cri-o","ID":"92102897b17cb0c0c81984cbb9f2014fd0591e9bae23a94bbb5532b66e484fd8"} pod="openstack/horizon-77948db5bb-97qh6" containerMessage="Container horizon failed startup probe, will be restarted" Feb 27 08:10:19 crc kubenswrapper[4612]: I0227 08:10:19.272995 4612 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-77948db5bb-97qh6" podUID="b72b541d-4661-44a7-a121-c2a8aef6db11" containerName="horizon" containerID="cri-o://92102897b17cb0c0c81984cbb9f2014fd0591e9bae23a94bbb5532b66e484fd8" gracePeriod=30 Feb 27 08:10:19 crc kubenswrapper[4612]: I0227 08:10:19.273570 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9ab32d5a-a797-47a8-961a-a534f55fdfa7-config-data" (OuterVolumeSpecName: "config-data") pod "9ab32d5a-a797-47a8-961a-a534f55fdfa7" (UID: "9ab32d5a-a797-47a8-961a-a534f55fdfa7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:10:19 crc kubenswrapper[4612]: I0227 08:10:19.275457 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9ab32d5a-a797-47a8-961a-a534f55fdfa7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9ab32d5a-a797-47a8-961a-a534f55fdfa7" (UID: "9ab32d5a-a797-47a8-961a-a534f55fdfa7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:10:19 crc kubenswrapper[4612]: I0227 08:10:19.291827 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bb25e0d6-95cd-42a8-a182-f714a14d0084-config-data" (OuterVolumeSpecName: "config-data") pod "bb25e0d6-95cd-42a8-a182-f714a14d0084" (UID: "bb25e0d6-95cd-42a8-a182-f714a14d0084"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:10:19 crc kubenswrapper[4612]: I0227 08:10:19.367824 4612 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bb25e0d6-95cd-42a8-a182-f714a14d0084-config-data\") on node \"crc\" DevicePath \"\"" Feb 27 08:10:19 crc kubenswrapper[4612]: I0227 08:10:19.367856 4612 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9ab32d5a-a797-47a8-961a-a534f55fdfa7-config-data\") on node \"crc\" DevicePath \"\"" Feb 27 08:10:19 crc kubenswrapper[4612]: I0227 08:10:19.367866 4612 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9ab32d5a-a797-47a8-961a-a534f55fdfa7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 08:10:19 crc kubenswrapper[4612]: I0227 08:10:19.636386 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-55cd89c446-4c5hx" event={"ID":"bb25e0d6-95cd-42a8-a182-f714a14d0084","Type":"ContainerDied","Data":"80abde5b490b842dde142302c01ef4b837956aa9c3244ea384215f0463274da5"} Feb 27 08:10:19 crc kubenswrapper[4612]: I0227 08:10:19.636445 4612 scope.go:117] "RemoveContainer" containerID="fe5438c31c8c7aa2197f6896cae35a810753162b8b67a1c965e8e2d4cc8a9dfa" Feb 27 08:10:19 crc kubenswrapper[4612]: I0227 08:10:19.636590 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-55cd89c446-4c5hx" Feb 27 08:10:19 crc kubenswrapper[4612]: I0227 08:10:19.646121 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 27 08:10:19 crc kubenswrapper[4612]: I0227 08:10:19.646125 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9ab32d5a-a797-47a8-961a-a534f55fdfa7","Type":"ContainerDied","Data":"0674026bc8c0b7f46f8592de3f04756fc96cc88411ef5b665e4e65ca68b325b1"} Feb 27 08:10:19 crc kubenswrapper[4612]: I0227 08:10:19.650630 4612 generic.go:334] "Generic (PLEG): container finished" podID="9a40110d-5675-40b6-87c6-63b1d2faba9d" containerID="c198f8b781fb884ae7ba10334f69026ead837cda8cfc728772b0e7b942cb069e" exitCode=143 Feb 27 08:10:19 crc kubenswrapper[4612]: I0227 08:10:19.650755 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"9a40110d-5675-40b6-87c6-63b1d2faba9d","Type":"ContainerDied","Data":"c198f8b781fb884ae7ba10334f69026ead837cda8cfc728772b0e7b942cb069e"} Feb 27 08:10:19 crc kubenswrapper[4612]: I0227 08:10:19.684292 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-55cd89c446-4c5hx"] Feb 27 08:10:19 crc kubenswrapper[4612]: I0227 08:10:19.694357 4612 scope.go:117] "RemoveContainer" containerID="093ddc25c2613ae5cc7e84e5832820695b3bdc3586636199cfdfdf0a520691c0" Feb 27 08:10:19 crc kubenswrapper[4612]: I0227 08:10:19.699867 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-55cd89c446-4c5hx"] Feb 27 08:10:19 crc kubenswrapper[4612]: I0227 08:10:19.760375 4612 scope.go:117] "RemoveContainer" containerID="e77b56aa24d4104786d48b89f361df289c7f02205d97f075d23ee055ec70f239" Feb 27 08:10:19 crc kubenswrapper[4612]: I0227 08:10:19.762184 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 27 08:10:19 crc kubenswrapper[4612]: I0227 08:10:19.780174 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Feb 27 08:10:19 crc kubenswrapper[4612]: I0227 08:10:19.806974 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Feb 27 08:10:19 crc kubenswrapper[4612]: E0227 08:10:19.807337 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9ab32d5a-a797-47a8-961a-a534f55fdfa7" containerName="ceilometer-notification-agent" Feb 27 08:10:19 crc kubenswrapper[4612]: I0227 08:10:19.807353 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="9ab32d5a-a797-47a8-961a-a534f55fdfa7" containerName="ceilometer-notification-agent" Feb 27 08:10:19 crc kubenswrapper[4612]: E0227 08:10:19.807363 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb25e0d6-95cd-42a8-a182-f714a14d0084" containerName="barbican-api-log" Feb 27 08:10:19 crc kubenswrapper[4612]: I0227 08:10:19.807370 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb25e0d6-95cd-42a8-a182-f714a14d0084" containerName="barbican-api-log" Feb 27 08:10:19 crc kubenswrapper[4612]: E0227 08:10:19.807382 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9ab32d5a-a797-47a8-961a-a534f55fdfa7" containerName="sg-core" Feb 27 08:10:19 crc kubenswrapper[4612]: I0227 08:10:19.807388 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="9ab32d5a-a797-47a8-961a-a534f55fdfa7" containerName="sg-core" Feb 27 08:10:19 crc kubenswrapper[4612]: E0227 08:10:19.807411 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb25e0d6-95cd-42a8-a182-f714a14d0084" containerName="barbican-api" Feb 27 08:10:19 crc kubenswrapper[4612]: I0227 08:10:19.807417 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb25e0d6-95cd-42a8-a182-f714a14d0084" containerName="barbican-api" Feb 27 08:10:19 crc kubenswrapper[4612]: I0227 08:10:19.807607 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="bb25e0d6-95cd-42a8-a182-f714a14d0084" containerName="barbican-api" Feb 27 08:10:19 crc kubenswrapper[4612]: I0227 08:10:19.807636 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="9ab32d5a-a797-47a8-961a-a534f55fdfa7" containerName="ceilometer-notification-agent" Feb 27 08:10:19 crc kubenswrapper[4612]: I0227 08:10:19.807659 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="9ab32d5a-a797-47a8-961a-a534f55fdfa7" containerName="sg-core" Feb 27 08:10:19 crc kubenswrapper[4612]: I0227 08:10:19.807675 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="bb25e0d6-95cd-42a8-a182-f714a14d0084" containerName="barbican-api-log" Feb 27 08:10:19 crc kubenswrapper[4612]: I0227 08:10:19.809274 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 27 08:10:19 crc kubenswrapper[4612]: I0227 08:10:19.815953 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Feb 27 08:10:19 crc kubenswrapper[4612]: I0227 08:10:19.816271 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Feb 27 08:10:19 crc kubenswrapper[4612]: I0227 08:10:19.821839 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 27 08:10:19 crc kubenswrapper[4612]: I0227 08:10:19.870259 4612 scope.go:117] "RemoveContainer" containerID="0b87c11e9c758edb6f988ef0ee90e9d6d19fa4c1d46725a5fdb737090d12c6be" Feb 27 08:10:19 crc kubenswrapper[4612]: I0227 08:10:19.880329 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/39aaf83c-4be0-4bd8-ae54-cb38271a5eb1-config-data\") pod \"ceilometer-0\" (UID: \"39aaf83c-4be0-4bd8-ae54-cb38271a5eb1\") " pod="openstack/ceilometer-0" Feb 27 08:10:19 crc kubenswrapper[4612]: I0227 08:10:19.880392 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q4b6g\" (UniqueName: \"kubernetes.io/projected/39aaf83c-4be0-4bd8-ae54-cb38271a5eb1-kube-api-access-q4b6g\") pod \"ceilometer-0\" (UID: \"39aaf83c-4be0-4bd8-ae54-cb38271a5eb1\") " pod="openstack/ceilometer-0" Feb 27 08:10:19 crc kubenswrapper[4612]: I0227 08:10:19.880419 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/39aaf83c-4be0-4bd8-ae54-cb38271a5eb1-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"39aaf83c-4be0-4bd8-ae54-cb38271a5eb1\") " pod="openstack/ceilometer-0" Feb 27 08:10:19 crc kubenswrapper[4612]: I0227 08:10:19.880469 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/39aaf83c-4be0-4bd8-ae54-cb38271a5eb1-scripts\") pod \"ceilometer-0\" (UID: \"39aaf83c-4be0-4bd8-ae54-cb38271a5eb1\") " pod="openstack/ceilometer-0" Feb 27 08:10:19 crc kubenswrapper[4612]: I0227 08:10:19.880512 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/39aaf83c-4be0-4bd8-ae54-cb38271a5eb1-run-httpd\") pod \"ceilometer-0\" (UID: \"39aaf83c-4be0-4bd8-ae54-cb38271a5eb1\") " pod="openstack/ceilometer-0" Feb 27 08:10:19 crc kubenswrapper[4612]: I0227 08:10:19.880560 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/39aaf83c-4be0-4bd8-ae54-cb38271a5eb1-log-httpd\") pod \"ceilometer-0\" (UID: \"39aaf83c-4be0-4bd8-ae54-cb38271a5eb1\") " pod="openstack/ceilometer-0" Feb 27 08:10:19 crc kubenswrapper[4612]: I0227 08:10:19.880625 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/39aaf83c-4be0-4bd8-ae54-cb38271a5eb1-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"39aaf83c-4be0-4bd8-ae54-cb38271a5eb1\") " pod="openstack/ceilometer-0" Feb 27 08:10:19 crc kubenswrapper[4612]: I0227 08:10:19.982652 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/39aaf83c-4be0-4bd8-ae54-cb38271a5eb1-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"39aaf83c-4be0-4bd8-ae54-cb38271a5eb1\") " pod="openstack/ceilometer-0" Feb 27 08:10:19 crc kubenswrapper[4612]: I0227 08:10:19.982744 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/39aaf83c-4be0-4bd8-ae54-cb38271a5eb1-config-data\") pod \"ceilometer-0\" (UID: \"39aaf83c-4be0-4bd8-ae54-cb38271a5eb1\") " pod="openstack/ceilometer-0" Feb 27 08:10:19 crc kubenswrapper[4612]: I0227 08:10:19.982784 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q4b6g\" (UniqueName: \"kubernetes.io/projected/39aaf83c-4be0-4bd8-ae54-cb38271a5eb1-kube-api-access-q4b6g\") pod \"ceilometer-0\" (UID: \"39aaf83c-4be0-4bd8-ae54-cb38271a5eb1\") " pod="openstack/ceilometer-0" Feb 27 08:10:19 crc kubenswrapper[4612]: I0227 08:10:19.982809 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/39aaf83c-4be0-4bd8-ae54-cb38271a5eb1-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"39aaf83c-4be0-4bd8-ae54-cb38271a5eb1\") " pod="openstack/ceilometer-0" Feb 27 08:10:19 crc kubenswrapper[4612]: I0227 08:10:19.982841 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/39aaf83c-4be0-4bd8-ae54-cb38271a5eb1-scripts\") pod \"ceilometer-0\" (UID: \"39aaf83c-4be0-4bd8-ae54-cb38271a5eb1\") " pod="openstack/ceilometer-0" Feb 27 08:10:19 crc kubenswrapper[4612]: I0227 08:10:19.982868 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/39aaf83c-4be0-4bd8-ae54-cb38271a5eb1-run-httpd\") pod \"ceilometer-0\" (UID: \"39aaf83c-4be0-4bd8-ae54-cb38271a5eb1\") " pod="openstack/ceilometer-0" Feb 27 08:10:19 crc kubenswrapper[4612]: I0227 08:10:19.982905 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/39aaf83c-4be0-4bd8-ae54-cb38271a5eb1-log-httpd\") pod \"ceilometer-0\" (UID: \"39aaf83c-4be0-4bd8-ae54-cb38271a5eb1\") " pod="openstack/ceilometer-0" Feb 27 08:10:19 crc kubenswrapper[4612]: I0227 08:10:19.983350 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/39aaf83c-4be0-4bd8-ae54-cb38271a5eb1-log-httpd\") pod \"ceilometer-0\" (UID: \"39aaf83c-4be0-4bd8-ae54-cb38271a5eb1\") " pod="openstack/ceilometer-0" Feb 27 08:10:19 crc kubenswrapper[4612]: I0227 08:10:19.983777 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/39aaf83c-4be0-4bd8-ae54-cb38271a5eb1-run-httpd\") pod \"ceilometer-0\" (UID: \"39aaf83c-4be0-4bd8-ae54-cb38271a5eb1\") " pod="openstack/ceilometer-0" Feb 27 08:10:19 crc kubenswrapper[4612]: I0227 08:10:19.997802 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/39aaf83c-4be0-4bd8-ae54-cb38271a5eb1-config-data\") pod \"ceilometer-0\" (UID: \"39aaf83c-4be0-4bd8-ae54-cb38271a5eb1\") " pod="openstack/ceilometer-0" Feb 27 08:10:19 crc kubenswrapper[4612]: I0227 08:10:19.998448 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/39aaf83c-4be0-4bd8-ae54-cb38271a5eb1-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"39aaf83c-4be0-4bd8-ae54-cb38271a5eb1\") " pod="openstack/ceilometer-0" Feb 27 08:10:20 crc kubenswrapper[4612]: I0227 08:10:20.007968 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/39aaf83c-4be0-4bd8-ae54-cb38271a5eb1-scripts\") pod \"ceilometer-0\" (UID: \"39aaf83c-4be0-4bd8-ae54-cb38271a5eb1\") " pod="openstack/ceilometer-0" Feb 27 08:10:20 crc kubenswrapper[4612]: I0227 08:10:20.017829 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/39aaf83c-4be0-4bd8-ae54-cb38271a5eb1-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"39aaf83c-4be0-4bd8-ae54-cb38271a5eb1\") " pod="openstack/ceilometer-0" Feb 27 08:10:20 crc kubenswrapper[4612]: I0227 08:10:20.071497 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q4b6g\" (UniqueName: \"kubernetes.io/projected/39aaf83c-4be0-4bd8-ae54-cb38271a5eb1-kube-api-access-q4b6g\") pod \"ceilometer-0\" (UID: \"39aaf83c-4be0-4bd8-ae54-cb38271a5eb1\") " pod="openstack/ceilometer-0" Feb 27 08:10:20 crc kubenswrapper[4612]: I0227 08:10:20.148249 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 27 08:10:20 crc kubenswrapper[4612]: I0227 08:10:20.274621 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536330-nck2q" Feb 27 08:10:20 crc kubenswrapper[4612]: I0227 08:10:20.404567 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fzgl5\" (UniqueName: \"kubernetes.io/projected/073232e9-6f4e-409e-bf29-989b354fcc39-kube-api-access-fzgl5\") pod \"073232e9-6f4e-409e-bf29-989b354fcc39\" (UID: \"073232e9-6f4e-409e-bf29-989b354fcc39\") " Feb 27 08:10:20 crc kubenswrapper[4612]: I0227 08:10:20.412921 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/073232e9-6f4e-409e-bf29-989b354fcc39-kube-api-access-fzgl5" (OuterVolumeSpecName: "kube-api-access-fzgl5") pod "073232e9-6f4e-409e-bf29-989b354fcc39" (UID: "073232e9-6f4e-409e-bf29-989b354fcc39"). InnerVolumeSpecName "kube-api-access-fzgl5". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 08:10:20 crc kubenswrapper[4612]: I0227 08:10:20.507220 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fzgl5\" (UniqueName: \"kubernetes.io/projected/073232e9-6f4e-409e-bf29-989b354fcc39-kube-api-access-fzgl5\") on node \"crc\" DevicePath \"\"" Feb 27 08:10:20 crc kubenswrapper[4612]: I0227 08:10:20.659965 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536330-nck2q" event={"ID":"073232e9-6f4e-409e-bf29-989b354fcc39","Type":"ContainerDied","Data":"1755fd2ce8358fd0df3e900e64f6e14e513b1b51402750e05b2e3cd2e246327c"} Feb 27 08:10:20 crc kubenswrapper[4612]: I0227 08:10:20.660010 4612 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1755fd2ce8358fd0df3e900e64f6e14e513b1b51402750e05b2e3cd2e246327c" Feb 27 08:10:20 crc kubenswrapper[4612]: I0227 08:10:20.660299 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536330-nck2q" Feb 27 08:10:20 crc kubenswrapper[4612]: I0227 08:10:20.753485 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29536324-2tpql"] Feb 27 08:10:20 crc kubenswrapper[4612]: I0227 08:10:20.762221 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29536324-2tpql"] Feb 27 08:10:20 crc kubenswrapper[4612]: I0227 08:10:20.799211 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 27 08:10:20 crc kubenswrapper[4612]: I0227 08:10:20.862793 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="79cb5a3d-2162-4f09-8508-c6102ab5e4cf" path="/var/lib/kubelet/pods/79cb5a3d-2162-4f09-8508-c6102ab5e4cf/volumes" Feb 27 08:10:20 crc kubenswrapper[4612]: I0227 08:10:20.863676 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9ab32d5a-a797-47a8-961a-a534f55fdfa7" path="/var/lib/kubelet/pods/9ab32d5a-a797-47a8-961a-a534f55fdfa7/volumes" Feb 27 08:10:20 crc kubenswrapper[4612]: I0227 08:10:20.864409 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bb25e0d6-95cd-42a8-a182-f714a14d0084" path="/var/lib/kubelet/pods/bb25e0d6-95cd-42a8-a182-f714a14d0084/volumes" Feb 27 08:10:21 crc kubenswrapper[4612]: I0227 08:10:21.277901 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Feb 27 08:10:21 crc kubenswrapper[4612]: I0227 08:10:21.297146 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Feb 27 08:10:21 crc kubenswrapper[4612]: I0227 08:10:21.676781 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"39aaf83c-4be0-4bd8-ae54-cb38271a5eb1","Type":"ContainerStarted","Data":"57e5093fb9f40dabcd4b4a0f88dd15403373494bdf7de2f87e395233ca1b7d99"} Feb 27 08:10:21 crc kubenswrapper[4612]: I0227 08:10:21.677205 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"39aaf83c-4be0-4bd8-ae54-cb38271a5eb1","Type":"ContainerStarted","Data":"c15e926ec4288e5cebe0309bb45c3d26e9cc12a650fdeb394602a26bce3029c1"} Feb 27 08:10:22 crc kubenswrapper[4612]: I0227 08:10:22.684566 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"39aaf83c-4be0-4bd8-ae54-cb38271a5eb1","Type":"ContainerStarted","Data":"27ac184f71751114a7ced8f7acf694e9812f35974b862b9042adbc00259c5c4b"} Feb 27 08:10:23 crc kubenswrapper[4612]: I0227 08:10:23.695608 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"39aaf83c-4be0-4bd8-ae54-cb38271a5eb1","Type":"ContainerStarted","Data":"4998f78f73290e3176d379a64fecaed3854ebacac2c7b7154e92c1cd0b3d0302"} Feb 27 08:10:24 crc kubenswrapper[4612]: I0227 08:10:24.042877 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6578955fd5-ds7jf" Feb 27 08:10:24 crc kubenswrapper[4612]: I0227 08:10:24.144324 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-848cf88cfc-tz267"] Feb 27 08:10:24 crc kubenswrapper[4612]: I0227 08:10:24.145459 4612 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-848cf88cfc-tz267" podUID="61899b16-fb85-4ba6-b057-65ad130ea015" containerName="dnsmasq-dns" containerID="cri-o://452de9407de7ea11ac17676dc5978b7bcc066fe94ddc3253fac1679600573f08" gracePeriod=10 Feb 27 08:10:24 crc kubenswrapper[4612]: I0227 08:10:24.723359 4612 generic.go:334] "Generic (PLEG): container finished" podID="61899b16-fb85-4ba6-b057-65ad130ea015" containerID="452de9407de7ea11ac17676dc5978b7bcc066fe94ddc3253fac1679600573f08" exitCode=0 Feb 27 08:10:24 crc kubenswrapper[4612]: I0227 08:10:24.723629 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-848cf88cfc-tz267" event={"ID":"61899b16-fb85-4ba6-b057-65ad130ea015","Type":"ContainerDied","Data":"452de9407de7ea11ac17676dc5978b7bcc066fe94ddc3253fac1679600573f08"} Feb 27 08:10:25 crc kubenswrapper[4612]: I0227 08:10:25.124122 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-848cf88cfc-tz267" Feb 27 08:10:25 crc kubenswrapper[4612]: I0227 08:10:25.228042 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/61899b16-fb85-4ba6-b057-65ad130ea015-ovsdbserver-nb\") pod \"61899b16-fb85-4ba6-b057-65ad130ea015\" (UID: \"61899b16-fb85-4ba6-b057-65ad130ea015\") " Feb 27 08:10:25 crc kubenswrapper[4612]: I0227 08:10:25.229709 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qp9jw\" (UniqueName: \"kubernetes.io/projected/61899b16-fb85-4ba6-b057-65ad130ea015-kube-api-access-qp9jw\") pod \"61899b16-fb85-4ba6-b057-65ad130ea015\" (UID: \"61899b16-fb85-4ba6-b057-65ad130ea015\") " Feb 27 08:10:25 crc kubenswrapper[4612]: I0227 08:10:25.231073 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/61899b16-fb85-4ba6-b057-65ad130ea015-config\") pod \"61899b16-fb85-4ba6-b057-65ad130ea015\" (UID: \"61899b16-fb85-4ba6-b057-65ad130ea015\") " Feb 27 08:10:25 crc kubenswrapper[4612]: I0227 08:10:25.231186 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/61899b16-fb85-4ba6-b057-65ad130ea015-ovsdbserver-sb\") pod \"61899b16-fb85-4ba6-b057-65ad130ea015\" (UID: \"61899b16-fb85-4ba6-b057-65ad130ea015\") " Feb 27 08:10:25 crc kubenswrapper[4612]: I0227 08:10:25.231276 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/61899b16-fb85-4ba6-b057-65ad130ea015-dns-swift-storage-0\") pod \"61899b16-fb85-4ba6-b057-65ad130ea015\" (UID: \"61899b16-fb85-4ba6-b057-65ad130ea015\") " Feb 27 08:10:25 crc kubenswrapper[4612]: I0227 08:10:25.231349 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/61899b16-fb85-4ba6-b057-65ad130ea015-dns-svc\") pod \"61899b16-fb85-4ba6-b057-65ad130ea015\" (UID: \"61899b16-fb85-4ba6-b057-65ad130ea015\") " Feb 27 08:10:25 crc kubenswrapper[4612]: I0227 08:10:25.262253 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/61899b16-fb85-4ba6-b057-65ad130ea015-kube-api-access-qp9jw" (OuterVolumeSpecName: "kube-api-access-qp9jw") pod "61899b16-fb85-4ba6-b057-65ad130ea015" (UID: "61899b16-fb85-4ba6-b057-65ad130ea015"). InnerVolumeSpecName "kube-api-access-qp9jw". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 08:10:25 crc kubenswrapper[4612]: I0227 08:10:25.333550 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qp9jw\" (UniqueName: \"kubernetes.io/projected/61899b16-fb85-4ba6-b057-65ad130ea015-kube-api-access-qp9jw\") on node \"crc\" DevicePath \"\"" Feb 27 08:10:25 crc kubenswrapper[4612]: I0227 08:10:25.347554 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/61899b16-fb85-4ba6-b057-65ad130ea015-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "61899b16-fb85-4ba6-b057-65ad130ea015" (UID: "61899b16-fb85-4ba6-b057-65ad130ea015"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 08:10:25 crc kubenswrapper[4612]: I0227 08:10:25.357121 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/61899b16-fb85-4ba6-b057-65ad130ea015-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "61899b16-fb85-4ba6-b057-65ad130ea015" (UID: "61899b16-fb85-4ba6-b057-65ad130ea015"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 08:10:25 crc kubenswrapper[4612]: I0227 08:10:25.366451 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/61899b16-fb85-4ba6-b057-65ad130ea015-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "61899b16-fb85-4ba6-b057-65ad130ea015" (UID: "61899b16-fb85-4ba6-b057-65ad130ea015"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 08:10:25 crc kubenswrapper[4612]: I0227 08:10:25.392579 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/61899b16-fb85-4ba6-b057-65ad130ea015-config" (OuterVolumeSpecName: "config") pod "61899b16-fb85-4ba6-b057-65ad130ea015" (UID: "61899b16-fb85-4ba6-b057-65ad130ea015"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 08:10:25 crc kubenswrapper[4612]: I0227 08:10:25.400487 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/61899b16-fb85-4ba6-b057-65ad130ea015-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "61899b16-fb85-4ba6-b057-65ad130ea015" (UID: "61899b16-fb85-4ba6-b057-65ad130ea015"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 08:10:25 crc kubenswrapper[4612]: I0227 08:10:25.404053 4612 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Feb 27 08:10:25 crc kubenswrapper[4612]: I0227 08:10:25.435520 4612 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/61899b16-fb85-4ba6-b057-65ad130ea015-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 27 08:10:25 crc kubenswrapper[4612]: I0227 08:10:25.435553 4612 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/61899b16-fb85-4ba6-b057-65ad130ea015-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Feb 27 08:10:25 crc kubenswrapper[4612]: I0227 08:10:25.435563 4612 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/61899b16-fb85-4ba6-b057-65ad130ea015-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 27 08:10:25 crc kubenswrapper[4612]: I0227 08:10:25.435571 4612 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/61899b16-fb85-4ba6-b057-65ad130ea015-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 27 08:10:25 crc kubenswrapper[4612]: I0227 08:10:25.435579 4612 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/61899b16-fb85-4ba6-b057-65ad130ea015-config\") on node \"crc\" DevicePath \"\"" Feb 27 08:10:25 crc kubenswrapper[4612]: I0227 08:10:25.486467 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Feb 27 08:10:25 crc kubenswrapper[4612]: I0227 08:10:25.733375 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"39aaf83c-4be0-4bd8-ae54-cb38271a5eb1","Type":"ContainerStarted","Data":"34e1ccd60759f81efa6f32f20783c47476ad2417b5e2dbeb9bd0428b84a2bf7d"} Feb 27 08:10:25 crc kubenswrapper[4612]: I0227 08:10:25.734713 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Feb 27 08:10:25 crc kubenswrapper[4612]: I0227 08:10:25.736886 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-848cf88cfc-tz267" event={"ID":"61899b16-fb85-4ba6-b057-65ad130ea015","Type":"ContainerDied","Data":"67dbb6b7b932909ae4d2afaae2e16ad844e84d4fa7023cf89cf20bf38cc72f6a"} Feb 27 08:10:25 crc kubenswrapper[4612]: I0227 08:10:25.736922 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-848cf88cfc-tz267" Feb 27 08:10:25 crc kubenswrapper[4612]: I0227 08:10:25.736953 4612 scope.go:117] "RemoveContainer" containerID="452de9407de7ea11ac17676dc5978b7bcc066fe94ddc3253fac1679600573f08" Feb 27 08:10:25 crc kubenswrapper[4612]: I0227 08:10:25.737649 4612 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="62b23ba9-5e68-4665-86cd-df78c13de661" containerName="probe" containerID="cri-o://370921685f3a020371fa0b5634f3b819b4505e8822283b03a1a586a24f9229e8" gracePeriod=30 Feb 27 08:10:25 crc kubenswrapper[4612]: I0227 08:10:25.737622 4612 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="62b23ba9-5e68-4665-86cd-df78c13de661" containerName="cinder-scheduler" containerID="cri-o://2b830865135697d89e47f2d5e4890e88bf23bb1fe36455912bcdc5ea5f0f65c1" gracePeriod=30 Feb 27 08:10:25 crc kubenswrapper[4612]: I0227 08:10:25.760426 4612 scope.go:117] "RemoveContainer" containerID="d825d98443adaf0ad96e63a1bf5e5533113dab4ebe6f67e171d94b86c28fd84f" Feb 27 08:10:25 crc kubenswrapper[4612]: I0227 08:10:25.767226 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.419072749 podStartE2EDuration="6.76720891s" podCreationTimestamp="2026-02-27 08:10:19 +0000 UTC" firstStartedPulling="2026-02-27 08:10:20.785021171 +0000 UTC m=+1278.638951169" lastFinishedPulling="2026-02-27 08:10:25.133157332 +0000 UTC m=+1282.987087330" observedRunningTime="2026-02-27 08:10:25.766368486 +0000 UTC m=+1283.620298484" watchObservedRunningTime="2026-02-27 08:10:25.76720891 +0000 UTC m=+1283.621138908" Feb 27 08:10:25 crc kubenswrapper[4612]: I0227 08:10:25.806602 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-848cf88cfc-tz267"] Feb 27 08:10:25 crc kubenswrapper[4612]: I0227 08:10:25.812204 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-848cf88cfc-tz267"] Feb 27 08:10:26 crc kubenswrapper[4612]: I0227 08:10:26.084445 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-9c6fdd6-lvn8p" Feb 27 08:10:26 crc kubenswrapper[4612]: I0227 08:10:26.431260 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-7d477cc4b6-5qtkx" Feb 27 08:10:26 crc kubenswrapper[4612]: I0227 08:10:26.435876 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-7d477cc4b6-5qtkx" Feb 27 08:10:26 crc kubenswrapper[4612]: I0227 08:10:26.745296 4612 generic.go:334] "Generic (PLEG): container finished" podID="62b23ba9-5e68-4665-86cd-df78c13de661" containerID="370921685f3a020371fa0b5634f3b819b4505e8822283b03a1a586a24f9229e8" exitCode=0 Feb 27 08:10:26 crc kubenswrapper[4612]: I0227 08:10:26.745600 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"62b23ba9-5e68-4665-86cd-df78c13de661","Type":"ContainerDied","Data":"370921685f3a020371fa0b5634f3b819b4505e8822283b03a1a586a24f9229e8"} Feb 27 08:10:26 crc kubenswrapper[4612]: I0227 08:10:26.862431 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="61899b16-fb85-4ba6-b057-65ad130ea015" path="/var/lib/kubelet/pods/61899b16-fb85-4ba6-b057-65ad130ea015/volumes" Feb 27 08:10:28 crc kubenswrapper[4612]: I0227 08:10:28.763412 4612 generic.go:334] "Generic (PLEG): container finished" podID="62b23ba9-5e68-4665-86cd-df78c13de661" containerID="2b830865135697d89e47f2d5e4890e88bf23bb1fe36455912bcdc5ea5f0f65c1" exitCode=0 Feb 27 08:10:28 crc kubenswrapper[4612]: I0227 08:10:28.763504 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"62b23ba9-5e68-4665-86cd-df78c13de661","Type":"ContainerDied","Data":"2b830865135697d89e47f2d5e4890e88bf23bb1fe36455912bcdc5ea5f0f65c1"} Feb 27 08:10:28 crc kubenswrapper[4612]: I0227 08:10:28.763710 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"62b23ba9-5e68-4665-86cd-df78c13de661","Type":"ContainerDied","Data":"e2c7ae14f9b40037b3a83bbdb57a7b74920b1025583ccd173903d55118596ce9"} Feb 27 08:10:28 crc kubenswrapper[4612]: I0227 08:10:28.763727 4612 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e2c7ae14f9b40037b3a83bbdb57a7b74920b1025583ccd173903d55118596ce9" Feb 27 08:10:28 crc kubenswrapper[4612]: I0227 08:10:28.772126 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Feb 27 08:10:28 crc kubenswrapper[4612]: I0227 08:10:28.915965 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/62b23ba9-5e68-4665-86cd-df78c13de661-scripts\") pod \"62b23ba9-5e68-4665-86cd-df78c13de661\" (UID: \"62b23ba9-5e68-4665-86cd-df78c13de661\") " Feb 27 08:10:28 crc kubenswrapper[4612]: I0227 08:10:28.916323 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/62b23ba9-5e68-4665-86cd-df78c13de661-config-data\") pod \"62b23ba9-5e68-4665-86cd-df78c13de661\" (UID: \"62b23ba9-5e68-4665-86cd-df78c13de661\") " Feb 27 08:10:28 crc kubenswrapper[4612]: I0227 08:10:28.916992 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-spf6v\" (UniqueName: \"kubernetes.io/projected/62b23ba9-5e68-4665-86cd-df78c13de661-kube-api-access-spf6v\") pod \"62b23ba9-5e68-4665-86cd-df78c13de661\" (UID: \"62b23ba9-5e68-4665-86cd-df78c13de661\") " Feb 27 08:10:28 crc kubenswrapper[4612]: I0227 08:10:28.917024 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62b23ba9-5e68-4665-86cd-df78c13de661-combined-ca-bundle\") pod \"62b23ba9-5e68-4665-86cd-df78c13de661\" (UID: \"62b23ba9-5e68-4665-86cd-df78c13de661\") " Feb 27 08:10:28 crc kubenswrapper[4612]: I0227 08:10:28.917052 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/62b23ba9-5e68-4665-86cd-df78c13de661-etc-machine-id\") pod \"62b23ba9-5e68-4665-86cd-df78c13de661\" (UID: \"62b23ba9-5e68-4665-86cd-df78c13de661\") " Feb 27 08:10:28 crc kubenswrapper[4612]: I0227 08:10:28.917090 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/62b23ba9-5e68-4665-86cd-df78c13de661-config-data-custom\") pod \"62b23ba9-5e68-4665-86cd-df78c13de661\" (UID: \"62b23ba9-5e68-4665-86cd-df78c13de661\") " Feb 27 08:10:28 crc kubenswrapper[4612]: I0227 08:10:28.918479 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/62b23ba9-5e68-4665-86cd-df78c13de661-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "62b23ba9-5e68-4665-86cd-df78c13de661" (UID: "62b23ba9-5e68-4665-86cd-df78c13de661"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 27 08:10:28 crc kubenswrapper[4612]: I0227 08:10:28.926223 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/62b23ba9-5e68-4665-86cd-df78c13de661-kube-api-access-spf6v" (OuterVolumeSpecName: "kube-api-access-spf6v") pod "62b23ba9-5e68-4665-86cd-df78c13de661" (UID: "62b23ba9-5e68-4665-86cd-df78c13de661"). InnerVolumeSpecName "kube-api-access-spf6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 08:10:28 crc kubenswrapper[4612]: I0227 08:10:28.937357 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/62b23ba9-5e68-4665-86cd-df78c13de661-scripts" (OuterVolumeSpecName: "scripts") pod "62b23ba9-5e68-4665-86cd-df78c13de661" (UID: "62b23ba9-5e68-4665-86cd-df78c13de661"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:10:28 crc kubenswrapper[4612]: I0227 08:10:28.937399 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/62b23ba9-5e68-4665-86cd-df78c13de661-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "62b23ba9-5e68-4665-86cd-df78c13de661" (UID: "62b23ba9-5e68-4665-86cd-df78c13de661"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:10:29 crc kubenswrapper[4612]: I0227 08:10:29.022771 4612 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/62b23ba9-5e68-4665-86cd-df78c13de661-scripts\") on node \"crc\" DevicePath \"\"" Feb 27 08:10:29 crc kubenswrapper[4612]: I0227 08:10:29.022816 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-spf6v\" (UniqueName: \"kubernetes.io/projected/62b23ba9-5e68-4665-86cd-df78c13de661-kube-api-access-spf6v\") on node \"crc\" DevicePath \"\"" Feb 27 08:10:29 crc kubenswrapper[4612]: I0227 08:10:29.022827 4612 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/62b23ba9-5e68-4665-86cd-df78c13de661-etc-machine-id\") on node \"crc\" DevicePath \"\"" Feb 27 08:10:29 crc kubenswrapper[4612]: I0227 08:10:29.022835 4612 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/62b23ba9-5e68-4665-86cd-df78c13de661-config-data-custom\") on node \"crc\" DevicePath \"\"" Feb 27 08:10:29 crc kubenswrapper[4612]: I0227 08:10:29.043428 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/62b23ba9-5e68-4665-86cd-df78c13de661-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "62b23ba9-5e68-4665-86cd-df78c13de661" (UID: "62b23ba9-5e68-4665-86cd-df78c13de661"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:10:29 crc kubenswrapper[4612]: I0227 08:10:29.106029 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Feb 27 08:10:29 crc kubenswrapper[4612]: I0227 08:10:29.124668 4612 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62b23ba9-5e68-4665-86cd-df78c13de661-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 08:10:29 crc kubenswrapper[4612]: I0227 08:10:29.153131 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/62b23ba9-5e68-4665-86cd-df78c13de661-config-data" (OuterVolumeSpecName: "config-data") pod "62b23ba9-5e68-4665-86cd-df78c13de661" (UID: "62b23ba9-5e68-4665-86cd-df78c13de661"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:10:29 crc kubenswrapper[4612]: I0227 08:10:29.226003 4612 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/62b23ba9-5e68-4665-86cd-df78c13de661-config-data\") on node \"crc\" DevicePath \"\"" Feb 27 08:10:29 crc kubenswrapper[4612]: I0227 08:10:29.498665 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Feb 27 08:10:29 crc kubenswrapper[4612]: E0227 08:10:29.499047 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="62b23ba9-5e68-4665-86cd-df78c13de661" containerName="probe" Feb 27 08:10:29 crc kubenswrapper[4612]: I0227 08:10:29.499065 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="62b23ba9-5e68-4665-86cd-df78c13de661" containerName="probe" Feb 27 08:10:29 crc kubenswrapper[4612]: E0227 08:10:29.499089 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="62b23ba9-5e68-4665-86cd-df78c13de661" containerName="cinder-scheduler" Feb 27 08:10:29 crc kubenswrapper[4612]: I0227 08:10:29.499095 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="62b23ba9-5e68-4665-86cd-df78c13de661" containerName="cinder-scheduler" Feb 27 08:10:29 crc kubenswrapper[4612]: E0227 08:10:29.499109 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="073232e9-6f4e-409e-bf29-989b354fcc39" containerName="oc" Feb 27 08:10:29 crc kubenswrapper[4612]: I0227 08:10:29.499116 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="073232e9-6f4e-409e-bf29-989b354fcc39" containerName="oc" Feb 27 08:10:29 crc kubenswrapper[4612]: E0227 08:10:29.499131 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="61899b16-fb85-4ba6-b057-65ad130ea015" containerName="dnsmasq-dns" Feb 27 08:10:29 crc kubenswrapper[4612]: I0227 08:10:29.499140 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="61899b16-fb85-4ba6-b057-65ad130ea015" containerName="dnsmasq-dns" Feb 27 08:10:29 crc kubenswrapper[4612]: E0227 08:10:29.499156 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="61899b16-fb85-4ba6-b057-65ad130ea015" containerName="init" Feb 27 08:10:29 crc kubenswrapper[4612]: I0227 08:10:29.499164 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="61899b16-fb85-4ba6-b057-65ad130ea015" containerName="init" Feb 27 08:10:29 crc kubenswrapper[4612]: I0227 08:10:29.499342 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="073232e9-6f4e-409e-bf29-989b354fcc39" containerName="oc" Feb 27 08:10:29 crc kubenswrapper[4612]: I0227 08:10:29.499352 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="62b23ba9-5e68-4665-86cd-df78c13de661" containerName="cinder-scheduler" Feb 27 08:10:29 crc kubenswrapper[4612]: I0227 08:10:29.499373 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="61899b16-fb85-4ba6-b057-65ad130ea015" containerName="dnsmasq-dns" Feb 27 08:10:29 crc kubenswrapper[4612]: I0227 08:10:29.499383 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="62b23ba9-5e68-4665-86cd-df78c13de661" containerName="probe" Feb 27 08:10:29 crc kubenswrapper[4612]: I0227 08:10:29.500063 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Feb 27 08:10:29 crc kubenswrapper[4612]: I0227 08:10:29.502627 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Feb 27 08:10:29 crc kubenswrapper[4612]: I0227 08:10:29.512629 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Feb 27 08:10:29 crc kubenswrapper[4612]: I0227 08:10:29.515412 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Feb 27 08:10:29 crc kubenswrapper[4612]: I0227 08:10:29.515649 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-crvfk" Feb 27 08:10:29 crc kubenswrapper[4612]: I0227 08:10:29.632524 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/9fd44447-aa5f-436f-8aaa-bf3b53d99e64-openstack-config\") pod \"openstackclient\" (UID: \"9fd44447-aa5f-436f-8aaa-bf3b53d99e64\") " pod="openstack/openstackclient" Feb 27 08:10:29 crc kubenswrapper[4612]: I0227 08:10:29.632589 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gm7zx\" (UniqueName: \"kubernetes.io/projected/9fd44447-aa5f-436f-8aaa-bf3b53d99e64-kube-api-access-gm7zx\") pod \"openstackclient\" (UID: \"9fd44447-aa5f-436f-8aaa-bf3b53d99e64\") " pod="openstack/openstackclient" Feb 27 08:10:29 crc kubenswrapper[4612]: I0227 08:10:29.632797 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9fd44447-aa5f-436f-8aaa-bf3b53d99e64-combined-ca-bundle\") pod \"openstackclient\" (UID: \"9fd44447-aa5f-436f-8aaa-bf3b53d99e64\") " pod="openstack/openstackclient" Feb 27 08:10:29 crc kubenswrapper[4612]: I0227 08:10:29.632853 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/9fd44447-aa5f-436f-8aaa-bf3b53d99e64-openstack-config-secret\") pod \"openstackclient\" (UID: \"9fd44447-aa5f-436f-8aaa-bf3b53d99e64\") " pod="openstack/openstackclient" Feb 27 08:10:29 crc kubenswrapper[4612]: I0227 08:10:29.734873 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9fd44447-aa5f-436f-8aaa-bf3b53d99e64-combined-ca-bundle\") pod \"openstackclient\" (UID: \"9fd44447-aa5f-436f-8aaa-bf3b53d99e64\") " pod="openstack/openstackclient" Feb 27 08:10:29 crc kubenswrapper[4612]: I0227 08:10:29.734927 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/9fd44447-aa5f-436f-8aaa-bf3b53d99e64-openstack-config-secret\") pod \"openstackclient\" (UID: \"9fd44447-aa5f-436f-8aaa-bf3b53d99e64\") " pod="openstack/openstackclient" Feb 27 08:10:29 crc kubenswrapper[4612]: I0227 08:10:29.734959 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/9fd44447-aa5f-436f-8aaa-bf3b53d99e64-openstack-config\") pod \"openstackclient\" (UID: \"9fd44447-aa5f-436f-8aaa-bf3b53d99e64\") " pod="openstack/openstackclient" Feb 27 08:10:29 crc kubenswrapper[4612]: I0227 08:10:29.734983 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gm7zx\" (UniqueName: \"kubernetes.io/projected/9fd44447-aa5f-436f-8aaa-bf3b53d99e64-kube-api-access-gm7zx\") pod \"openstackclient\" (UID: \"9fd44447-aa5f-436f-8aaa-bf3b53d99e64\") " pod="openstack/openstackclient" Feb 27 08:10:29 crc kubenswrapper[4612]: I0227 08:10:29.736091 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/9fd44447-aa5f-436f-8aaa-bf3b53d99e64-openstack-config\") pod \"openstackclient\" (UID: \"9fd44447-aa5f-436f-8aaa-bf3b53d99e64\") " pod="openstack/openstackclient" Feb 27 08:10:29 crc kubenswrapper[4612]: I0227 08:10:29.741268 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/9fd44447-aa5f-436f-8aaa-bf3b53d99e64-openstack-config-secret\") pod \"openstackclient\" (UID: \"9fd44447-aa5f-436f-8aaa-bf3b53d99e64\") " pod="openstack/openstackclient" Feb 27 08:10:29 crc kubenswrapper[4612]: I0227 08:10:29.743244 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9fd44447-aa5f-436f-8aaa-bf3b53d99e64-combined-ca-bundle\") pod \"openstackclient\" (UID: \"9fd44447-aa5f-436f-8aaa-bf3b53d99e64\") " pod="openstack/openstackclient" Feb 27 08:10:29 crc kubenswrapper[4612]: I0227 08:10:29.753507 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gm7zx\" (UniqueName: \"kubernetes.io/projected/9fd44447-aa5f-436f-8aaa-bf3b53d99e64-kube-api-access-gm7zx\") pod \"openstackclient\" (UID: \"9fd44447-aa5f-436f-8aaa-bf3b53d99e64\") " pod="openstack/openstackclient" Feb 27 08:10:29 crc kubenswrapper[4612]: I0227 08:10:29.771624 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Feb 27 08:10:29 crc kubenswrapper[4612]: I0227 08:10:29.818423 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Feb 27 08:10:29 crc kubenswrapper[4612]: I0227 08:10:29.827966 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Feb 27 08:10:29 crc kubenswrapper[4612]: I0227 08:10:29.841810 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Feb 27 08:10:29 crc kubenswrapper[4612]: I0227 08:10:29.865350 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Feb 27 08:10:29 crc kubenswrapper[4612]: I0227 08:10:29.875953 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Feb 27 08:10:29 crc kubenswrapper[4612]: I0227 08:10:29.883515 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Feb 27 08:10:29 crc kubenswrapper[4612]: I0227 08:10:29.923491 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Feb 27 08:10:29 crc kubenswrapper[4612]: I0227 08:10:29.943124 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0cf1b618-59b9-4688-85b9-1a7f6d2af657-config-data\") pod \"cinder-scheduler-0\" (UID: \"0cf1b618-59b9-4688-85b9-1a7f6d2af657\") " pod="openstack/cinder-scheduler-0" Feb 27 08:10:29 crc kubenswrapper[4612]: I0227 08:10:29.943177 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0cf1b618-59b9-4688-85b9-1a7f6d2af657-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"0cf1b618-59b9-4688-85b9-1a7f6d2af657\") " pod="openstack/cinder-scheduler-0" Feb 27 08:10:29 crc kubenswrapper[4612]: I0227 08:10:29.943327 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fr8pp\" (UniqueName: \"kubernetes.io/projected/0cf1b618-59b9-4688-85b9-1a7f6d2af657-kube-api-access-fr8pp\") pod \"cinder-scheduler-0\" (UID: \"0cf1b618-59b9-4688-85b9-1a7f6d2af657\") " pod="openstack/cinder-scheduler-0" Feb 27 08:10:29 crc kubenswrapper[4612]: I0227 08:10:29.943418 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0cf1b618-59b9-4688-85b9-1a7f6d2af657-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"0cf1b618-59b9-4688-85b9-1a7f6d2af657\") " pod="openstack/cinder-scheduler-0" Feb 27 08:10:29 crc kubenswrapper[4612]: I0227 08:10:29.945626 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0cf1b618-59b9-4688-85b9-1a7f6d2af657-scripts\") pod \"cinder-scheduler-0\" (UID: \"0cf1b618-59b9-4688-85b9-1a7f6d2af657\") " pod="openstack/cinder-scheduler-0" Feb 27 08:10:29 crc kubenswrapper[4612]: I0227 08:10:29.945661 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0cf1b618-59b9-4688-85b9-1a7f6d2af657-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"0cf1b618-59b9-4688-85b9-1a7f6d2af657\") " pod="openstack/cinder-scheduler-0" Feb 27 08:10:30 crc kubenswrapper[4612]: I0227 08:10:30.047939 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0cf1b618-59b9-4688-85b9-1a7f6d2af657-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"0cf1b618-59b9-4688-85b9-1a7f6d2af657\") " pod="openstack/cinder-scheduler-0" Feb 27 08:10:30 crc kubenswrapper[4612]: I0227 08:10:30.048079 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0cf1b618-59b9-4688-85b9-1a7f6d2af657-config-data\") pod \"cinder-scheduler-0\" (UID: \"0cf1b618-59b9-4688-85b9-1a7f6d2af657\") " pod="openstack/cinder-scheduler-0" Feb 27 08:10:30 crc kubenswrapper[4612]: I0227 08:10:30.048102 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0cf1b618-59b9-4688-85b9-1a7f6d2af657-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"0cf1b618-59b9-4688-85b9-1a7f6d2af657\") " pod="openstack/cinder-scheduler-0" Feb 27 08:10:30 crc kubenswrapper[4612]: I0227 08:10:30.048171 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fr8pp\" (UniqueName: \"kubernetes.io/projected/0cf1b618-59b9-4688-85b9-1a7f6d2af657-kube-api-access-fr8pp\") pod \"cinder-scheduler-0\" (UID: \"0cf1b618-59b9-4688-85b9-1a7f6d2af657\") " pod="openstack/cinder-scheduler-0" Feb 27 08:10:30 crc kubenswrapper[4612]: I0227 08:10:30.048215 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0cf1b618-59b9-4688-85b9-1a7f6d2af657-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"0cf1b618-59b9-4688-85b9-1a7f6d2af657\") " pod="openstack/cinder-scheduler-0" Feb 27 08:10:30 crc kubenswrapper[4612]: I0227 08:10:30.048261 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0cf1b618-59b9-4688-85b9-1a7f6d2af657-scripts\") pod \"cinder-scheduler-0\" (UID: \"0cf1b618-59b9-4688-85b9-1a7f6d2af657\") " pod="openstack/cinder-scheduler-0" Feb 27 08:10:30 crc kubenswrapper[4612]: I0227 08:10:30.050055 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0cf1b618-59b9-4688-85b9-1a7f6d2af657-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"0cf1b618-59b9-4688-85b9-1a7f6d2af657\") " pod="openstack/cinder-scheduler-0" Feb 27 08:10:30 crc kubenswrapper[4612]: I0227 08:10:30.053093 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0cf1b618-59b9-4688-85b9-1a7f6d2af657-scripts\") pod \"cinder-scheduler-0\" (UID: \"0cf1b618-59b9-4688-85b9-1a7f6d2af657\") " pod="openstack/cinder-scheduler-0" Feb 27 08:10:30 crc kubenswrapper[4612]: I0227 08:10:30.057321 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0cf1b618-59b9-4688-85b9-1a7f6d2af657-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"0cf1b618-59b9-4688-85b9-1a7f6d2af657\") " pod="openstack/cinder-scheduler-0" Feb 27 08:10:30 crc kubenswrapper[4612]: I0227 08:10:30.058654 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0cf1b618-59b9-4688-85b9-1a7f6d2af657-config-data\") pod \"cinder-scheduler-0\" (UID: \"0cf1b618-59b9-4688-85b9-1a7f6d2af657\") " pod="openstack/cinder-scheduler-0" Feb 27 08:10:30 crc kubenswrapper[4612]: I0227 08:10:30.062323 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0cf1b618-59b9-4688-85b9-1a7f6d2af657-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"0cf1b618-59b9-4688-85b9-1a7f6d2af657\") " pod="openstack/cinder-scheduler-0" Feb 27 08:10:30 crc kubenswrapper[4612]: I0227 08:10:30.085658 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fr8pp\" (UniqueName: \"kubernetes.io/projected/0cf1b618-59b9-4688-85b9-1a7f6d2af657-kube-api-access-fr8pp\") pod \"cinder-scheduler-0\" (UID: \"0cf1b618-59b9-4688-85b9-1a7f6d2af657\") " pod="openstack/cinder-scheduler-0" Feb 27 08:10:30 crc kubenswrapper[4612]: I0227 08:10:30.282505 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Feb 27 08:10:30 crc kubenswrapper[4612]: I0227 08:10:30.482636 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Feb 27 08:10:30 crc kubenswrapper[4612]: I0227 08:10:30.785051 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"9fd44447-aa5f-436f-8aaa-bf3b53d99e64","Type":"ContainerStarted","Data":"728b3d028b0d1360b1dc28193f24a72f63638cd97fade68d7dd4b90127035290"} Feb 27 08:10:30 crc kubenswrapper[4612]: I0227 08:10:30.862852 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="62b23ba9-5e68-4665-86cd-df78c13de661" path="/var/lib/kubelet/pods/62b23ba9-5e68-4665-86cd-df78c13de661/volumes" Feb 27 08:10:30 crc kubenswrapper[4612]: W0227 08:10:30.939565 4612 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0cf1b618_59b9_4688_85b9_1a7f6d2af657.slice/crio-a94ffb8d465f5207e6f69b34d67dc2551a047d03ad836e2f26a408601ff88557 WatchSource:0}: Error finding container a94ffb8d465f5207e6f69b34d67dc2551a047d03ad836e2f26a408601ff88557: Status 404 returned error can't find the container with id a94ffb8d465f5207e6f69b34d67dc2551a047d03ad836e2f26a408601ff88557 Feb 27 08:10:30 crc kubenswrapper[4612]: I0227 08:10:30.945540 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Feb 27 08:10:31 crc kubenswrapper[4612]: I0227 08:10:31.818830 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"0cf1b618-59b9-4688-85b9-1a7f6d2af657","Type":"ContainerStarted","Data":"d8dab9418257228988b15c18c519983334f70d77b3bf48cd8c5e3e24645b141d"} Feb 27 08:10:31 crc kubenswrapper[4612]: I0227 08:10:31.819087 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"0cf1b618-59b9-4688-85b9-1a7f6d2af657","Type":"ContainerStarted","Data":"a94ffb8d465f5207e6f69b34d67dc2551a047d03ad836e2f26a408601ff88557"} Feb 27 08:10:32 crc kubenswrapper[4612]: I0227 08:10:32.837091 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"0cf1b618-59b9-4688-85b9-1a7f6d2af657","Type":"ContainerStarted","Data":"9f41cdb7bdfe7d603d53391385925b109264102dab986b9e1b7fce8ef7685d6b"} Feb 27 08:10:32 crc kubenswrapper[4612]: I0227 08:10:32.860337 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.860321126 podStartE2EDuration="3.860321126s" podCreationTimestamp="2026-02-27 08:10:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 08:10:32.854291374 +0000 UTC m=+1290.708221372" watchObservedRunningTime="2026-02-27 08:10:32.860321126 +0000 UTC m=+1290.714251124" Feb 27 08:10:35 crc kubenswrapper[4612]: I0227 08:10:35.282846 4612 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Feb 27 08:10:35 crc kubenswrapper[4612]: I0227 08:10:35.457193 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-proxy-59ddd57cfc-qvv4h"] Feb 27 08:10:35 crc kubenswrapper[4612]: I0227 08:10:35.458773 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-59ddd57cfc-qvv4h" Feb 27 08:10:35 crc kubenswrapper[4612]: I0227 08:10:35.467854 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Feb 27 08:10:35 crc kubenswrapper[4612]: I0227 08:10:35.468255 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-public-svc" Feb 27 08:10:35 crc kubenswrapper[4612]: I0227 08:10:35.468441 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-internal-svc" Feb 27 08:10:35 crc kubenswrapper[4612]: I0227 08:10:35.482559 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-59ddd57cfc-qvv4h"] Feb 27 08:10:35 crc kubenswrapper[4612]: I0227 08:10:35.568477 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e631b0d-d44b-472b-90d9-6794cf71d3fd-config-data\") pod \"swift-proxy-59ddd57cfc-qvv4h\" (UID: \"2e631b0d-d44b-472b-90d9-6794cf71d3fd\") " pod="openstack/swift-proxy-59ddd57cfc-qvv4h" Feb 27 08:10:35 crc kubenswrapper[4612]: I0227 08:10:35.568900 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2e631b0d-d44b-472b-90d9-6794cf71d3fd-internal-tls-certs\") pod \"swift-proxy-59ddd57cfc-qvv4h\" (UID: \"2e631b0d-d44b-472b-90d9-6794cf71d3fd\") " pod="openstack/swift-proxy-59ddd57cfc-qvv4h" Feb 27 08:10:35 crc kubenswrapper[4612]: I0227 08:10:35.568930 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w4kct\" (UniqueName: \"kubernetes.io/projected/2e631b0d-d44b-472b-90d9-6794cf71d3fd-kube-api-access-w4kct\") pod \"swift-proxy-59ddd57cfc-qvv4h\" (UID: \"2e631b0d-d44b-472b-90d9-6794cf71d3fd\") " pod="openstack/swift-proxy-59ddd57cfc-qvv4h" Feb 27 08:10:35 crc kubenswrapper[4612]: I0227 08:10:35.568950 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2e631b0d-d44b-472b-90d9-6794cf71d3fd-log-httpd\") pod \"swift-proxy-59ddd57cfc-qvv4h\" (UID: \"2e631b0d-d44b-472b-90d9-6794cf71d3fd\") " pod="openstack/swift-proxy-59ddd57cfc-qvv4h" Feb 27 08:10:35 crc kubenswrapper[4612]: I0227 08:10:35.568969 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2e631b0d-d44b-472b-90d9-6794cf71d3fd-public-tls-certs\") pod \"swift-proxy-59ddd57cfc-qvv4h\" (UID: \"2e631b0d-d44b-472b-90d9-6794cf71d3fd\") " pod="openstack/swift-proxy-59ddd57cfc-qvv4h" Feb 27 08:10:35 crc kubenswrapper[4612]: I0227 08:10:35.569022 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/2e631b0d-d44b-472b-90d9-6794cf71d3fd-etc-swift\") pod \"swift-proxy-59ddd57cfc-qvv4h\" (UID: \"2e631b0d-d44b-472b-90d9-6794cf71d3fd\") " pod="openstack/swift-proxy-59ddd57cfc-qvv4h" Feb 27 08:10:35 crc kubenswrapper[4612]: I0227 08:10:35.569041 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e631b0d-d44b-472b-90d9-6794cf71d3fd-combined-ca-bundle\") pod \"swift-proxy-59ddd57cfc-qvv4h\" (UID: \"2e631b0d-d44b-472b-90d9-6794cf71d3fd\") " pod="openstack/swift-proxy-59ddd57cfc-qvv4h" Feb 27 08:10:35 crc kubenswrapper[4612]: I0227 08:10:35.569063 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2e631b0d-d44b-472b-90d9-6794cf71d3fd-run-httpd\") pod \"swift-proxy-59ddd57cfc-qvv4h\" (UID: \"2e631b0d-d44b-472b-90d9-6794cf71d3fd\") " pod="openstack/swift-proxy-59ddd57cfc-qvv4h" Feb 27 08:10:35 crc kubenswrapper[4612]: I0227 08:10:35.670524 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2e631b0d-d44b-472b-90d9-6794cf71d3fd-internal-tls-certs\") pod \"swift-proxy-59ddd57cfc-qvv4h\" (UID: \"2e631b0d-d44b-472b-90d9-6794cf71d3fd\") " pod="openstack/swift-proxy-59ddd57cfc-qvv4h" Feb 27 08:10:35 crc kubenswrapper[4612]: I0227 08:10:35.671431 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w4kct\" (UniqueName: \"kubernetes.io/projected/2e631b0d-d44b-472b-90d9-6794cf71d3fd-kube-api-access-w4kct\") pod \"swift-proxy-59ddd57cfc-qvv4h\" (UID: \"2e631b0d-d44b-472b-90d9-6794cf71d3fd\") " pod="openstack/swift-proxy-59ddd57cfc-qvv4h" Feb 27 08:10:35 crc kubenswrapper[4612]: I0227 08:10:35.671461 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2e631b0d-d44b-472b-90d9-6794cf71d3fd-log-httpd\") pod \"swift-proxy-59ddd57cfc-qvv4h\" (UID: \"2e631b0d-d44b-472b-90d9-6794cf71d3fd\") " pod="openstack/swift-proxy-59ddd57cfc-qvv4h" Feb 27 08:10:35 crc kubenswrapper[4612]: I0227 08:10:35.671490 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2e631b0d-d44b-472b-90d9-6794cf71d3fd-public-tls-certs\") pod \"swift-proxy-59ddd57cfc-qvv4h\" (UID: \"2e631b0d-d44b-472b-90d9-6794cf71d3fd\") " pod="openstack/swift-proxy-59ddd57cfc-qvv4h" Feb 27 08:10:35 crc kubenswrapper[4612]: I0227 08:10:35.671554 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/2e631b0d-d44b-472b-90d9-6794cf71d3fd-etc-swift\") pod \"swift-proxy-59ddd57cfc-qvv4h\" (UID: \"2e631b0d-d44b-472b-90d9-6794cf71d3fd\") " pod="openstack/swift-proxy-59ddd57cfc-qvv4h" Feb 27 08:10:35 crc kubenswrapper[4612]: I0227 08:10:35.671577 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e631b0d-d44b-472b-90d9-6794cf71d3fd-combined-ca-bundle\") pod \"swift-proxy-59ddd57cfc-qvv4h\" (UID: \"2e631b0d-d44b-472b-90d9-6794cf71d3fd\") " pod="openstack/swift-proxy-59ddd57cfc-qvv4h" Feb 27 08:10:35 crc kubenswrapper[4612]: I0227 08:10:35.671598 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2e631b0d-d44b-472b-90d9-6794cf71d3fd-run-httpd\") pod \"swift-proxy-59ddd57cfc-qvv4h\" (UID: \"2e631b0d-d44b-472b-90d9-6794cf71d3fd\") " pod="openstack/swift-proxy-59ddd57cfc-qvv4h" Feb 27 08:10:35 crc kubenswrapper[4612]: I0227 08:10:35.671662 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e631b0d-d44b-472b-90d9-6794cf71d3fd-config-data\") pod \"swift-proxy-59ddd57cfc-qvv4h\" (UID: \"2e631b0d-d44b-472b-90d9-6794cf71d3fd\") " pod="openstack/swift-proxy-59ddd57cfc-qvv4h" Feb 27 08:10:35 crc kubenswrapper[4612]: I0227 08:10:35.672195 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2e631b0d-d44b-472b-90d9-6794cf71d3fd-log-httpd\") pod \"swift-proxy-59ddd57cfc-qvv4h\" (UID: \"2e631b0d-d44b-472b-90d9-6794cf71d3fd\") " pod="openstack/swift-proxy-59ddd57cfc-qvv4h" Feb 27 08:10:35 crc kubenswrapper[4612]: I0227 08:10:35.673109 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2e631b0d-d44b-472b-90d9-6794cf71d3fd-run-httpd\") pod \"swift-proxy-59ddd57cfc-qvv4h\" (UID: \"2e631b0d-d44b-472b-90d9-6794cf71d3fd\") " pod="openstack/swift-proxy-59ddd57cfc-qvv4h" Feb 27 08:10:35 crc kubenswrapper[4612]: I0227 08:10:35.677577 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/2e631b0d-d44b-472b-90d9-6794cf71d3fd-etc-swift\") pod \"swift-proxy-59ddd57cfc-qvv4h\" (UID: \"2e631b0d-d44b-472b-90d9-6794cf71d3fd\") " pod="openstack/swift-proxy-59ddd57cfc-qvv4h" Feb 27 08:10:35 crc kubenswrapper[4612]: I0227 08:10:35.679779 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e631b0d-d44b-472b-90d9-6794cf71d3fd-config-data\") pod \"swift-proxy-59ddd57cfc-qvv4h\" (UID: \"2e631b0d-d44b-472b-90d9-6794cf71d3fd\") " pod="openstack/swift-proxy-59ddd57cfc-qvv4h" Feb 27 08:10:35 crc kubenswrapper[4612]: I0227 08:10:35.680321 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2e631b0d-d44b-472b-90d9-6794cf71d3fd-public-tls-certs\") pod \"swift-proxy-59ddd57cfc-qvv4h\" (UID: \"2e631b0d-d44b-472b-90d9-6794cf71d3fd\") " pod="openstack/swift-proxy-59ddd57cfc-qvv4h" Feb 27 08:10:35 crc kubenswrapper[4612]: I0227 08:10:35.680469 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e631b0d-d44b-472b-90d9-6794cf71d3fd-combined-ca-bundle\") pod \"swift-proxy-59ddd57cfc-qvv4h\" (UID: \"2e631b0d-d44b-472b-90d9-6794cf71d3fd\") " pod="openstack/swift-proxy-59ddd57cfc-qvv4h" Feb 27 08:10:35 crc kubenswrapper[4612]: I0227 08:10:35.687000 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2e631b0d-d44b-472b-90d9-6794cf71d3fd-internal-tls-certs\") pod \"swift-proxy-59ddd57cfc-qvv4h\" (UID: \"2e631b0d-d44b-472b-90d9-6794cf71d3fd\") " pod="openstack/swift-proxy-59ddd57cfc-qvv4h" Feb 27 08:10:35 crc kubenswrapper[4612]: I0227 08:10:35.694951 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w4kct\" (UniqueName: \"kubernetes.io/projected/2e631b0d-d44b-472b-90d9-6794cf71d3fd-kube-api-access-w4kct\") pod \"swift-proxy-59ddd57cfc-qvv4h\" (UID: \"2e631b0d-d44b-472b-90d9-6794cf71d3fd\") " pod="openstack/swift-proxy-59ddd57cfc-qvv4h" Feb 27 08:10:35 crc kubenswrapper[4612]: I0227 08:10:35.774990 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-59ddd57cfc-qvv4h" Feb 27 08:10:36 crc kubenswrapper[4612]: I0227 08:10:36.372566 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-cc6d746d7-b8wbb" Feb 27 08:10:36 crc kubenswrapper[4612]: I0227 08:10:36.449463 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-77854b564-8nkbp"] Feb 27 08:10:36 crc kubenswrapper[4612]: I0227 08:10:36.449682 4612 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-77854b564-8nkbp" podUID="a9397555-a3e3-4a49-b725-882006d2ef0d" containerName="neutron-api" containerID="cri-o://a282f0ed98a69a8ea94fc9525a85432465089708bc9f78755024b91a5fd52a03" gracePeriod=30 Feb 27 08:10:36 crc kubenswrapper[4612]: I0227 08:10:36.450066 4612 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-77854b564-8nkbp" podUID="a9397555-a3e3-4a49-b725-882006d2ef0d" containerName="neutron-httpd" containerID="cri-o://99db33fb196282606d81b167cb3314cc2ecbee32c696435c83491070eba30343" gracePeriod=30 Feb 27 08:10:36 crc kubenswrapper[4612]: I0227 08:10:36.585433 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-59ddd57cfc-qvv4h"] Feb 27 08:10:36 crc kubenswrapper[4612]: I0227 08:10:36.894516 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-59ddd57cfc-qvv4h" event={"ID":"2e631b0d-d44b-472b-90d9-6794cf71d3fd","Type":"ContainerStarted","Data":"d286f0766cc665faf8e50dcadfcaf9854bffbeb9c9c3b476275ed118407b7ef5"} Feb 27 08:10:37 crc kubenswrapper[4612]: I0227 08:10:37.593345 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 27 08:10:37 crc kubenswrapper[4612]: I0227 08:10:37.593875 4612 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="39aaf83c-4be0-4bd8-ae54-cb38271a5eb1" containerName="ceilometer-central-agent" containerID="cri-o://57e5093fb9f40dabcd4b4a0f88dd15403373494bdf7de2f87e395233ca1b7d99" gracePeriod=30 Feb 27 08:10:37 crc kubenswrapper[4612]: I0227 08:10:37.594268 4612 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="39aaf83c-4be0-4bd8-ae54-cb38271a5eb1" containerName="proxy-httpd" containerID="cri-o://34e1ccd60759f81efa6f32f20783c47476ad2417b5e2dbeb9bd0428b84a2bf7d" gracePeriod=30 Feb 27 08:10:37 crc kubenswrapper[4612]: I0227 08:10:37.594313 4612 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="39aaf83c-4be0-4bd8-ae54-cb38271a5eb1" containerName="sg-core" containerID="cri-o://4998f78f73290e3176d379a64fecaed3854ebacac2c7b7154e92c1cd0b3d0302" gracePeriod=30 Feb 27 08:10:37 crc kubenswrapper[4612]: I0227 08:10:37.594345 4612 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="39aaf83c-4be0-4bd8-ae54-cb38271a5eb1" containerName="ceilometer-notification-agent" containerID="cri-o://27ac184f71751114a7ced8f7acf694e9812f35974b862b9042adbc00259c5c4b" gracePeriod=30 Feb 27 08:10:37 crc kubenswrapper[4612]: I0227 08:10:37.617677 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Feb 27 08:10:37 crc kubenswrapper[4612]: I0227 08:10:37.924117 4612 generic.go:334] "Generic (PLEG): container finished" podID="39aaf83c-4be0-4bd8-ae54-cb38271a5eb1" containerID="4998f78f73290e3176d379a64fecaed3854ebacac2c7b7154e92c1cd0b3d0302" exitCode=2 Feb 27 08:10:37 crc kubenswrapper[4612]: I0227 08:10:37.924418 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"39aaf83c-4be0-4bd8-ae54-cb38271a5eb1","Type":"ContainerDied","Data":"4998f78f73290e3176d379a64fecaed3854ebacac2c7b7154e92c1cd0b3d0302"} Feb 27 08:10:37 crc kubenswrapper[4612]: I0227 08:10:37.931706 4612 generic.go:334] "Generic (PLEG): container finished" podID="a9397555-a3e3-4a49-b725-882006d2ef0d" containerID="99db33fb196282606d81b167cb3314cc2ecbee32c696435c83491070eba30343" exitCode=0 Feb 27 08:10:37 crc kubenswrapper[4612]: I0227 08:10:37.931761 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-77854b564-8nkbp" event={"ID":"a9397555-a3e3-4a49-b725-882006d2ef0d","Type":"ContainerDied","Data":"99db33fb196282606d81b167cb3314cc2ecbee32c696435c83491070eba30343"} Feb 27 08:10:37 crc kubenswrapper[4612]: I0227 08:10:37.942101 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-59ddd57cfc-qvv4h" event={"ID":"2e631b0d-d44b-472b-90d9-6794cf71d3fd","Type":"ContainerStarted","Data":"386c59f79c2ff7fb454b9eb1d02c1afb0e5bd08d3df7c0bfed4e89d497ff7510"} Feb 27 08:10:37 crc kubenswrapper[4612]: I0227 08:10:37.942144 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-59ddd57cfc-qvv4h" event={"ID":"2e631b0d-d44b-472b-90d9-6794cf71d3fd","Type":"ContainerStarted","Data":"bb13ff90c6411b0705138a2cc0ea39cabbed1fafb5fd9c3df28c55b9d6aa384c"} Feb 27 08:10:37 crc kubenswrapper[4612]: I0227 08:10:37.942513 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-59ddd57cfc-qvv4h" Feb 27 08:10:37 crc kubenswrapper[4612]: I0227 08:10:37.942644 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-59ddd57cfc-qvv4h" Feb 27 08:10:37 crc kubenswrapper[4612]: I0227 08:10:37.999820 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-proxy-59ddd57cfc-qvv4h" podStartSLOduration=2.9998011460000003 podStartE2EDuration="2.999801146s" podCreationTimestamp="2026-02-27 08:10:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 08:10:37.996139831 +0000 UTC m=+1295.850069829" watchObservedRunningTime="2026-02-27 08:10:37.999801146 +0000 UTC m=+1295.853731144" Feb 27 08:10:38 crc kubenswrapper[4612]: I0227 08:10:38.952524 4612 generic.go:334] "Generic (PLEG): container finished" podID="39aaf83c-4be0-4bd8-ae54-cb38271a5eb1" containerID="27ac184f71751114a7ced8f7acf694e9812f35974b862b9042adbc00259c5c4b" exitCode=0 Feb 27 08:10:38 crc kubenswrapper[4612]: I0227 08:10:38.953349 4612 generic.go:334] "Generic (PLEG): container finished" podID="39aaf83c-4be0-4bd8-ae54-cb38271a5eb1" containerID="57e5093fb9f40dabcd4b4a0f88dd15403373494bdf7de2f87e395233ca1b7d99" exitCode=0 Feb 27 08:10:38 crc kubenswrapper[4612]: I0227 08:10:38.954125 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"39aaf83c-4be0-4bd8-ae54-cb38271a5eb1","Type":"ContainerDied","Data":"27ac184f71751114a7ced8f7acf694e9812f35974b862b9042adbc00259c5c4b"} Feb 27 08:10:38 crc kubenswrapper[4612]: I0227 08:10:38.954250 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"39aaf83c-4be0-4bd8-ae54-cb38271a5eb1","Type":"ContainerDied","Data":"57e5093fb9f40dabcd4b4a0f88dd15403373494bdf7de2f87e395233ca1b7d99"} Feb 27 08:10:39 crc kubenswrapper[4612]: I0227 08:10:39.979137 4612 generic.go:334] "Generic (PLEG): container finished" podID="39aaf83c-4be0-4bd8-ae54-cb38271a5eb1" containerID="34e1ccd60759f81efa6f32f20783c47476ad2417b5e2dbeb9bd0428b84a2bf7d" exitCode=0 Feb 27 08:10:39 crc kubenswrapper[4612]: I0227 08:10:39.979511 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"39aaf83c-4be0-4bd8-ae54-cb38271a5eb1","Type":"ContainerDied","Data":"34e1ccd60759f81efa6f32f20783c47476ad2417b5e2dbeb9bd0428b84a2bf7d"} Feb 27 08:10:40 crc kubenswrapper[4612]: I0227 08:10:40.555915 4612 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Feb 27 08:10:42 crc kubenswrapper[4612]: I0227 08:10:42.000238 4612 generic.go:334] "Generic (PLEG): container finished" podID="a9397555-a3e3-4a49-b725-882006d2ef0d" containerID="a282f0ed98a69a8ea94fc9525a85432465089708bc9f78755024b91a5fd52a03" exitCode=0 Feb 27 08:10:42 crc kubenswrapper[4612]: I0227 08:10:42.000320 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-77854b564-8nkbp" event={"ID":"a9397555-a3e3-4a49-b725-882006d2ef0d","Type":"ContainerDied","Data":"a282f0ed98a69a8ea94fc9525a85432465089708bc9f78755024b91a5fd52a03"} Feb 27 08:10:45 crc kubenswrapper[4612]: I0227 08:10:45.785731 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-59ddd57cfc-qvv4h" Feb 27 08:10:45 crc kubenswrapper[4612]: I0227 08:10:45.787656 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-59ddd57cfc-qvv4h" Feb 27 08:10:46 crc kubenswrapper[4612]: I0227 08:10:45.986559 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 27 08:10:46 crc kubenswrapper[4612]: I0227 08:10:46.027613 4612 patch_prober.go:28] interesting pod/machine-config-daemon-924vb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 27 08:10:46 crc kubenswrapper[4612]: I0227 08:10:46.027661 4612 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 27 08:10:46 crc kubenswrapper[4612]: I0227 08:10:46.049343 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"39aaf83c-4be0-4bd8-ae54-cb38271a5eb1","Type":"ContainerDied","Data":"c15e926ec4288e5cebe0309bb45c3d26e9cc12a650fdeb394602a26bce3029c1"} Feb 27 08:10:46 crc kubenswrapper[4612]: I0227 08:10:46.049391 4612 scope.go:117] "RemoveContainer" containerID="34e1ccd60759f81efa6f32f20783c47476ad2417b5e2dbeb9bd0428b84a2bf7d" Feb 27 08:10:46 crc kubenswrapper[4612]: I0227 08:10:46.049508 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 27 08:10:46 crc kubenswrapper[4612]: I0227 08:10:46.067053 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"9fd44447-aa5f-436f-8aaa-bf3b53d99e64","Type":"ContainerStarted","Data":"90e277bcaa2146d90a72628cec0af9fe97de11e2b81287795449471a66a6f0a8"} Feb 27 08:10:46 crc kubenswrapper[4612]: I0227 08:10:46.099382 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=1.9614313650000001 podStartE2EDuration="17.099360431s" podCreationTimestamp="2026-02-27 08:10:29 +0000 UTC" firstStartedPulling="2026-02-27 08:10:30.481040259 +0000 UTC m=+1288.334970257" lastFinishedPulling="2026-02-27 08:10:45.618969325 +0000 UTC m=+1303.472899323" observedRunningTime="2026-02-27 08:10:46.080483989 +0000 UTC m=+1303.934413987" watchObservedRunningTime="2026-02-27 08:10:46.099360431 +0000 UTC m=+1303.953290429" Feb 27 08:10:46 crc kubenswrapper[4612]: I0227 08:10:46.102921 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q4b6g\" (UniqueName: \"kubernetes.io/projected/39aaf83c-4be0-4bd8-ae54-cb38271a5eb1-kube-api-access-q4b6g\") pod \"39aaf83c-4be0-4bd8-ae54-cb38271a5eb1\" (UID: \"39aaf83c-4be0-4bd8-ae54-cb38271a5eb1\") " Feb 27 08:10:46 crc kubenswrapper[4612]: I0227 08:10:46.102995 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/39aaf83c-4be0-4bd8-ae54-cb38271a5eb1-scripts\") pod \"39aaf83c-4be0-4bd8-ae54-cb38271a5eb1\" (UID: \"39aaf83c-4be0-4bd8-ae54-cb38271a5eb1\") " Feb 27 08:10:46 crc kubenswrapper[4612]: I0227 08:10:46.103031 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/39aaf83c-4be0-4bd8-ae54-cb38271a5eb1-combined-ca-bundle\") pod \"39aaf83c-4be0-4bd8-ae54-cb38271a5eb1\" (UID: \"39aaf83c-4be0-4bd8-ae54-cb38271a5eb1\") " Feb 27 08:10:46 crc kubenswrapper[4612]: I0227 08:10:46.103096 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/39aaf83c-4be0-4bd8-ae54-cb38271a5eb1-log-httpd\") pod \"39aaf83c-4be0-4bd8-ae54-cb38271a5eb1\" (UID: \"39aaf83c-4be0-4bd8-ae54-cb38271a5eb1\") " Feb 27 08:10:46 crc kubenswrapper[4612]: I0227 08:10:46.103187 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/39aaf83c-4be0-4bd8-ae54-cb38271a5eb1-sg-core-conf-yaml\") pod \"39aaf83c-4be0-4bd8-ae54-cb38271a5eb1\" (UID: \"39aaf83c-4be0-4bd8-ae54-cb38271a5eb1\") " Feb 27 08:10:46 crc kubenswrapper[4612]: I0227 08:10:46.103224 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/39aaf83c-4be0-4bd8-ae54-cb38271a5eb1-config-data\") pod \"39aaf83c-4be0-4bd8-ae54-cb38271a5eb1\" (UID: \"39aaf83c-4be0-4bd8-ae54-cb38271a5eb1\") " Feb 27 08:10:46 crc kubenswrapper[4612]: I0227 08:10:46.103258 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/39aaf83c-4be0-4bd8-ae54-cb38271a5eb1-run-httpd\") pod \"39aaf83c-4be0-4bd8-ae54-cb38271a5eb1\" (UID: \"39aaf83c-4be0-4bd8-ae54-cb38271a5eb1\") " Feb 27 08:10:46 crc kubenswrapper[4612]: I0227 08:10:46.105559 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/39aaf83c-4be0-4bd8-ae54-cb38271a5eb1-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "39aaf83c-4be0-4bd8-ae54-cb38271a5eb1" (UID: "39aaf83c-4be0-4bd8-ae54-cb38271a5eb1"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 08:10:46 crc kubenswrapper[4612]: I0227 08:10:46.106214 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/39aaf83c-4be0-4bd8-ae54-cb38271a5eb1-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "39aaf83c-4be0-4bd8-ae54-cb38271a5eb1" (UID: "39aaf83c-4be0-4bd8-ae54-cb38271a5eb1"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 08:10:46 crc kubenswrapper[4612]: I0227 08:10:46.108883 4612 scope.go:117] "RemoveContainer" containerID="4998f78f73290e3176d379a64fecaed3854ebacac2c7b7154e92c1cd0b3d0302" Feb 27 08:10:46 crc kubenswrapper[4612]: I0227 08:10:46.109713 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/39aaf83c-4be0-4bd8-ae54-cb38271a5eb1-kube-api-access-q4b6g" (OuterVolumeSpecName: "kube-api-access-q4b6g") pod "39aaf83c-4be0-4bd8-ae54-cb38271a5eb1" (UID: "39aaf83c-4be0-4bd8-ae54-cb38271a5eb1"). InnerVolumeSpecName "kube-api-access-q4b6g". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 08:10:46 crc kubenswrapper[4612]: I0227 08:10:46.116445 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/39aaf83c-4be0-4bd8-ae54-cb38271a5eb1-scripts" (OuterVolumeSpecName: "scripts") pod "39aaf83c-4be0-4bd8-ae54-cb38271a5eb1" (UID: "39aaf83c-4be0-4bd8-ae54-cb38271a5eb1"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:10:46 crc kubenswrapper[4612]: I0227 08:10:46.133473 4612 scope.go:117] "RemoveContainer" containerID="27ac184f71751114a7ced8f7acf694e9812f35974b862b9042adbc00259c5c4b" Feb 27 08:10:46 crc kubenswrapper[4612]: I0227 08:10:46.146768 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-77854b564-8nkbp" Feb 27 08:10:46 crc kubenswrapper[4612]: I0227 08:10:46.190823 4612 scope.go:117] "RemoveContainer" containerID="57e5093fb9f40dabcd4b4a0f88dd15403373494bdf7de2f87e395233ca1b7d99" Feb 27 08:10:46 crc kubenswrapper[4612]: I0227 08:10:46.204898 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/39aaf83c-4be0-4bd8-ae54-cb38271a5eb1-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "39aaf83c-4be0-4bd8-ae54-cb38271a5eb1" (UID: "39aaf83c-4be0-4bd8-ae54-cb38271a5eb1"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:10:46 crc kubenswrapper[4612]: I0227 08:10:46.205639 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-njhrs\" (UniqueName: \"kubernetes.io/projected/a9397555-a3e3-4a49-b725-882006d2ef0d-kube-api-access-njhrs\") pod \"a9397555-a3e3-4a49-b725-882006d2ef0d\" (UID: \"a9397555-a3e3-4a49-b725-882006d2ef0d\") " Feb 27 08:10:46 crc kubenswrapper[4612]: I0227 08:10:46.206029 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/a9397555-a3e3-4a49-b725-882006d2ef0d-config\") pod \"a9397555-a3e3-4a49-b725-882006d2ef0d\" (UID: \"a9397555-a3e3-4a49-b725-882006d2ef0d\") " Feb 27 08:10:46 crc kubenswrapper[4612]: I0227 08:10:46.206067 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/a9397555-a3e3-4a49-b725-882006d2ef0d-ovndb-tls-certs\") pod \"a9397555-a3e3-4a49-b725-882006d2ef0d\" (UID: \"a9397555-a3e3-4a49-b725-882006d2ef0d\") " Feb 27 08:10:46 crc kubenswrapper[4612]: I0227 08:10:46.206096 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/a9397555-a3e3-4a49-b725-882006d2ef0d-httpd-config\") pod \"a9397555-a3e3-4a49-b725-882006d2ef0d\" (UID: \"a9397555-a3e3-4a49-b725-882006d2ef0d\") " Feb 27 08:10:46 crc kubenswrapper[4612]: I0227 08:10:46.206580 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a9397555-a3e3-4a49-b725-882006d2ef0d-combined-ca-bundle\") pod \"a9397555-a3e3-4a49-b725-882006d2ef0d\" (UID: \"a9397555-a3e3-4a49-b725-882006d2ef0d\") " Feb 27 08:10:46 crc kubenswrapper[4612]: I0227 08:10:46.207091 4612 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/39aaf83c-4be0-4bd8-ae54-cb38271a5eb1-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Feb 27 08:10:46 crc kubenswrapper[4612]: I0227 08:10:46.207111 4612 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/39aaf83c-4be0-4bd8-ae54-cb38271a5eb1-run-httpd\") on node \"crc\" DevicePath \"\"" Feb 27 08:10:46 crc kubenswrapper[4612]: I0227 08:10:46.207119 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q4b6g\" (UniqueName: \"kubernetes.io/projected/39aaf83c-4be0-4bd8-ae54-cb38271a5eb1-kube-api-access-q4b6g\") on node \"crc\" DevicePath \"\"" Feb 27 08:10:46 crc kubenswrapper[4612]: I0227 08:10:46.207129 4612 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/39aaf83c-4be0-4bd8-ae54-cb38271a5eb1-scripts\") on node \"crc\" DevicePath \"\"" Feb 27 08:10:46 crc kubenswrapper[4612]: I0227 08:10:46.207137 4612 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/39aaf83c-4be0-4bd8-ae54-cb38271a5eb1-log-httpd\") on node \"crc\" DevicePath \"\"" Feb 27 08:10:46 crc kubenswrapper[4612]: I0227 08:10:46.225521 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a9397555-a3e3-4a49-b725-882006d2ef0d-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "a9397555-a3e3-4a49-b725-882006d2ef0d" (UID: "a9397555-a3e3-4a49-b725-882006d2ef0d"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:10:46 crc kubenswrapper[4612]: I0227 08:10:46.226711 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a9397555-a3e3-4a49-b725-882006d2ef0d-kube-api-access-njhrs" (OuterVolumeSpecName: "kube-api-access-njhrs") pod "a9397555-a3e3-4a49-b725-882006d2ef0d" (UID: "a9397555-a3e3-4a49-b725-882006d2ef0d"). InnerVolumeSpecName "kube-api-access-njhrs". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 08:10:46 crc kubenswrapper[4612]: I0227 08:10:46.238518 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/39aaf83c-4be0-4bd8-ae54-cb38271a5eb1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "39aaf83c-4be0-4bd8-ae54-cb38271a5eb1" (UID: "39aaf83c-4be0-4bd8-ae54-cb38271a5eb1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:10:46 crc kubenswrapper[4612]: I0227 08:10:46.276930 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/39aaf83c-4be0-4bd8-ae54-cb38271a5eb1-config-data" (OuterVolumeSpecName: "config-data") pod "39aaf83c-4be0-4bd8-ae54-cb38271a5eb1" (UID: "39aaf83c-4be0-4bd8-ae54-cb38271a5eb1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:10:46 crc kubenswrapper[4612]: I0227 08:10:46.287577 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a9397555-a3e3-4a49-b725-882006d2ef0d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a9397555-a3e3-4a49-b725-882006d2ef0d" (UID: "a9397555-a3e3-4a49-b725-882006d2ef0d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:10:46 crc kubenswrapper[4612]: I0227 08:10:46.304437 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a9397555-a3e3-4a49-b725-882006d2ef0d-config" (OuterVolumeSpecName: "config") pod "a9397555-a3e3-4a49-b725-882006d2ef0d" (UID: "a9397555-a3e3-4a49-b725-882006d2ef0d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:10:46 crc kubenswrapper[4612]: I0227 08:10:46.307660 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a9397555-a3e3-4a49-b725-882006d2ef0d-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "a9397555-a3e3-4a49-b725-882006d2ef0d" (UID: "a9397555-a3e3-4a49-b725-882006d2ef0d"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:10:46 crc kubenswrapper[4612]: I0227 08:10:46.314767 4612 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/a9397555-a3e3-4a49-b725-882006d2ef0d-config\") on node \"crc\" DevicePath \"\"" Feb 27 08:10:46 crc kubenswrapper[4612]: I0227 08:10:46.314801 4612 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/a9397555-a3e3-4a49-b725-882006d2ef0d-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 27 08:10:46 crc kubenswrapper[4612]: I0227 08:10:46.314812 4612 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/a9397555-a3e3-4a49-b725-882006d2ef0d-httpd-config\") on node \"crc\" DevicePath \"\"" Feb 27 08:10:46 crc kubenswrapper[4612]: I0227 08:10:46.314821 4612 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/39aaf83c-4be0-4bd8-ae54-cb38271a5eb1-config-data\") on node \"crc\" DevicePath \"\"" Feb 27 08:10:46 crc kubenswrapper[4612]: I0227 08:10:46.314832 4612 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a9397555-a3e3-4a49-b725-882006d2ef0d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 08:10:46 crc kubenswrapper[4612]: I0227 08:10:46.314840 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-njhrs\" (UniqueName: \"kubernetes.io/projected/a9397555-a3e3-4a49-b725-882006d2ef0d-kube-api-access-njhrs\") on node \"crc\" DevicePath \"\"" Feb 27 08:10:46 crc kubenswrapper[4612]: I0227 08:10:46.314849 4612 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/39aaf83c-4be0-4bd8-ae54-cb38271a5eb1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 08:10:46 crc kubenswrapper[4612]: I0227 08:10:46.377068 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 27 08:10:46 crc kubenswrapper[4612]: I0227 08:10:46.387249 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Feb 27 08:10:46 crc kubenswrapper[4612]: I0227 08:10:46.410209 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Feb 27 08:10:46 crc kubenswrapper[4612]: E0227 08:10:46.422250 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a9397555-a3e3-4a49-b725-882006d2ef0d" containerName="neutron-httpd" Feb 27 08:10:46 crc kubenswrapper[4612]: I0227 08:10:46.422286 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="a9397555-a3e3-4a49-b725-882006d2ef0d" containerName="neutron-httpd" Feb 27 08:10:46 crc kubenswrapper[4612]: E0227 08:10:46.422317 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a9397555-a3e3-4a49-b725-882006d2ef0d" containerName="neutron-api" Feb 27 08:10:46 crc kubenswrapper[4612]: I0227 08:10:46.422327 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="a9397555-a3e3-4a49-b725-882006d2ef0d" containerName="neutron-api" Feb 27 08:10:46 crc kubenswrapper[4612]: E0227 08:10:46.422348 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="39aaf83c-4be0-4bd8-ae54-cb38271a5eb1" containerName="ceilometer-notification-agent" Feb 27 08:10:46 crc kubenswrapper[4612]: I0227 08:10:46.422354 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="39aaf83c-4be0-4bd8-ae54-cb38271a5eb1" containerName="ceilometer-notification-agent" Feb 27 08:10:46 crc kubenswrapper[4612]: E0227 08:10:46.422372 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="39aaf83c-4be0-4bd8-ae54-cb38271a5eb1" containerName="sg-core" Feb 27 08:10:46 crc kubenswrapper[4612]: I0227 08:10:46.422377 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="39aaf83c-4be0-4bd8-ae54-cb38271a5eb1" containerName="sg-core" Feb 27 08:10:46 crc kubenswrapper[4612]: E0227 08:10:46.422390 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="39aaf83c-4be0-4bd8-ae54-cb38271a5eb1" containerName="ceilometer-central-agent" Feb 27 08:10:46 crc kubenswrapper[4612]: I0227 08:10:46.422396 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="39aaf83c-4be0-4bd8-ae54-cb38271a5eb1" containerName="ceilometer-central-agent" Feb 27 08:10:46 crc kubenswrapper[4612]: E0227 08:10:46.422436 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="39aaf83c-4be0-4bd8-ae54-cb38271a5eb1" containerName="proxy-httpd" Feb 27 08:10:46 crc kubenswrapper[4612]: I0227 08:10:46.422442 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="39aaf83c-4be0-4bd8-ae54-cb38271a5eb1" containerName="proxy-httpd" Feb 27 08:10:46 crc kubenswrapper[4612]: I0227 08:10:46.422914 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="a9397555-a3e3-4a49-b725-882006d2ef0d" containerName="neutron-api" Feb 27 08:10:46 crc kubenswrapper[4612]: I0227 08:10:46.422936 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="39aaf83c-4be0-4bd8-ae54-cb38271a5eb1" containerName="proxy-httpd" Feb 27 08:10:46 crc kubenswrapper[4612]: I0227 08:10:46.422949 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="39aaf83c-4be0-4bd8-ae54-cb38271a5eb1" containerName="ceilometer-central-agent" Feb 27 08:10:46 crc kubenswrapper[4612]: I0227 08:10:46.422971 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="a9397555-a3e3-4a49-b725-882006d2ef0d" containerName="neutron-httpd" Feb 27 08:10:46 crc kubenswrapper[4612]: I0227 08:10:46.422979 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="39aaf83c-4be0-4bd8-ae54-cb38271a5eb1" containerName="sg-core" Feb 27 08:10:46 crc kubenswrapper[4612]: I0227 08:10:46.422999 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="39aaf83c-4be0-4bd8-ae54-cb38271a5eb1" containerName="ceilometer-notification-agent" Feb 27 08:10:46 crc kubenswrapper[4612]: I0227 08:10:46.426590 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 27 08:10:46 crc kubenswrapper[4612]: I0227 08:10:46.433302 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Feb 27 08:10:46 crc kubenswrapper[4612]: I0227 08:10:46.433524 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Feb 27 08:10:46 crc kubenswrapper[4612]: I0227 08:10:46.458785 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 27 08:10:46 crc kubenswrapper[4612]: I0227 08:10:46.526207 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9a05b2bc-7c21-412d-b7ba-26e77517f60d-config-data\") pod \"ceilometer-0\" (UID: \"9a05b2bc-7c21-412d-b7ba-26e77517f60d\") " pod="openstack/ceilometer-0" Feb 27 08:10:46 crc kubenswrapper[4612]: I0227 08:10:46.526288 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9a05b2bc-7c21-412d-b7ba-26e77517f60d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"9a05b2bc-7c21-412d-b7ba-26e77517f60d\") " pod="openstack/ceilometer-0" Feb 27 08:10:46 crc kubenswrapper[4612]: I0227 08:10:46.526315 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9a05b2bc-7c21-412d-b7ba-26e77517f60d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"9a05b2bc-7c21-412d-b7ba-26e77517f60d\") " pod="openstack/ceilometer-0" Feb 27 08:10:46 crc kubenswrapper[4612]: I0227 08:10:46.526341 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hnh2l\" (UniqueName: \"kubernetes.io/projected/9a05b2bc-7c21-412d-b7ba-26e77517f60d-kube-api-access-hnh2l\") pod \"ceilometer-0\" (UID: \"9a05b2bc-7c21-412d-b7ba-26e77517f60d\") " pod="openstack/ceilometer-0" Feb 27 08:10:46 crc kubenswrapper[4612]: I0227 08:10:46.526358 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9a05b2bc-7c21-412d-b7ba-26e77517f60d-log-httpd\") pod \"ceilometer-0\" (UID: \"9a05b2bc-7c21-412d-b7ba-26e77517f60d\") " pod="openstack/ceilometer-0" Feb 27 08:10:46 crc kubenswrapper[4612]: I0227 08:10:46.526379 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9a05b2bc-7c21-412d-b7ba-26e77517f60d-scripts\") pod \"ceilometer-0\" (UID: \"9a05b2bc-7c21-412d-b7ba-26e77517f60d\") " pod="openstack/ceilometer-0" Feb 27 08:10:46 crc kubenswrapper[4612]: I0227 08:10:46.526407 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9a05b2bc-7c21-412d-b7ba-26e77517f60d-run-httpd\") pod \"ceilometer-0\" (UID: \"9a05b2bc-7c21-412d-b7ba-26e77517f60d\") " pod="openstack/ceilometer-0" Feb 27 08:10:46 crc kubenswrapper[4612]: I0227 08:10:46.628021 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9a05b2bc-7c21-412d-b7ba-26e77517f60d-run-httpd\") pod \"ceilometer-0\" (UID: \"9a05b2bc-7c21-412d-b7ba-26e77517f60d\") " pod="openstack/ceilometer-0" Feb 27 08:10:46 crc kubenswrapper[4612]: I0227 08:10:46.628155 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9a05b2bc-7c21-412d-b7ba-26e77517f60d-config-data\") pod \"ceilometer-0\" (UID: \"9a05b2bc-7c21-412d-b7ba-26e77517f60d\") " pod="openstack/ceilometer-0" Feb 27 08:10:46 crc kubenswrapper[4612]: I0227 08:10:46.628210 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9a05b2bc-7c21-412d-b7ba-26e77517f60d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"9a05b2bc-7c21-412d-b7ba-26e77517f60d\") " pod="openstack/ceilometer-0" Feb 27 08:10:46 crc kubenswrapper[4612]: I0227 08:10:46.628246 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9a05b2bc-7c21-412d-b7ba-26e77517f60d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"9a05b2bc-7c21-412d-b7ba-26e77517f60d\") " pod="openstack/ceilometer-0" Feb 27 08:10:46 crc kubenswrapper[4612]: I0227 08:10:46.628283 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hnh2l\" (UniqueName: \"kubernetes.io/projected/9a05b2bc-7c21-412d-b7ba-26e77517f60d-kube-api-access-hnh2l\") pod \"ceilometer-0\" (UID: \"9a05b2bc-7c21-412d-b7ba-26e77517f60d\") " pod="openstack/ceilometer-0" Feb 27 08:10:46 crc kubenswrapper[4612]: I0227 08:10:46.628307 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9a05b2bc-7c21-412d-b7ba-26e77517f60d-log-httpd\") pod \"ceilometer-0\" (UID: \"9a05b2bc-7c21-412d-b7ba-26e77517f60d\") " pod="openstack/ceilometer-0" Feb 27 08:10:46 crc kubenswrapper[4612]: I0227 08:10:46.628338 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9a05b2bc-7c21-412d-b7ba-26e77517f60d-scripts\") pod \"ceilometer-0\" (UID: \"9a05b2bc-7c21-412d-b7ba-26e77517f60d\") " pod="openstack/ceilometer-0" Feb 27 08:10:46 crc kubenswrapper[4612]: I0227 08:10:46.629217 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9a05b2bc-7c21-412d-b7ba-26e77517f60d-run-httpd\") pod \"ceilometer-0\" (UID: \"9a05b2bc-7c21-412d-b7ba-26e77517f60d\") " pod="openstack/ceilometer-0" Feb 27 08:10:46 crc kubenswrapper[4612]: I0227 08:10:46.629646 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9a05b2bc-7c21-412d-b7ba-26e77517f60d-log-httpd\") pod \"ceilometer-0\" (UID: \"9a05b2bc-7c21-412d-b7ba-26e77517f60d\") " pod="openstack/ceilometer-0" Feb 27 08:10:46 crc kubenswrapper[4612]: I0227 08:10:46.633151 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9a05b2bc-7c21-412d-b7ba-26e77517f60d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"9a05b2bc-7c21-412d-b7ba-26e77517f60d\") " pod="openstack/ceilometer-0" Feb 27 08:10:46 crc kubenswrapper[4612]: I0227 08:10:46.639186 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9a05b2bc-7c21-412d-b7ba-26e77517f60d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"9a05b2bc-7c21-412d-b7ba-26e77517f60d\") " pod="openstack/ceilometer-0" Feb 27 08:10:46 crc kubenswrapper[4612]: I0227 08:10:46.639597 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9a05b2bc-7c21-412d-b7ba-26e77517f60d-scripts\") pod \"ceilometer-0\" (UID: \"9a05b2bc-7c21-412d-b7ba-26e77517f60d\") " pod="openstack/ceilometer-0" Feb 27 08:10:46 crc kubenswrapper[4612]: I0227 08:10:46.643746 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9a05b2bc-7c21-412d-b7ba-26e77517f60d-config-data\") pod \"ceilometer-0\" (UID: \"9a05b2bc-7c21-412d-b7ba-26e77517f60d\") " pod="openstack/ceilometer-0" Feb 27 08:10:46 crc kubenswrapper[4612]: I0227 08:10:46.645231 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hnh2l\" (UniqueName: \"kubernetes.io/projected/9a05b2bc-7c21-412d-b7ba-26e77517f60d-kube-api-access-hnh2l\") pod \"ceilometer-0\" (UID: \"9a05b2bc-7c21-412d-b7ba-26e77517f60d\") " pod="openstack/ceilometer-0" Feb 27 08:10:46 crc kubenswrapper[4612]: I0227 08:10:46.743897 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 27 08:10:46 crc kubenswrapper[4612]: I0227 08:10:46.871383 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="39aaf83c-4be0-4bd8-ae54-cb38271a5eb1" path="/var/lib/kubelet/pods/39aaf83c-4be0-4bd8-ae54-cb38271a5eb1/volumes" Feb 27 08:10:47 crc kubenswrapper[4612]: I0227 08:10:47.073798 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-77854b564-8nkbp" event={"ID":"a9397555-a3e3-4a49-b725-882006d2ef0d","Type":"ContainerDied","Data":"0b2aa2610321a1eab88845caf61ec74398c53cf0faad98a6a45c83957ad8e63e"} Feb 27 08:10:47 crc kubenswrapper[4612]: I0227 08:10:47.073852 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-77854b564-8nkbp" Feb 27 08:10:47 crc kubenswrapper[4612]: I0227 08:10:47.073867 4612 scope.go:117] "RemoveContainer" containerID="99db33fb196282606d81b167cb3314cc2ecbee32c696435c83491070eba30343" Feb 27 08:10:47 crc kubenswrapper[4612]: I0227 08:10:47.102666 4612 scope.go:117] "RemoveContainer" containerID="a282f0ed98a69a8ea94fc9525a85432465089708bc9f78755024b91a5fd52a03" Feb 27 08:10:47 crc kubenswrapper[4612]: I0227 08:10:47.106995 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-77854b564-8nkbp"] Feb 27 08:10:47 crc kubenswrapper[4612]: I0227 08:10:47.117791 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-77854b564-8nkbp"] Feb 27 08:10:47 crc kubenswrapper[4612]: I0227 08:10:47.238964 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 27 08:10:47 crc kubenswrapper[4612]: W0227 08:10:47.240163 4612 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9a05b2bc_7c21_412d_b7ba_26e77517f60d.slice/crio-f7d6c1da1dc0352307fccccf8c1045c0c7a5161657111f79d8d8dfab4d214fa1 WatchSource:0}: Error finding container f7d6c1da1dc0352307fccccf8c1045c0c7a5161657111f79d8d8dfab4d214fa1: Status 404 returned error can't find the container with id f7d6c1da1dc0352307fccccf8c1045c0c7a5161657111f79d8d8dfab4d214fa1 Feb 27 08:10:47 crc kubenswrapper[4612]: I0227 08:10:47.860955 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 27 08:10:47 crc kubenswrapper[4612]: I0227 08:10:47.900548 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Feb 27 08:10:47 crc kubenswrapper[4612]: I0227 08:10:47.900881 4612 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="7c51bfaf-2c6a-49aa-afe7-eb4789351857" containerName="kube-state-metrics" containerID="cri-o://6631aea4a97b9613eab069c3da8c94e2ca649e395147956fd3276596b9588520" gracePeriod=30 Feb 27 08:10:48 crc kubenswrapper[4612]: I0227 08:10:48.097183 4612 generic.go:334] "Generic (PLEG): container finished" podID="7c51bfaf-2c6a-49aa-afe7-eb4789351857" containerID="6631aea4a97b9613eab069c3da8c94e2ca649e395147956fd3276596b9588520" exitCode=2 Feb 27 08:10:48 crc kubenswrapper[4612]: I0227 08:10:48.097291 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"7c51bfaf-2c6a-49aa-afe7-eb4789351857","Type":"ContainerDied","Data":"6631aea4a97b9613eab069c3da8c94e2ca649e395147956fd3276596b9588520"} Feb 27 08:10:48 crc kubenswrapper[4612]: I0227 08:10:48.099661 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9a05b2bc-7c21-412d-b7ba-26e77517f60d","Type":"ContainerStarted","Data":"c9c7680af8a800caac979e94217d6e08c6a44613c883dc27baa42950bf08b3c4"} Feb 27 08:10:48 crc kubenswrapper[4612]: I0227 08:10:48.099722 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9a05b2bc-7c21-412d-b7ba-26e77517f60d","Type":"ContainerStarted","Data":"f7d6c1da1dc0352307fccccf8c1045c0c7a5161657111f79d8d8dfab4d214fa1"} Feb 27 08:10:48 crc kubenswrapper[4612]: I0227 08:10:48.353619 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Feb 27 08:10:48 crc kubenswrapper[4612]: I0227 08:10:48.469326 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b9h48\" (UniqueName: \"kubernetes.io/projected/7c51bfaf-2c6a-49aa-afe7-eb4789351857-kube-api-access-b9h48\") pod \"7c51bfaf-2c6a-49aa-afe7-eb4789351857\" (UID: \"7c51bfaf-2c6a-49aa-afe7-eb4789351857\") " Feb 27 08:10:48 crc kubenswrapper[4612]: I0227 08:10:48.476879 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7c51bfaf-2c6a-49aa-afe7-eb4789351857-kube-api-access-b9h48" (OuterVolumeSpecName: "kube-api-access-b9h48") pod "7c51bfaf-2c6a-49aa-afe7-eb4789351857" (UID: "7c51bfaf-2c6a-49aa-afe7-eb4789351857"). InnerVolumeSpecName "kube-api-access-b9h48". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 08:10:48 crc kubenswrapper[4612]: I0227 08:10:48.572107 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b9h48\" (UniqueName: \"kubernetes.io/projected/7c51bfaf-2c6a-49aa-afe7-eb4789351857-kube-api-access-b9h48\") on node \"crc\" DevicePath \"\"" Feb 27 08:10:48 crc kubenswrapper[4612]: I0227 08:10:48.863830 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a9397555-a3e3-4a49-b725-882006d2ef0d" path="/var/lib/kubelet/pods/a9397555-a3e3-4a49-b725-882006d2ef0d/volumes" Feb 27 08:10:49 crc kubenswrapper[4612]: I0227 08:10:49.076533 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Feb 27 08:10:49 crc kubenswrapper[4612]: I0227 08:10:49.124770 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"7c51bfaf-2c6a-49aa-afe7-eb4789351857","Type":"ContainerDied","Data":"a6c6141ac16efb9dd3d28c920fec9a6735719a3bc398e647edc3c62558c8250b"} Feb 27 08:10:49 crc kubenswrapper[4612]: I0227 08:10:49.124819 4612 scope.go:117] "RemoveContainer" containerID="6631aea4a97b9613eab069c3da8c94e2ca649e395147956fd3276596b9588520" Feb 27 08:10:49 crc kubenswrapper[4612]: I0227 08:10:49.124921 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Feb 27 08:10:49 crc kubenswrapper[4612]: I0227 08:10:49.135310 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9a05b2bc-7c21-412d-b7ba-26e77517f60d","Type":"ContainerStarted","Data":"031c23bd61e46974556b76e828661e3ce104da5c8221a415ba11063a8de5e6af"} Feb 27 08:10:49 crc kubenswrapper[4612]: I0227 08:10:49.145754 4612 generic.go:334] "Generic (PLEG): container finished" podID="9a40110d-5675-40b6-87c6-63b1d2faba9d" containerID="6a08fb66db51ba5788882f4d7cb293f89bccd46dba71bf71878951639068185f" exitCode=137 Feb 27 08:10:49 crc kubenswrapper[4612]: I0227 08:10:49.145846 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"9a40110d-5675-40b6-87c6-63b1d2faba9d","Type":"ContainerDied","Data":"6a08fb66db51ba5788882f4d7cb293f89bccd46dba71bf71878951639068185f"} Feb 27 08:10:49 crc kubenswrapper[4612]: I0227 08:10:49.145894 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"9a40110d-5675-40b6-87c6-63b1d2faba9d","Type":"ContainerDied","Data":"dbbfaac3a7d15a84f4fdaecc71f20fb511c1ae04bfb068e83c0f5eab387bc2f6"} Feb 27 08:10:49 crc kubenswrapper[4612]: I0227 08:10:49.145970 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Feb 27 08:10:49 crc kubenswrapper[4612]: I0227 08:10:49.149497 4612 scope.go:117] "RemoveContainer" containerID="6a08fb66db51ba5788882f4d7cb293f89bccd46dba71bf71878951639068185f" Feb 27 08:10:49 crc kubenswrapper[4612]: I0227 08:10:49.178798 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Feb 27 08:10:49 crc kubenswrapper[4612]: I0227 08:10:49.182154 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9a40110d-5675-40b6-87c6-63b1d2faba9d-scripts\") pod \"9a40110d-5675-40b6-87c6-63b1d2faba9d\" (UID: \"9a40110d-5675-40b6-87c6-63b1d2faba9d\") " Feb 27 08:10:49 crc kubenswrapper[4612]: I0227 08:10:49.182199 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9a40110d-5675-40b6-87c6-63b1d2faba9d-config-data\") pod \"9a40110d-5675-40b6-87c6-63b1d2faba9d\" (UID: \"9a40110d-5675-40b6-87c6-63b1d2faba9d\") " Feb 27 08:10:49 crc kubenswrapper[4612]: I0227 08:10:49.182222 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nkgbh\" (UniqueName: \"kubernetes.io/projected/9a40110d-5675-40b6-87c6-63b1d2faba9d-kube-api-access-nkgbh\") pod \"9a40110d-5675-40b6-87c6-63b1d2faba9d\" (UID: \"9a40110d-5675-40b6-87c6-63b1d2faba9d\") " Feb 27 08:10:49 crc kubenswrapper[4612]: I0227 08:10:49.182321 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9a40110d-5675-40b6-87c6-63b1d2faba9d-config-data-custom\") pod \"9a40110d-5675-40b6-87c6-63b1d2faba9d\" (UID: \"9a40110d-5675-40b6-87c6-63b1d2faba9d\") " Feb 27 08:10:49 crc kubenswrapper[4612]: I0227 08:10:49.182361 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9a40110d-5675-40b6-87c6-63b1d2faba9d-combined-ca-bundle\") pod \"9a40110d-5675-40b6-87c6-63b1d2faba9d\" (UID: \"9a40110d-5675-40b6-87c6-63b1d2faba9d\") " Feb 27 08:10:49 crc kubenswrapper[4612]: I0227 08:10:49.182430 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9a40110d-5675-40b6-87c6-63b1d2faba9d-logs\") pod \"9a40110d-5675-40b6-87c6-63b1d2faba9d\" (UID: \"9a40110d-5675-40b6-87c6-63b1d2faba9d\") " Feb 27 08:10:49 crc kubenswrapper[4612]: I0227 08:10:49.182525 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/9a40110d-5675-40b6-87c6-63b1d2faba9d-etc-machine-id\") pod \"9a40110d-5675-40b6-87c6-63b1d2faba9d\" (UID: \"9a40110d-5675-40b6-87c6-63b1d2faba9d\") " Feb 27 08:10:49 crc kubenswrapper[4612]: I0227 08:10:49.186197 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9a40110d-5675-40b6-87c6-63b1d2faba9d-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "9a40110d-5675-40b6-87c6-63b1d2faba9d" (UID: "9a40110d-5675-40b6-87c6-63b1d2faba9d"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 27 08:10:49 crc kubenswrapper[4612]: I0227 08:10:49.201389 4612 scope.go:117] "RemoveContainer" containerID="c198f8b781fb884ae7ba10334f69026ead837cda8cfc728772b0e7b942cb069e" Feb 27 08:10:49 crc kubenswrapper[4612]: I0227 08:10:49.201910 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9a40110d-5675-40b6-87c6-63b1d2faba9d-logs" (OuterVolumeSpecName: "logs") pod "9a40110d-5675-40b6-87c6-63b1d2faba9d" (UID: "9a40110d-5675-40b6-87c6-63b1d2faba9d"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 08:10:49 crc kubenswrapper[4612]: I0227 08:10:49.205800 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9a40110d-5675-40b6-87c6-63b1d2faba9d-scripts" (OuterVolumeSpecName: "scripts") pod "9a40110d-5675-40b6-87c6-63b1d2faba9d" (UID: "9a40110d-5675-40b6-87c6-63b1d2faba9d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:10:49 crc kubenswrapper[4612]: I0227 08:10:49.212874 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9a40110d-5675-40b6-87c6-63b1d2faba9d-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "9a40110d-5675-40b6-87c6-63b1d2faba9d" (UID: "9a40110d-5675-40b6-87c6-63b1d2faba9d"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:10:49 crc kubenswrapper[4612]: I0227 08:10:49.216200 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9a40110d-5675-40b6-87c6-63b1d2faba9d-kube-api-access-nkgbh" (OuterVolumeSpecName: "kube-api-access-nkgbh") pod "9a40110d-5675-40b6-87c6-63b1d2faba9d" (UID: "9a40110d-5675-40b6-87c6-63b1d2faba9d"). InnerVolumeSpecName "kube-api-access-nkgbh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 08:10:49 crc kubenswrapper[4612]: I0227 08:10:49.218820 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Feb 27 08:10:49 crc kubenswrapper[4612]: I0227 08:10:49.269832 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Feb 27 08:10:49 crc kubenswrapper[4612]: E0227 08:10:49.270542 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7c51bfaf-2c6a-49aa-afe7-eb4789351857" containerName="kube-state-metrics" Feb 27 08:10:49 crc kubenswrapper[4612]: I0227 08:10:49.270562 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c51bfaf-2c6a-49aa-afe7-eb4789351857" containerName="kube-state-metrics" Feb 27 08:10:49 crc kubenswrapper[4612]: E0227 08:10:49.270574 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a40110d-5675-40b6-87c6-63b1d2faba9d" containerName="cinder-api-log" Feb 27 08:10:49 crc kubenswrapper[4612]: I0227 08:10:49.270581 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a40110d-5675-40b6-87c6-63b1d2faba9d" containerName="cinder-api-log" Feb 27 08:10:49 crc kubenswrapper[4612]: E0227 08:10:49.270630 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a40110d-5675-40b6-87c6-63b1d2faba9d" containerName="cinder-api" Feb 27 08:10:49 crc kubenswrapper[4612]: I0227 08:10:49.270639 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a40110d-5675-40b6-87c6-63b1d2faba9d" containerName="cinder-api" Feb 27 08:10:49 crc kubenswrapper[4612]: I0227 08:10:49.270877 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="7c51bfaf-2c6a-49aa-afe7-eb4789351857" containerName="kube-state-metrics" Feb 27 08:10:49 crc kubenswrapper[4612]: I0227 08:10:49.270890 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="9a40110d-5675-40b6-87c6-63b1d2faba9d" containerName="cinder-api" Feb 27 08:10:49 crc kubenswrapper[4612]: I0227 08:10:49.270904 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="9a40110d-5675-40b6-87c6-63b1d2faba9d" containerName="cinder-api-log" Feb 27 08:10:49 crc kubenswrapper[4612]: I0227 08:10:49.271640 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Feb 27 08:10:49 crc kubenswrapper[4612]: I0227 08:10:49.279129 4612 scope.go:117] "RemoveContainer" containerID="6a08fb66db51ba5788882f4d7cb293f89bccd46dba71bf71878951639068185f" Feb 27 08:10:49 crc kubenswrapper[4612]: I0227 08:10:49.279482 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Feb 27 08:10:49 crc kubenswrapper[4612]: I0227 08:10:49.279670 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Feb 27 08:10:49 crc kubenswrapper[4612]: E0227 08:10:49.283835 4612 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6a08fb66db51ba5788882f4d7cb293f89bccd46dba71bf71878951639068185f\": container with ID starting with 6a08fb66db51ba5788882f4d7cb293f89bccd46dba71bf71878951639068185f not found: ID does not exist" containerID="6a08fb66db51ba5788882f4d7cb293f89bccd46dba71bf71878951639068185f" Feb 27 08:10:49 crc kubenswrapper[4612]: I0227 08:10:49.283878 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6a08fb66db51ba5788882f4d7cb293f89bccd46dba71bf71878951639068185f"} err="failed to get container status \"6a08fb66db51ba5788882f4d7cb293f89bccd46dba71bf71878951639068185f\": rpc error: code = NotFound desc = could not find container \"6a08fb66db51ba5788882f4d7cb293f89bccd46dba71bf71878951639068185f\": container with ID starting with 6a08fb66db51ba5788882f4d7cb293f89bccd46dba71bf71878951639068185f not found: ID does not exist" Feb 27 08:10:49 crc kubenswrapper[4612]: I0227 08:10:49.283900 4612 scope.go:117] "RemoveContainer" containerID="c198f8b781fb884ae7ba10334f69026ead837cda8cfc728772b0e7b942cb069e" Feb 27 08:10:49 crc kubenswrapper[4612]: E0227 08:10:49.284960 4612 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c198f8b781fb884ae7ba10334f69026ead837cda8cfc728772b0e7b942cb069e\": container with ID starting with c198f8b781fb884ae7ba10334f69026ead837cda8cfc728772b0e7b942cb069e not found: ID does not exist" containerID="c198f8b781fb884ae7ba10334f69026ead837cda8cfc728772b0e7b942cb069e" Feb 27 08:10:49 crc kubenswrapper[4612]: I0227 08:10:49.284984 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c198f8b781fb884ae7ba10334f69026ead837cda8cfc728772b0e7b942cb069e"} err="failed to get container status \"c198f8b781fb884ae7ba10334f69026ead837cda8cfc728772b0e7b942cb069e\": rpc error: code = NotFound desc = could not find container \"c198f8b781fb884ae7ba10334f69026ead837cda8cfc728772b0e7b942cb069e\": container with ID starting with c198f8b781fb884ae7ba10334f69026ead837cda8cfc728772b0e7b942cb069e not found: ID does not exist" Feb 27 08:10:49 crc kubenswrapper[4612]: I0227 08:10:49.287188 4612 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9a40110d-5675-40b6-87c6-63b1d2faba9d-logs\") on node \"crc\" DevicePath \"\"" Feb 27 08:10:49 crc kubenswrapper[4612]: I0227 08:10:49.287207 4612 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/9a40110d-5675-40b6-87c6-63b1d2faba9d-etc-machine-id\") on node \"crc\" DevicePath \"\"" Feb 27 08:10:49 crc kubenswrapper[4612]: I0227 08:10:49.287216 4612 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9a40110d-5675-40b6-87c6-63b1d2faba9d-scripts\") on node \"crc\" DevicePath \"\"" Feb 27 08:10:49 crc kubenswrapper[4612]: I0227 08:10:49.287226 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nkgbh\" (UniqueName: \"kubernetes.io/projected/9a40110d-5675-40b6-87c6-63b1d2faba9d-kube-api-access-nkgbh\") on node \"crc\" DevicePath \"\"" Feb 27 08:10:49 crc kubenswrapper[4612]: I0227 08:10:49.287234 4612 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9a40110d-5675-40b6-87c6-63b1d2faba9d-config-data-custom\") on node \"crc\" DevicePath \"\"" Feb 27 08:10:49 crc kubenswrapper[4612]: I0227 08:10:49.290625 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Feb 27 08:10:49 crc kubenswrapper[4612]: I0227 08:10:49.304879 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9a40110d-5675-40b6-87c6-63b1d2faba9d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9a40110d-5675-40b6-87c6-63b1d2faba9d" (UID: "9a40110d-5675-40b6-87c6-63b1d2faba9d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:10:49 crc kubenswrapper[4612]: I0227 08:10:49.390036 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7a3af89c-deaa-4d32-b00b-ae3ca1de8ee9-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"7a3af89c-deaa-4d32-b00b-ae3ca1de8ee9\") " pod="openstack/kube-state-metrics-0" Feb 27 08:10:49 crc kubenswrapper[4612]: I0227 08:10:49.390138 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9777h\" (UniqueName: \"kubernetes.io/projected/7a3af89c-deaa-4d32-b00b-ae3ca1de8ee9-kube-api-access-9777h\") pod \"kube-state-metrics-0\" (UID: \"7a3af89c-deaa-4d32-b00b-ae3ca1de8ee9\") " pod="openstack/kube-state-metrics-0" Feb 27 08:10:49 crc kubenswrapper[4612]: I0227 08:10:49.390181 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/7a3af89c-deaa-4d32-b00b-ae3ca1de8ee9-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"7a3af89c-deaa-4d32-b00b-ae3ca1de8ee9\") " pod="openstack/kube-state-metrics-0" Feb 27 08:10:49 crc kubenswrapper[4612]: I0227 08:10:49.390322 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/7a3af89c-deaa-4d32-b00b-ae3ca1de8ee9-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"7a3af89c-deaa-4d32-b00b-ae3ca1de8ee9\") " pod="openstack/kube-state-metrics-0" Feb 27 08:10:49 crc kubenswrapper[4612]: I0227 08:10:49.390451 4612 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9a40110d-5675-40b6-87c6-63b1d2faba9d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 08:10:49 crc kubenswrapper[4612]: I0227 08:10:49.400262 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9a40110d-5675-40b6-87c6-63b1d2faba9d-config-data" (OuterVolumeSpecName: "config-data") pod "9a40110d-5675-40b6-87c6-63b1d2faba9d" (UID: "9a40110d-5675-40b6-87c6-63b1d2faba9d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:10:49 crc kubenswrapper[4612]: I0227 08:10:49.492177 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7a3af89c-deaa-4d32-b00b-ae3ca1de8ee9-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"7a3af89c-deaa-4d32-b00b-ae3ca1de8ee9\") " pod="openstack/kube-state-metrics-0" Feb 27 08:10:49 crc kubenswrapper[4612]: I0227 08:10:49.492250 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9777h\" (UniqueName: \"kubernetes.io/projected/7a3af89c-deaa-4d32-b00b-ae3ca1de8ee9-kube-api-access-9777h\") pod \"kube-state-metrics-0\" (UID: \"7a3af89c-deaa-4d32-b00b-ae3ca1de8ee9\") " pod="openstack/kube-state-metrics-0" Feb 27 08:10:49 crc kubenswrapper[4612]: I0227 08:10:49.492273 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/7a3af89c-deaa-4d32-b00b-ae3ca1de8ee9-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"7a3af89c-deaa-4d32-b00b-ae3ca1de8ee9\") " pod="openstack/kube-state-metrics-0" Feb 27 08:10:49 crc kubenswrapper[4612]: I0227 08:10:49.492352 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/7a3af89c-deaa-4d32-b00b-ae3ca1de8ee9-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"7a3af89c-deaa-4d32-b00b-ae3ca1de8ee9\") " pod="openstack/kube-state-metrics-0" Feb 27 08:10:49 crc kubenswrapper[4612]: I0227 08:10:49.492412 4612 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9a40110d-5675-40b6-87c6-63b1d2faba9d-config-data\") on node \"crc\" DevicePath \"\"" Feb 27 08:10:49 crc kubenswrapper[4612]: I0227 08:10:49.497340 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Feb 27 08:10:49 crc kubenswrapper[4612]: I0227 08:10:49.497909 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/7a3af89c-deaa-4d32-b00b-ae3ca1de8ee9-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"7a3af89c-deaa-4d32-b00b-ae3ca1de8ee9\") " pod="openstack/kube-state-metrics-0" Feb 27 08:10:49 crc kubenswrapper[4612]: I0227 08:10:49.507904 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Feb 27 08:10:49 crc kubenswrapper[4612]: I0227 08:10:49.509786 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7a3af89c-deaa-4d32-b00b-ae3ca1de8ee9-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"7a3af89c-deaa-4d32-b00b-ae3ca1de8ee9\") " pod="openstack/kube-state-metrics-0" Feb 27 08:10:49 crc kubenswrapper[4612]: I0227 08:10:49.532450 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/7a3af89c-deaa-4d32-b00b-ae3ca1de8ee9-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"7a3af89c-deaa-4d32-b00b-ae3ca1de8ee9\") " pod="openstack/kube-state-metrics-0" Feb 27 08:10:49 crc kubenswrapper[4612]: I0227 08:10:49.534521 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9777h\" (UniqueName: \"kubernetes.io/projected/7a3af89c-deaa-4d32-b00b-ae3ca1de8ee9-kube-api-access-9777h\") pod \"kube-state-metrics-0\" (UID: \"7a3af89c-deaa-4d32-b00b-ae3ca1de8ee9\") " pod="openstack/kube-state-metrics-0" Feb 27 08:10:49 crc kubenswrapper[4612]: I0227 08:10:49.542736 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Feb 27 08:10:49 crc kubenswrapper[4612]: I0227 08:10:49.544193 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Feb 27 08:10:49 crc kubenswrapper[4612]: I0227 08:10:49.552987 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Feb 27 08:10:49 crc kubenswrapper[4612]: I0227 08:10:49.555972 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Feb 27 08:10:49 crc kubenswrapper[4612]: I0227 08:10:49.554921 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Feb 27 08:10:49 crc kubenswrapper[4612]: I0227 08:10:49.555144 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Feb 27 08:10:49 crc kubenswrapper[4612]: I0227 08:10:49.594944 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9ba31a95-0fce-4eb6-98be-6292b7146e21-scripts\") pod \"cinder-api-0\" (UID: \"9ba31a95-0fce-4eb6-98be-6292b7146e21\") " pod="openstack/cinder-api-0" Feb 27 08:10:49 crc kubenswrapper[4612]: I0227 08:10:49.595003 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9ba31a95-0fce-4eb6-98be-6292b7146e21-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"9ba31a95-0fce-4eb6-98be-6292b7146e21\") " pod="openstack/cinder-api-0" Feb 27 08:10:49 crc kubenswrapper[4612]: I0227 08:10:49.595067 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9ba31a95-0fce-4eb6-98be-6292b7146e21-config-data-custom\") pod \"cinder-api-0\" (UID: \"9ba31a95-0fce-4eb6-98be-6292b7146e21\") " pod="openstack/cinder-api-0" Feb 27 08:10:49 crc kubenswrapper[4612]: I0227 08:10:49.595086 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/9ba31a95-0fce-4eb6-98be-6292b7146e21-etc-machine-id\") pod \"cinder-api-0\" (UID: \"9ba31a95-0fce-4eb6-98be-6292b7146e21\") " pod="openstack/cinder-api-0" Feb 27 08:10:49 crc kubenswrapper[4612]: I0227 08:10:49.595167 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9ba31a95-0fce-4eb6-98be-6292b7146e21-logs\") pod \"cinder-api-0\" (UID: \"9ba31a95-0fce-4eb6-98be-6292b7146e21\") " pod="openstack/cinder-api-0" Feb 27 08:10:49 crc kubenswrapper[4612]: I0227 08:10:49.595207 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-79rsv\" (UniqueName: \"kubernetes.io/projected/9ba31a95-0fce-4eb6-98be-6292b7146e21-kube-api-access-79rsv\") pod \"cinder-api-0\" (UID: \"9ba31a95-0fce-4eb6-98be-6292b7146e21\") " pod="openstack/cinder-api-0" Feb 27 08:10:49 crc kubenswrapper[4612]: I0227 08:10:49.595227 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9ba31a95-0fce-4eb6-98be-6292b7146e21-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"9ba31a95-0fce-4eb6-98be-6292b7146e21\") " pod="openstack/cinder-api-0" Feb 27 08:10:49 crc kubenswrapper[4612]: I0227 08:10:49.595253 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9ba31a95-0fce-4eb6-98be-6292b7146e21-config-data\") pod \"cinder-api-0\" (UID: \"9ba31a95-0fce-4eb6-98be-6292b7146e21\") " pod="openstack/cinder-api-0" Feb 27 08:10:49 crc kubenswrapper[4612]: I0227 08:10:49.595294 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9ba31a95-0fce-4eb6-98be-6292b7146e21-public-tls-certs\") pod \"cinder-api-0\" (UID: \"9ba31a95-0fce-4eb6-98be-6292b7146e21\") " pod="openstack/cinder-api-0" Feb 27 08:10:49 crc kubenswrapper[4612]: I0227 08:10:49.631486 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Feb 27 08:10:49 crc kubenswrapper[4612]: I0227 08:10:49.699910 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9ba31a95-0fce-4eb6-98be-6292b7146e21-logs\") pod \"cinder-api-0\" (UID: \"9ba31a95-0fce-4eb6-98be-6292b7146e21\") " pod="openstack/cinder-api-0" Feb 27 08:10:49 crc kubenswrapper[4612]: I0227 08:10:49.700065 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-79rsv\" (UniqueName: \"kubernetes.io/projected/9ba31a95-0fce-4eb6-98be-6292b7146e21-kube-api-access-79rsv\") pod \"cinder-api-0\" (UID: \"9ba31a95-0fce-4eb6-98be-6292b7146e21\") " pod="openstack/cinder-api-0" Feb 27 08:10:49 crc kubenswrapper[4612]: I0227 08:10:49.700140 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9ba31a95-0fce-4eb6-98be-6292b7146e21-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"9ba31a95-0fce-4eb6-98be-6292b7146e21\") " pod="openstack/cinder-api-0" Feb 27 08:10:49 crc kubenswrapper[4612]: I0227 08:10:49.700215 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9ba31a95-0fce-4eb6-98be-6292b7146e21-config-data\") pod \"cinder-api-0\" (UID: \"9ba31a95-0fce-4eb6-98be-6292b7146e21\") " pod="openstack/cinder-api-0" Feb 27 08:10:49 crc kubenswrapper[4612]: I0227 08:10:49.700482 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9ba31a95-0fce-4eb6-98be-6292b7146e21-public-tls-certs\") pod \"cinder-api-0\" (UID: \"9ba31a95-0fce-4eb6-98be-6292b7146e21\") " pod="openstack/cinder-api-0" Feb 27 08:10:49 crc kubenswrapper[4612]: I0227 08:10:49.701172 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9ba31a95-0fce-4eb6-98be-6292b7146e21-logs\") pod \"cinder-api-0\" (UID: \"9ba31a95-0fce-4eb6-98be-6292b7146e21\") " pod="openstack/cinder-api-0" Feb 27 08:10:49 crc kubenswrapper[4612]: I0227 08:10:49.700523 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9ba31a95-0fce-4eb6-98be-6292b7146e21-scripts\") pod \"cinder-api-0\" (UID: \"9ba31a95-0fce-4eb6-98be-6292b7146e21\") " pod="openstack/cinder-api-0" Feb 27 08:10:49 crc kubenswrapper[4612]: I0227 08:10:49.703000 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9ba31a95-0fce-4eb6-98be-6292b7146e21-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"9ba31a95-0fce-4eb6-98be-6292b7146e21\") " pod="openstack/cinder-api-0" Feb 27 08:10:49 crc kubenswrapper[4612]: I0227 08:10:49.703420 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9ba31a95-0fce-4eb6-98be-6292b7146e21-config-data-custom\") pod \"cinder-api-0\" (UID: \"9ba31a95-0fce-4eb6-98be-6292b7146e21\") " pod="openstack/cinder-api-0" Feb 27 08:10:49 crc kubenswrapper[4612]: I0227 08:10:49.704020 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/9ba31a95-0fce-4eb6-98be-6292b7146e21-etc-machine-id\") pod \"cinder-api-0\" (UID: \"9ba31a95-0fce-4eb6-98be-6292b7146e21\") " pod="openstack/cinder-api-0" Feb 27 08:10:49 crc kubenswrapper[4612]: I0227 08:10:49.705110 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/9ba31a95-0fce-4eb6-98be-6292b7146e21-etc-machine-id\") pod \"cinder-api-0\" (UID: \"9ba31a95-0fce-4eb6-98be-6292b7146e21\") " pod="openstack/cinder-api-0" Feb 27 08:10:49 crc kubenswrapper[4612]: I0227 08:10:49.716583 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9ba31a95-0fce-4eb6-98be-6292b7146e21-public-tls-certs\") pod \"cinder-api-0\" (UID: \"9ba31a95-0fce-4eb6-98be-6292b7146e21\") " pod="openstack/cinder-api-0" Feb 27 08:10:49 crc kubenswrapper[4612]: I0227 08:10:49.717230 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9ba31a95-0fce-4eb6-98be-6292b7146e21-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"9ba31a95-0fce-4eb6-98be-6292b7146e21\") " pod="openstack/cinder-api-0" Feb 27 08:10:49 crc kubenswrapper[4612]: I0227 08:10:49.719008 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9ba31a95-0fce-4eb6-98be-6292b7146e21-config-data\") pod \"cinder-api-0\" (UID: \"9ba31a95-0fce-4eb6-98be-6292b7146e21\") " pod="openstack/cinder-api-0" Feb 27 08:10:49 crc kubenswrapper[4612]: I0227 08:10:49.719483 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9ba31a95-0fce-4eb6-98be-6292b7146e21-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"9ba31a95-0fce-4eb6-98be-6292b7146e21\") " pod="openstack/cinder-api-0" Feb 27 08:10:49 crc kubenswrapper[4612]: I0227 08:10:49.722491 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9ba31a95-0fce-4eb6-98be-6292b7146e21-config-data-custom\") pod \"cinder-api-0\" (UID: \"9ba31a95-0fce-4eb6-98be-6292b7146e21\") " pod="openstack/cinder-api-0" Feb 27 08:10:49 crc kubenswrapper[4612]: I0227 08:10:49.738138 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9ba31a95-0fce-4eb6-98be-6292b7146e21-scripts\") pod \"cinder-api-0\" (UID: \"9ba31a95-0fce-4eb6-98be-6292b7146e21\") " pod="openstack/cinder-api-0" Feb 27 08:10:49 crc kubenswrapper[4612]: I0227 08:10:49.738751 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-79rsv\" (UniqueName: \"kubernetes.io/projected/9ba31a95-0fce-4eb6-98be-6292b7146e21-kube-api-access-79rsv\") pod \"cinder-api-0\" (UID: \"9ba31a95-0fce-4eb6-98be-6292b7146e21\") " pod="openstack/cinder-api-0" Feb 27 08:10:49 crc kubenswrapper[4612]: I0227 08:10:49.876483 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Feb 27 08:10:50 crc kubenswrapper[4612]: W0227 08:10:50.144185 4612 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7a3af89c_deaa_4d32_b00b_ae3ca1de8ee9.slice/crio-0d0b57f4603f542a8ab85fa0b73fee4f1d63b12ce39a0f29c59ab98a9d6fca42 WatchSource:0}: Error finding container 0d0b57f4603f542a8ab85fa0b73fee4f1d63b12ce39a0f29c59ab98a9d6fca42: Status 404 returned error can't find the container with id 0d0b57f4603f542a8ab85fa0b73fee4f1d63b12ce39a0f29c59ab98a9d6fca42 Feb 27 08:10:50 crc kubenswrapper[4612]: I0227 08:10:50.147715 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Feb 27 08:10:50 crc kubenswrapper[4612]: I0227 08:10:50.171215 4612 generic.go:334] "Generic (PLEG): container finished" podID="76009243-16b3-4652-abfe-a6cdba363724" containerID="a0baf671a66706916e8b02ab97ccdead951de20e3f62e64d913a8590aa359cb3" exitCode=137 Feb 27 08:10:50 crc kubenswrapper[4612]: I0227 08:10:50.171301 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6b985547d4-mkfpv" event={"ID":"76009243-16b3-4652-abfe-a6cdba363724","Type":"ContainerDied","Data":"a0baf671a66706916e8b02ab97ccdead951de20e3f62e64d913a8590aa359cb3"} Feb 27 08:10:50 crc kubenswrapper[4612]: I0227 08:10:50.171330 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6b985547d4-mkfpv" event={"ID":"76009243-16b3-4652-abfe-a6cdba363724","Type":"ContainerStarted","Data":"36510ad0ab2650616b7e36d7449187c82f7e47b4ef8d4d802e348d602213aa5a"} Feb 27 08:10:50 crc kubenswrapper[4612]: I0227 08:10:50.172910 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"7a3af89c-deaa-4d32-b00b-ae3ca1de8ee9","Type":"ContainerStarted","Data":"0d0b57f4603f542a8ab85fa0b73fee4f1d63b12ce39a0f29c59ab98a9d6fca42"} Feb 27 08:10:50 crc kubenswrapper[4612]: I0227 08:10:50.188150 4612 generic.go:334] "Generic (PLEG): container finished" podID="b72b541d-4661-44a7-a121-c2a8aef6db11" containerID="92102897b17cb0c0c81984cbb9f2014fd0591e9bae23a94bbb5532b66e484fd8" exitCode=137 Feb 27 08:10:50 crc kubenswrapper[4612]: I0227 08:10:50.188185 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-77948db5bb-97qh6" event={"ID":"b72b541d-4661-44a7-a121-c2a8aef6db11","Type":"ContainerDied","Data":"92102897b17cb0c0c81984cbb9f2014fd0591e9bae23a94bbb5532b66e484fd8"} Feb 27 08:10:50 crc kubenswrapper[4612]: I0227 08:10:50.188621 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-77948db5bb-97qh6" event={"ID":"b72b541d-4661-44a7-a121-c2a8aef6db11","Type":"ContainerStarted","Data":"30c91de62cca0b2977a82998a3b3ae2c064180b42fa458f4c7412775535106de"} Feb 27 08:10:50 crc kubenswrapper[4612]: I0227 08:10:50.781997 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-pqvhc"] Feb 27 08:10:50 crc kubenswrapper[4612]: I0227 08:10:50.783359 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-pqvhc" Feb 27 08:10:50 crc kubenswrapper[4612]: I0227 08:10:50.786514 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-pqvhc"] Feb 27 08:10:50 crc kubenswrapper[4612]: I0227 08:10:50.907828 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7c51bfaf-2c6a-49aa-afe7-eb4789351857" path="/var/lib/kubelet/pods/7c51bfaf-2c6a-49aa-afe7-eb4789351857/volumes" Feb 27 08:10:50 crc kubenswrapper[4612]: I0227 08:10:50.908574 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9a40110d-5675-40b6-87c6-63b1d2faba9d" path="/var/lib/kubelet/pods/9a40110d-5675-40b6-87c6-63b1d2faba9d/volumes" Feb 27 08:10:50 crc kubenswrapper[4612]: I0227 08:10:50.922053 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-c7w6b"] Feb 27 08:10:50 crc kubenswrapper[4612]: I0227 08:10:50.930223 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-c7w6b" Feb 27 08:10:50 crc kubenswrapper[4612]: I0227 08:10:50.930647 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5c30c00d-ed94-4980-a218-a2e3b95bd8eb-operator-scripts\") pod \"nova-api-db-create-pqvhc\" (UID: \"5c30c00d-ed94-4980-a218-a2e3b95bd8eb\") " pod="openstack/nova-api-db-create-pqvhc" Feb 27 08:10:50 crc kubenswrapper[4612]: I0227 08:10:50.930762 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mgqhr\" (UniqueName: \"kubernetes.io/projected/5c30c00d-ed94-4980-a218-a2e3b95bd8eb-kube-api-access-mgqhr\") pod \"nova-api-db-create-pqvhc\" (UID: \"5c30c00d-ed94-4980-a218-a2e3b95bd8eb\") " pod="openstack/nova-api-db-create-pqvhc" Feb 27 08:10:50 crc kubenswrapper[4612]: I0227 08:10:50.940383 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-c7w6b"] Feb 27 08:10:51 crc kubenswrapper[4612]: I0227 08:10:51.032672 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5c30c00d-ed94-4980-a218-a2e3b95bd8eb-operator-scripts\") pod \"nova-api-db-create-pqvhc\" (UID: \"5c30c00d-ed94-4980-a218-a2e3b95bd8eb\") " pod="openstack/nova-api-db-create-pqvhc" Feb 27 08:10:51 crc kubenswrapper[4612]: I0227 08:10:51.032745 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/06a51ad9-c1fb-452b-9c2b-62e4d2c80266-operator-scripts\") pod \"nova-cell0-db-create-c7w6b\" (UID: \"06a51ad9-c1fb-452b-9c2b-62e4d2c80266\") " pod="openstack/nova-cell0-db-create-c7w6b" Feb 27 08:10:51 crc kubenswrapper[4612]: I0227 08:10:51.032798 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mgqhr\" (UniqueName: \"kubernetes.io/projected/5c30c00d-ed94-4980-a218-a2e3b95bd8eb-kube-api-access-mgqhr\") pod \"nova-api-db-create-pqvhc\" (UID: \"5c30c00d-ed94-4980-a218-a2e3b95bd8eb\") " pod="openstack/nova-api-db-create-pqvhc" Feb 27 08:10:51 crc kubenswrapper[4612]: I0227 08:10:51.032851 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ls4rt\" (UniqueName: \"kubernetes.io/projected/06a51ad9-c1fb-452b-9c2b-62e4d2c80266-kube-api-access-ls4rt\") pod \"nova-cell0-db-create-c7w6b\" (UID: \"06a51ad9-c1fb-452b-9c2b-62e4d2c80266\") " pod="openstack/nova-cell0-db-create-c7w6b" Feb 27 08:10:51 crc kubenswrapper[4612]: I0227 08:10:51.040272 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5c30c00d-ed94-4980-a218-a2e3b95bd8eb-operator-scripts\") pod \"nova-api-db-create-pqvhc\" (UID: \"5c30c00d-ed94-4980-a218-a2e3b95bd8eb\") " pod="openstack/nova-api-db-create-pqvhc" Feb 27 08:10:51 crc kubenswrapper[4612]: I0227 08:10:51.067224 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mgqhr\" (UniqueName: \"kubernetes.io/projected/5c30c00d-ed94-4980-a218-a2e3b95bd8eb-kube-api-access-mgqhr\") pod \"nova-api-db-create-pqvhc\" (UID: \"5c30c00d-ed94-4980-a218-a2e3b95bd8eb\") " pod="openstack/nova-api-db-create-pqvhc" Feb 27 08:10:51 crc kubenswrapper[4612]: I0227 08:10:51.126873 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-pqvhc" Feb 27 08:10:51 crc kubenswrapper[4612]: I0227 08:10:51.128646 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-tfl6p"] Feb 27 08:10:51 crc kubenswrapper[4612]: I0227 08:10:51.131672 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-tfl6p" Feb 27 08:10:51 crc kubenswrapper[4612]: I0227 08:10:51.138177 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a5162728-e496-4f30-9efb-a411e8e52ba4-operator-scripts\") pod \"nova-cell1-db-create-tfl6p\" (UID: \"a5162728-e496-4f30-9efb-a411e8e52ba4\") " pod="openstack/nova-cell1-db-create-tfl6p" Feb 27 08:10:51 crc kubenswrapper[4612]: I0227 08:10:51.138267 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/06a51ad9-c1fb-452b-9c2b-62e4d2c80266-operator-scripts\") pod \"nova-cell0-db-create-c7w6b\" (UID: \"06a51ad9-c1fb-452b-9c2b-62e4d2c80266\") " pod="openstack/nova-cell0-db-create-c7w6b" Feb 27 08:10:51 crc kubenswrapper[4612]: I0227 08:10:51.138291 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m546h\" (UniqueName: \"kubernetes.io/projected/a5162728-e496-4f30-9efb-a411e8e52ba4-kube-api-access-m546h\") pod \"nova-cell1-db-create-tfl6p\" (UID: \"a5162728-e496-4f30-9efb-a411e8e52ba4\") " pod="openstack/nova-cell1-db-create-tfl6p" Feb 27 08:10:51 crc kubenswrapper[4612]: I0227 08:10:51.138376 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ls4rt\" (UniqueName: \"kubernetes.io/projected/06a51ad9-c1fb-452b-9c2b-62e4d2c80266-kube-api-access-ls4rt\") pod \"nova-cell0-db-create-c7w6b\" (UID: \"06a51ad9-c1fb-452b-9c2b-62e4d2c80266\") " pod="openstack/nova-cell0-db-create-c7w6b" Feb 27 08:10:51 crc kubenswrapper[4612]: I0227 08:10:51.154313 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-tfl6p"] Feb 27 08:10:51 crc kubenswrapper[4612]: I0227 08:10:51.155355 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/06a51ad9-c1fb-452b-9c2b-62e4d2c80266-operator-scripts\") pod \"nova-cell0-db-create-c7w6b\" (UID: \"06a51ad9-c1fb-452b-9c2b-62e4d2c80266\") " pod="openstack/nova-cell0-db-create-c7w6b" Feb 27 08:10:51 crc kubenswrapper[4612]: I0227 08:10:51.169894 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-2ac7-account-create-update-hnvdn"] Feb 27 08:10:51 crc kubenswrapper[4612]: I0227 08:10:51.179843 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-2ac7-account-create-update-hnvdn" Feb 27 08:10:51 crc kubenswrapper[4612]: I0227 08:10:51.182451 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Feb 27 08:10:51 crc kubenswrapper[4612]: I0227 08:10:51.197863 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ls4rt\" (UniqueName: \"kubernetes.io/projected/06a51ad9-c1fb-452b-9c2b-62e4d2c80266-kube-api-access-ls4rt\") pod \"nova-cell0-db-create-c7w6b\" (UID: \"06a51ad9-c1fb-452b-9c2b-62e4d2c80266\") " pod="openstack/nova-cell0-db-create-c7w6b" Feb 27 08:10:51 crc kubenswrapper[4612]: I0227 08:10:51.214639 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-2ac7-account-create-update-hnvdn"] Feb 27 08:10:51 crc kubenswrapper[4612]: I0227 08:10:51.240678 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d1b7e951-57b4-47f3-b337-b3468e7c6637-operator-scripts\") pod \"nova-api-2ac7-account-create-update-hnvdn\" (UID: \"d1b7e951-57b4-47f3-b337-b3468e7c6637\") " pod="openstack/nova-api-2ac7-account-create-update-hnvdn" Feb 27 08:10:51 crc kubenswrapper[4612]: I0227 08:10:51.240770 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a5162728-e496-4f30-9efb-a411e8e52ba4-operator-scripts\") pod \"nova-cell1-db-create-tfl6p\" (UID: \"a5162728-e496-4f30-9efb-a411e8e52ba4\") " pod="openstack/nova-cell1-db-create-tfl6p" Feb 27 08:10:51 crc kubenswrapper[4612]: I0227 08:10:51.240814 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kzdkv\" (UniqueName: \"kubernetes.io/projected/d1b7e951-57b4-47f3-b337-b3468e7c6637-kube-api-access-kzdkv\") pod \"nova-api-2ac7-account-create-update-hnvdn\" (UID: \"d1b7e951-57b4-47f3-b337-b3468e7c6637\") " pod="openstack/nova-api-2ac7-account-create-update-hnvdn" Feb 27 08:10:51 crc kubenswrapper[4612]: I0227 08:10:51.241738 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a5162728-e496-4f30-9efb-a411e8e52ba4-operator-scripts\") pod \"nova-cell1-db-create-tfl6p\" (UID: \"a5162728-e496-4f30-9efb-a411e8e52ba4\") " pod="openstack/nova-cell1-db-create-tfl6p" Feb 27 08:10:51 crc kubenswrapper[4612]: I0227 08:10:51.240931 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m546h\" (UniqueName: \"kubernetes.io/projected/a5162728-e496-4f30-9efb-a411e8e52ba4-kube-api-access-m546h\") pod \"nova-cell1-db-create-tfl6p\" (UID: \"a5162728-e496-4f30-9efb-a411e8e52ba4\") " pod="openstack/nova-cell1-db-create-tfl6p" Feb 27 08:10:51 crc kubenswrapper[4612]: I0227 08:10:51.271822 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-c7w6b" Feb 27 08:10:51 crc kubenswrapper[4612]: I0227 08:10:51.273209 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m546h\" (UniqueName: \"kubernetes.io/projected/a5162728-e496-4f30-9efb-a411e8e52ba4-kube-api-access-m546h\") pod \"nova-cell1-db-create-tfl6p\" (UID: \"a5162728-e496-4f30-9efb-a411e8e52ba4\") " pod="openstack/nova-cell1-db-create-tfl6p" Feb 27 08:10:51 crc kubenswrapper[4612]: I0227 08:10:51.273268 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9a05b2bc-7c21-412d-b7ba-26e77517f60d","Type":"ContainerStarted","Data":"371987e64261489b75417093a2cfe845de762e34592593234573c534290c186b"} Feb 27 08:10:51 crc kubenswrapper[4612]: I0227 08:10:51.292466 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Feb 27 08:10:51 crc kubenswrapper[4612]: I0227 08:10:51.352446 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kzdkv\" (UniqueName: \"kubernetes.io/projected/d1b7e951-57b4-47f3-b337-b3468e7c6637-kube-api-access-kzdkv\") pod \"nova-api-2ac7-account-create-update-hnvdn\" (UID: \"d1b7e951-57b4-47f3-b337-b3468e7c6637\") " pod="openstack/nova-api-2ac7-account-create-update-hnvdn" Feb 27 08:10:51 crc kubenswrapper[4612]: I0227 08:10:51.364138 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d1b7e951-57b4-47f3-b337-b3468e7c6637-operator-scripts\") pod \"nova-api-2ac7-account-create-update-hnvdn\" (UID: \"d1b7e951-57b4-47f3-b337-b3468e7c6637\") " pod="openstack/nova-api-2ac7-account-create-update-hnvdn" Feb 27 08:10:51 crc kubenswrapper[4612]: I0227 08:10:51.366793 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d1b7e951-57b4-47f3-b337-b3468e7c6637-operator-scripts\") pod \"nova-api-2ac7-account-create-update-hnvdn\" (UID: \"d1b7e951-57b4-47f3-b337-b3468e7c6637\") " pod="openstack/nova-api-2ac7-account-create-update-hnvdn" Feb 27 08:10:51 crc kubenswrapper[4612]: I0227 08:10:51.386330 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kzdkv\" (UniqueName: \"kubernetes.io/projected/d1b7e951-57b4-47f3-b337-b3468e7c6637-kube-api-access-kzdkv\") pod \"nova-api-2ac7-account-create-update-hnvdn\" (UID: \"d1b7e951-57b4-47f3-b337-b3468e7c6637\") " pod="openstack/nova-api-2ac7-account-create-update-hnvdn" Feb 27 08:10:51 crc kubenswrapper[4612]: I0227 08:10:51.481485 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-5f7d-account-create-update-fptbd"] Feb 27 08:10:51 crc kubenswrapper[4612]: I0227 08:10:51.482634 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-5f7d-account-create-update-fptbd" Feb 27 08:10:51 crc kubenswrapper[4612]: I0227 08:10:51.486213 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Feb 27 08:10:51 crc kubenswrapper[4612]: I0227 08:10:51.488388 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-5f7d-account-create-update-fptbd"] Feb 27 08:10:51 crc kubenswrapper[4612]: I0227 08:10:51.504081 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-tfl6p" Feb 27 08:10:51 crc kubenswrapper[4612]: I0227 08:10:51.505851 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-2ac7-account-create-update-hnvdn" Feb 27 08:10:51 crc kubenswrapper[4612]: I0227 08:10:51.618727 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-36be-account-create-update-7grtc"] Feb 27 08:10:51 crc kubenswrapper[4612]: I0227 08:10:51.618879 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ed43d122-6dcc-4bce-9ed1-184dcb96647c-operator-scripts\") pod \"nova-cell0-5f7d-account-create-update-fptbd\" (UID: \"ed43d122-6dcc-4bce-9ed1-184dcb96647c\") " pod="openstack/nova-cell0-5f7d-account-create-update-fptbd" Feb 27 08:10:51 crc kubenswrapper[4612]: I0227 08:10:51.621052 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8krrv\" (UniqueName: \"kubernetes.io/projected/ed43d122-6dcc-4bce-9ed1-184dcb96647c-kube-api-access-8krrv\") pod \"nova-cell0-5f7d-account-create-update-fptbd\" (UID: \"ed43d122-6dcc-4bce-9ed1-184dcb96647c\") " pod="openstack/nova-cell0-5f7d-account-create-update-fptbd" Feb 27 08:10:51 crc kubenswrapper[4612]: I0227 08:10:51.621896 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-36be-account-create-update-7grtc" Feb 27 08:10:51 crc kubenswrapper[4612]: I0227 08:10:51.630943 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Feb 27 08:10:51 crc kubenswrapper[4612]: I0227 08:10:51.648567 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-36be-account-create-update-7grtc"] Feb 27 08:10:51 crc kubenswrapper[4612]: I0227 08:10:51.724051 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e63e127a-2950-4a4c-a194-668b5142adb8-operator-scripts\") pod \"nova-cell1-36be-account-create-update-7grtc\" (UID: \"e63e127a-2950-4a4c-a194-668b5142adb8\") " pod="openstack/nova-cell1-36be-account-create-update-7grtc" Feb 27 08:10:51 crc kubenswrapper[4612]: I0227 08:10:51.724161 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x2vlv\" (UniqueName: \"kubernetes.io/projected/e63e127a-2950-4a4c-a194-668b5142adb8-kube-api-access-x2vlv\") pod \"nova-cell1-36be-account-create-update-7grtc\" (UID: \"e63e127a-2950-4a4c-a194-668b5142adb8\") " pod="openstack/nova-cell1-36be-account-create-update-7grtc" Feb 27 08:10:51 crc kubenswrapper[4612]: I0227 08:10:51.724193 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ed43d122-6dcc-4bce-9ed1-184dcb96647c-operator-scripts\") pod \"nova-cell0-5f7d-account-create-update-fptbd\" (UID: \"ed43d122-6dcc-4bce-9ed1-184dcb96647c\") " pod="openstack/nova-cell0-5f7d-account-create-update-fptbd" Feb 27 08:10:51 crc kubenswrapper[4612]: I0227 08:10:51.724229 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8krrv\" (UniqueName: \"kubernetes.io/projected/ed43d122-6dcc-4bce-9ed1-184dcb96647c-kube-api-access-8krrv\") pod \"nova-cell0-5f7d-account-create-update-fptbd\" (UID: \"ed43d122-6dcc-4bce-9ed1-184dcb96647c\") " pod="openstack/nova-cell0-5f7d-account-create-update-fptbd" Feb 27 08:10:51 crc kubenswrapper[4612]: I0227 08:10:51.725237 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ed43d122-6dcc-4bce-9ed1-184dcb96647c-operator-scripts\") pod \"nova-cell0-5f7d-account-create-update-fptbd\" (UID: \"ed43d122-6dcc-4bce-9ed1-184dcb96647c\") " pod="openstack/nova-cell0-5f7d-account-create-update-fptbd" Feb 27 08:10:51 crc kubenswrapper[4612]: I0227 08:10:51.749360 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8krrv\" (UniqueName: \"kubernetes.io/projected/ed43d122-6dcc-4bce-9ed1-184dcb96647c-kube-api-access-8krrv\") pod \"nova-cell0-5f7d-account-create-update-fptbd\" (UID: \"ed43d122-6dcc-4bce-9ed1-184dcb96647c\") " pod="openstack/nova-cell0-5f7d-account-create-update-fptbd" Feb 27 08:10:51 crc kubenswrapper[4612]: I0227 08:10:51.827863 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x2vlv\" (UniqueName: \"kubernetes.io/projected/e63e127a-2950-4a4c-a194-668b5142adb8-kube-api-access-x2vlv\") pod \"nova-cell1-36be-account-create-update-7grtc\" (UID: \"e63e127a-2950-4a4c-a194-668b5142adb8\") " pod="openstack/nova-cell1-36be-account-create-update-7grtc" Feb 27 08:10:51 crc kubenswrapper[4612]: I0227 08:10:51.828205 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e63e127a-2950-4a4c-a194-668b5142adb8-operator-scripts\") pod \"nova-cell1-36be-account-create-update-7grtc\" (UID: \"e63e127a-2950-4a4c-a194-668b5142adb8\") " pod="openstack/nova-cell1-36be-account-create-update-7grtc" Feb 27 08:10:51 crc kubenswrapper[4612]: I0227 08:10:51.829088 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e63e127a-2950-4a4c-a194-668b5142adb8-operator-scripts\") pod \"nova-cell1-36be-account-create-update-7grtc\" (UID: \"e63e127a-2950-4a4c-a194-668b5142adb8\") " pod="openstack/nova-cell1-36be-account-create-update-7grtc" Feb 27 08:10:51 crc kubenswrapper[4612]: I0227 08:10:51.844762 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x2vlv\" (UniqueName: \"kubernetes.io/projected/e63e127a-2950-4a4c-a194-668b5142adb8-kube-api-access-x2vlv\") pod \"nova-cell1-36be-account-create-update-7grtc\" (UID: \"e63e127a-2950-4a4c-a194-668b5142adb8\") " pod="openstack/nova-cell1-36be-account-create-update-7grtc" Feb 27 08:10:51 crc kubenswrapper[4612]: I0227 08:10:51.857674 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-pqvhc"] Feb 27 08:10:51 crc kubenswrapper[4612]: I0227 08:10:51.941322 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-5f7d-account-create-update-fptbd" Feb 27 08:10:52 crc kubenswrapper[4612]: I0227 08:10:52.115405 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-36be-account-create-update-7grtc" Feb 27 08:10:52 crc kubenswrapper[4612]: I0227 08:10:52.170871 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-c7w6b"] Feb 27 08:10:52 crc kubenswrapper[4612]: I0227 08:10:52.290137 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-pqvhc" event={"ID":"5c30c00d-ed94-4980-a218-a2e3b95bd8eb","Type":"ContainerStarted","Data":"72f46a1607606dd0aaa5a24ee872882b463e59b7d0f973888e5d048e09f7a541"} Feb 27 08:10:52 crc kubenswrapper[4612]: I0227 08:10:52.291495 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-c7w6b" event={"ID":"06a51ad9-c1fb-452b-9c2b-62e4d2c80266","Type":"ContainerStarted","Data":"a74a6ae364b9adc5f3d373f3c65f93353162f32546bccca3fe1f2c1206591088"} Feb 27 08:10:52 crc kubenswrapper[4612]: I0227 08:10:52.293906 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"9ba31a95-0fce-4eb6-98be-6292b7146e21","Type":"ContainerStarted","Data":"d12f633d523ce8782a8383f6dc271cc78bf910174623d850830aaad901aa7b67"} Feb 27 08:10:52 crc kubenswrapper[4612]: I0227 08:10:52.296164 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"7a3af89c-deaa-4d32-b00b-ae3ca1de8ee9","Type":"ContainerStarted","Data":"8719ec6ad79e4b73a2fdc2b2d69d4280ab3f808c75bf5ec631e299c4cc02a684"} Feb 27 08:10:52 crc kubenswrapper[4612]: I0227 08:10:52.296317 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Feb 27 08:10:52 crc kubenswrapper[4612]: I0227 08:10:52.329364 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=2.906723948 podStartE2EDuration="3.329346384s" podCreationTimestamp="2026-02-27 08:10:49 +0000 UTC" firstStartedPulling="2026-02-27 08:10:50.555885304 +0000 UTC m=+1308.409815302" lastFinishedPulling="2026-02-27 08:10:50.97850774 +0000 UTC m=+1308.832437738" observedRunningTime="2026-02-27 08:10:52.312592053 +0000 UTC m=+1310.166522051" watchObservedRunningTime="2026-02-27 08:10:52.329346384 +0000 UTC m=+1310.183276382" Feb 27 08:10:52 crc kubenswrapper[4612]: I0227 08:10:52.383589 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-tfl6p"] Feb 27 08:10:52 crc kubenswrapper[4612]: I0227 08:10:52.405812 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-2ac7-account-create-update-hnvdn"] Feb 27 08:10:52 crc kubenswrapper[4612]: I0227 08:10:52.660897 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-5f7d-account-create-update-fptbd"] Feb 27 08:10:52 crc kubenswrapper[4612]: I0227 08:10:52.875243 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-36be-account-create-update-7grtc"] Feb 27 08:10:53 crc kubenswrapper[4612]: I0227 08:10:53.311722 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"9ba31a95-0fce-4eb6-98be-6292b7146e21","Type":"ContainerStarted","Data":"eb76d87319de96f3f16bebe54a9e4fe96c55f34c5804498be87010f97b3576b9"} Feb 27 08:10:53 crc kubenswrapper[4612]: I0227 08:10:53.315215 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-5f7d-account-create-update-fptbd" event={"ID":"ed43d122-6dcc-4bce-9ed1-184dcb96647c","Type":"ContainerStarted","Data":"76720259afaeb97d1994408be42c233f8079c49ec8b77bf9ebb83772ca20fb35"} Feb 27 08:10:53 crc kubenswrapper[4612]: I0227 08:10:53.315257 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-5f7d-account-create-update-fptbd" event={"ID":"ed43d122-6dcc-4bce-9ed1-184dcb96647c","Type":"ContainerStarted","Data":"3f035dc833d2fed58f08924c97c7fe7b4ddf5a74892934d4fa7cd390dfe5763d"} Feb 27 08:10:53 crc kubenswrapper[4612]: I0227 08:10:53.337256 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-2ac7-account-create-update-hnvdn" event={"ID":"d1b7e951-57b4-47f3-b337-b3468e7c6637","Type":"ContainerStarted","Data":"2cc2b1f42a72c486e6a061a906a7e9f77b61101b86e5eb8b8cd7dddefa341eb9"} Feb 27 08:10:53 crc kubenswrapper[4612]: I0227 08:10:53.337489 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-2ac7-account-create-update-hnvdn" event={"ID":"d1b7e951-57b4-47f3-b337-b3468e7c6637","Type":"ContainerStarted","Data":"7ce27fd45040abdd7bde16a6fb28d92c377217b2d0f61511643b36ac566e1f03"} Feb 27 08:10:53 crc kubenswrapper[4612]: I0227 08:10:53.347612 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-tfl6p" event={"ID":"a5162728-e496-4f30-9efb-a411e8e52ba4","Type":"ContainerStarted","Data":"143969eaf93c6f453f01dc76dd51a7faea9680aaf6f7dde73247b621512f99e2"} Feb 27 08:10:53 crc kubenswrapper[4612]: I0227 08:10:53.347849 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-tfl6p" event={"ID":"a5162728-e496-4f30-9efb-a411e8e52ba4","Type":"ContainerStarted","Data":"4e9319ce04de4731fd5845c2868786a7476ed6adc59f015bc3c618818b366cd6"} Feb 27 08:10:53 crc kubenswrapper[4612]: I0227 08:10:53.349801 4612 generic.go:334] "Generic (PLEG): container finished" podID="5c30c00d-ed94-4980-a218-a2e3b95bd8eb" containerID="fa004161ac9acb310bf260abbf49e83dab05575fae08eb7b55a7941699bab387" exitCode=0 Feb 27 08:10:53 crc kubenswrapper[4612]: I0227 08:10:53.349862 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-pqvhc" event={"ID":"5c30c00d-ed94-4980-a218-a2e3b95bd8eb","Type":"ContainerDied","Data":"fa004161ac9acb310bf260abbf49e83dab05575fae08eb7b55a7941699bab387"} Feb 27 08:10:53 crc kubenswrapper[4612]: I0227 08:10:53.350220 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-5f7d-account-create-update-fptbd" podStartSLOduration=2.350208801 podStartE2EDuration="2.350208801s" podCreationTimestamp="2026-02-27 08:10:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 08:10:53.337008372 +0000 UTC m=+1311.190938380" watchObservedRunningTime="2026-02-27 08:10:53.350208801 +0000 UTC m=+1311.204138799" Feb 27 08:10:53 crc kubenswrapper[4612]: I0227 08:10:53.360351 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-2ac7-account-create-update-hnvdn" podStartSLOduration=2.360334342 podStartE2EDuration="2.360334342s" podCreationTimestamp="2026-02-27 08:10:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 08:10:53.360121216 +0000 UTC m=+1311.214051214" watchObservedRunningTime="2026-02-27 08:10:53.360334342 +0000 UTC m=+1311.214264330" Feb 27 08:10:53 crc kubenswrapper[4612]: I0227 08:10:53.386724 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-db-create-tfl6p" podStartSLOduration=2.386705609 podStartE2EDuration="2.386705609s" podCreationTimestamp="2026-02-27 08:10:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 08:10:53.375404955 +0000 UTC m=+1311.229334953" watchObservedRunningTime="2026-02-27 08:10:53.386705609 +0000 UTC m=+1311.240635607" Feb 27 08:10:53 crc kubenswrapper[4612]: I0227 08:10:53.410159 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9a05b2bc-7c21-412d-b7ba-26e77517f60d","Type":"ContainerStarted","Data":"7b9eafefccc9df5427ba8a3a5ccd2e2d5026ab9bbca43a0ff2af18a3f38ffd5b"} Feb 27 08:10:53 crc kubenswrapper[4612]: I0227 08:10:53.410567 4612 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="9a05b2bc-7c21-412d-b7ba-26e77517f60d" containerName="ceilometer-central-agent" containerID="cri-o://c9c7680af8a800caac979e94217d6e08c6a44613c883dc27baa42950bf08b3c4" gracePeriod=30 Feb 27 08:10:53 crc kubenswrapper[4612]: I0227 08:10:53.410734 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Feb 27 08:10:53 crc kubenswrapper[4612]: I0227 08:10:53.411112 4612 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="9a05b2bc-7c21-412d-b7ba-26e77517f60d" containerName="sg-core" containerID="cri-o://371987e64261489b75417093a2cfe845de762e34592593234573c534290c186b" gracePeriod=30 Feb 27 08:10:53 crc kubenswrapper[4612]: I0227 08:10:53.411136 4612 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="9a05b2bc-7c21-412d-b7ba-26e77517f60d" containerName="proxy-httpd" containerID="cri-o://7b9eafefccc9df5427ba8a3a5ccd2e2d5026ab9bbca43a0ff2af18a3f38ffd5b" gracePeriod=30 Feb 27 08:10:53 crc kubenswrapper[4612]: I0227 08:10:53.411188 4612 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="9a05b2bc-7c21-412d-b7ba-26e77517f60d" containerName="ceilometer-notification-agent" containerID="cri-o://031c23bd61e46974556b76e828661e3ce104da5c8221a415ba11063a8de5e6af" gracePeriod=30 Feb 27 08:10:53 crc kubenswrapper[4612]: I0227 08:10:53.417699 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-36be-account-create-update-7grtc" event={"ID":"e63e127a-2950-4a4c-a194-668b5142adb8","Type":"ContainerStarted","Data":"cdd9f292b431cd4aefc9a1fbba353ade1e867f18ac1d560e581e8d61e7bfc478"} Feb 27 08:10:53 crc kubenswrapper[4612]: I0227 08:10:53.417931 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-36be-account-create-update-7grtc" event={"ID":"e63e127a-2950-4a4c-a194-668b5142adb8","Type":"ContainerStarted","Data":"840e146d38a8fb95b33ef622a441eb43aa0420e18a17d68bcc902fe4051d0651"} Feb 27 08:10:53 crc kubenswrapper[4612]: I0227 08:10:53.430227 4612 generic.go:334] "Generic (PLEG): container finished" podID="06a51ad9-c1fb-452b-9c2b-62e4d2c80266" containerID="95d77cdca63cf202190670df3149188fa1b7295ea5a44c2f79b867a58f183e6e" exitCode=0 Feb 27 08:10:53 crc kubenswrapper[4612]: I0227 08:10:53.431078 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-c7w6b" event={"ID":"06a51ad9-c1fb-452b-9c2b-62e4d2c80266","Type":"ContainerDied","Data":"95d77cdca63cf202190670df3149188fa1b7295ea5a44c2f79b867a58f183e6e"} Feb 27 08:10:53 crc kubenswrapper[4612]: I0227 08:10:53.450865 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.708913684 podStartE2EDuration="7.450846071s" podCreationTimestamp="2026-02-27 08:10:46 +0000 UTC" firstStartedPulling="2026-02-27 08:10:47.243369175 +0000 UTC m=+1305.097299173" lastFinishedPulling="2026-02-27 08:10:52.985301572 +0000 UTC m=+1310.839231560" observedRunningTime="2026-02-27 08:10:53.436373556 +0000 UTC m=+1311.290303564" watchObservedRunningTime="2026-02-27 08:10:53.450846071 +0000 UTC m=+1311.304776069" Feb 27 08:10:53 crc kubenswrapper[4612]: I0227 08:10:53.464784 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-36be-account-create-update-7grtc" podStartSLOduration=2.464764331 podStartE2EDuration="2.464764331s" podCreationTimestamp="2026-02-27 08:10:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 08:10:53.456501614 +0000 UTC m=+1311.310431612" watchObservedRunningTime="2026-02-27 08:10:53.464764331 +0000 UTC m=+1311.318694329" Feb 27 08:10:54 crc kubenswrapper[4612]: I0227 08:10:54.439773 4612 generic.go:334] "Generic (PLEG): container finished" podID="a5162728-e496-4f30-9efb-a411e8e52ba4" containerID="143969eaf93c6f453f01dc76dd51a7faea9680aaf6f7dde73247b621512f99e2" exitCode=0 Feb 27 08:10:54 crc kubenswrapper[4612]: I0227 08:10:54.440315 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-tfl6p" event={"ID":"a5162728-e496-4f30-9efb-a411e8e52ba4","Type":"ContainerDied","Data":"143969eaf93c6f453f01dc76dd51a7faea9680aaf6f7dde73247b621512f99e2"} Feb 27 08:10:54 crc kubenswrapper[4612]: I0227 08:10:54.442934 4612 generic.go:334] "Generic (PLEG): container finished" podID="9a05b2bc-7c21-412d-b7ba-26e77517f60d" containerID="7b9eafefccc9df5427ba8a3a5ccd2e2d5026ab9bbca43a0ff2af18a3f38ffd5b" exitCode=0 Feb 27 08:10:54 crc kubenswrapper[4612]: I0227 08:10:54.442956 4612 generic.go:334] "Generic (PLEG): container finished" podID="9a05b2bc-7c21-412d-b7ba-26e77517f60d" containerID="371987e64261489b75417093a2cfe845de762e34592593234573c534290c186b" exitCode=2 Feb 27 08:10:54 crc kubenswrapper[4612]: I0227 08:10:54.442964 4612 generic.go:334] "Generic (PLEG): container finished" podID="9a05b2bc-7c21-412d-b7ba-26e77517f60d" containerID="031c23bd61e46974556b76e828661e3ce104da5c8221a415ba11063a8de5e6af" exitCode=0 Feb 27 08:10:54 crc kubenswrapper[4612]: I0227 08:10:54.442975 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9a05b2bc-7c21-412d-b7ba-26e77517f60d","Type":"ContainerDied","Data":"7b9eafefccc9df5427ba8a3a5ccd2e2d5026ab9bbca43a0ff2af18a3f38ffd5b"} Feb 27 08:10:54 crc kubenswrapper[4612]: I0227 08:10:54.443030 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9a05b2bc-7c21-412d-b7ba-26e77517f60d","Type":"ContainerDied","Data":"371987e64261489b75417093a2cfe845de762e34592593234573c534290c186b"} Feb 27 08:10:54 crc kubenswrapper[4612]: I0227 08:10:54.443044 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9a05b2bc-7c21-412d-b7ba-26e77517f60d","Type":"ContainerDied","Data":"031c23bd61e46974556b76e828661e3ce104da5c8221a415ba11063a8de5e6af"} Feb 27 08:10:54 crc kubenswrapper[4612]: I0227 08:10:54.444754 4612 generic.go:334] "Generic (PLEG): container finished" podID="e63e127a-2950-4a4c-a194-668b5142adb8" containerID="cdd9f292b431cd4aefc9a1fbba353ade1e867f18ac1d560e581e8d61e7bfc478" exitCode=0 Feb 27 08:10:54 crc kubenswrapper[4612]: I0227 08:10:54.444802 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-36be-account-create-update-7grtc" event={"ID":"e63e127a-2950-4a4c-a194-668b5142adb8","Type":"ContainerDied","Data":"cdd9f292b431cd4aefc9a1fbba353ade1e867f18ac1d560e581e8d61e7bfc478"} Feb 27 08:10:54 crc kubenswrapper[4612]: I0227 08:10:54.446968 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"9ba31a95-0fce-4eb6-98be-6292b7146e21","Type":"ContainerStarted","Data":"76d7ca4ab9641e4aee3a5575d42b884b6ec73d608f66abfcf8dd9a4fa051156f"} Feb 27 08:10:54 crc kubenswrapper[4612]: I0227 08:10:54.447089 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Feb 27 08:10:54 crc kubenswrapper[4612]: I0227 08:10:54.448630 4612 generic.go:334] "Generic (PLEG): container finished" podID="ed43d122-6dcc-4bce-9ed1-184dcb96647c" containerID="76720259afaeb97d1994408be42c233f8079c49ec8b77bf9ebb83772ca20fb35" exitCode=0 Feb 27 08:10:54 crc kubenswrapper[4612]: I0227 08:10:54.448685 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-5f7d-account-create-update-fptbd" event={"ID":"ed43d122-6dcc-4bce-9ed1-184dcb96647c","Type":"ContainerDied","Data":"76720259afaeb97d1994408be42c233f8079c49ec8b77bf9ebb83772ca20fb35"} Feb 27 08:10:54 crc kubenswrapper[4612]: I0227 08:10:54.450299 4612 generic.go:334] "Generic (PLEG): container finished" podID="d1b7e951-57b4-47f3-b337-b3468e7c6637" containerID="2cc2b1f42a72c486e6a061a906a7e9f77b61101b86e5eb8b8cd7dddefa341eb9" exitCode=0 Feb 27 08:10:54 crc kubenswrapper[4612]: I0227 08:10:54.450378 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-2ac7-account-create-update-hnvdn" event={"ID":"d1b7e951-57b4-47f3-b337-b3468e7c6637","Type":"ContainerDied","Data":"2cc2b1f42a72c486e6a061a906a7e9f77b61101b86e5eb8b8cd7dddefa341eb9"} Feb 27 08:10:54 crc kubenswrapper[4612]: I0227 08:10:54.515158 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=5.5151373150000005 podStartE2EDuration="5.515137315s" podCreationTimestamp="2026-02-27 08:10:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 08:10:54.509652318 +0000 UTC m=+1312.363582316" watchObservedRunningTime="2026-02-27 08:10:54.515137315 +0000 UTC m=+1312.369067313" Feb 27 08:10:56 crc kubenswrapper[4612]: I0227 08:10:56.023536 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-c7w6b" Feb 27 08:10:56 crc kubenswrapper[4612]: I0227 08:10:56.146307 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/06a51ad9-c1fb-452b-9c2b-62e4d2c80266-operator-scripts\") pod \"06a51ad9-c1fb-452b-9c2b-62e4d2c80266\" (UID: \"06a51ad9-c1fb-452b-9c2b-62e4d2c80266\") " Feb 27 08:10:56 crc kubenswrapper[4612]: I0227 08:10:56.146368 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ls4rt\" (UniqueName: \"kubernetes.io/projected/06a51ad9-c1fb-452b-9c2b-62e4d2c80266-kube-api-access-ls4rt\") pod \"06a51ad9-c1fb-452b-9c2b-62e4d2c80266\" (UID: \"06a51ad9-c1fb-452b-9c2b-62e4d2c80266\") " Feb 27 08:10:56 crc kubenswrapper[4612]: I0227 08:10:56.148786 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/06a51ad9-c1fb-452b-9c2b-62e4d2c80266-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "06a51ad9-c1fb-452b-9c2b-62e4d2c80266" (UID: "06a51ad9-c1fb-452b-9c2b-62e4d2c80266"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 08:10:56 crc kubenswrapper[4612]: I0227 08:10:56.188801 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/06a51ad9-c1fb-452b-9c2b-62e4d2c80266-kube-api-access-ls4rt" (OuterVolumeSpecName: "kube-api-access-ls4rt") pod "06a51ad9-c1fb-452b-9c2b-62e4d2c80266" (UID: "06a51ad9-c1fb-452b-9c2b-62e4d2c80266"). InnerVolumeSpecName "kube-api-access-ls4rt". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 08:10:56 crc kubenswrapper[4612]: I0227 08:10:56.265878 4612 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/06a51ad9-c1fb-452b-9c2b-62e4d2c80266-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 27 08:10:56 crc kubenswrapper[4612]: I0227 08:10:56.265916 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ls4rt\" (UniqueName: \"kubernetes.io/projected/06a51ad9-c1fb-452b-9c2b-62e4d2c80266-kube-api-access-ls4rt\") on node \"crc\" DevicePath \"\"" Feb 27 08:10:56 crc kubenswrapper[4612]: I0227 08:10:56.387946 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-pqvhc" Feb 27 08:10:56 crc kubenswrapper[4612]: I0227 08:10:56.400042 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-tfl6p" Feb 27 08:10:56 crc kubenswrapper[4612]: I0227 08:10:56.401446 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-36be-account-create-update-7grtc" Feb 27 08:10:56 crc kubenswrapper[4612]: I0227 08:10:56.420419 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-2ac7-account-create-update-hnvdn" Feb 27 08:10:56 crc kubenswrapper[4612]: I0227 08:10:56.438176 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-5f7d-account-create-update-fptbd" Feb 27 08:10:56 crc kubenswrapper[4612]: I0227 08:10:56.472532 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a5162728-e496-4f30-9efb-a411e8e52ba4-operator-scripts\") pod \"a5162728-e496-4f30-9efb-a411e8e52ba4\" (UID: \"a5162728-e496-4f30-9efb-a411e8e52ba4\") " Feb 27 08:10:56 crc kubenswrapper[4612]: I0227 08:10:56.472634 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mgqhr\" (UniqueName: \"kubernetes.io/projected/5c30c00d-ed94-4980-a218-a2e3b95bd8eb-kube-api-access-mgqhr\") pod \"5c30c00d-ed94-4980-a218-a2e3b95bd8eb\" (UID: \"5c30c00d-ed94-4980-a218-a2e3b95bd8eb\") " Feb 27 08:10:56 crc kubenswrapper[4612]: I0227 08:10:56.472694 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ed43d122-6dcc-4bce-9ed1-184dcb96647c-operator-scripts\") pod \"ed43d122-6dcc-4bce-9ed1-184dcb96647c\" (UID: \"ed43d122-6dcc-4bce-9ed1-184dcb96647c\") " Feb 27 08:10:56 crc kubenswrapper[4612]: I0227 08:10:56.472713 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2vlv\" (UniqueName: \"kubernetes.io/projected/e63e127a-2950-4a4c-a194-668b5142adb8-kube-api-access-x2vlv\") pod \"e63e127a-2950-4a4c-a194-668b5142adb8\" (UID: \"e63e127a-2950-4a4c-a194-668b5142adb8\") " Feb 27 08:10:56 crc kubenswrapper[4612]: I0227 08:10:56.472734 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8krrv\" (UniqueName: \"kubernetes.io/projected/ed43d122-6dcc-4bce-9ed1-184dcb96647c-kube-api-access-8krrv\") pod \"ed43d122-6dcc-4bce-9ed1-184dcb96647c\" (UID: \"ed43d122-6dcc-4bce-9ed1-184dcb96647c\") " Feb 27 08:10:56 crc kubenswrapper[4612]: I0227 08:10:56.473156 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e63e127a-2950-4a4c-a194-668b5142adb8-operator-scripts\") pod \"e63e127a-2950-4a4c-a194-668b5142adb8\" (UID: \"e63e127a-2950-4a4c-a194-668b5142adb8\") " Feb 27 08:10:56 crc kubenswrapper[4612]: I0227 08:10:56.473205 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m546h\" (UniqueName: \"kubernetes.io/projected/a5162728-e496-4f30-9efb-a411e8e52ba4-kube-api-access-m546h\") pod \"a5162728-e496-4f30-9efb-a411e8e52ba4\" (UID: \"a5162728-e496-4f30-9efb-a411e8e52ba4\") " Feb 27 08:10:56 crc kubenswrapper[4612]: I0227 08:10:56.473280 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5c30c00d-ed94-4980-a218-a2e3b95bd8eb-operator-scripts\") pod \"5c30c00d-ed94-4980-a218-a2e3b95bd8eb\" (UID: \"5c30c00d-ed94-4980-a218-a2e3b95bd8eb\") " Feb 27 08:10:56 crc kubenswrapper[4612]: I0227 08:10:56.473317 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kzdkv\" (UniqueName: \"kubernetes.io/projected/d1b7e951-57b4-47f3-b337-b3468e7c6637-kube-api-access-kzdkv\") pod \"d1b7e951-57b4-47f3-b337-b3468e7c6637\" (UID: \"d1b7e951-57b4-47f3-b337-b3468e7c6637\") " Feb 27 08:10:56 crc kubenswrapper[4612]: I0227 08:10:56.473392 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d1b7e951-57b4-47f3-b337-b3468e7c6637-operator-scripts\") pod \"d1b7e951-57b4-47f3-b337-b3468e7c6637\" (UID: \"d1b7e951-57b4-47f3-b337-b3468e7c6637\") " Feb 27 08:10:56 crc kubenswrapper[4612]: I0227 08:10:56.474558 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d1b7e951-57b4-47f3-b337-b3468e7c6637-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "d1b7e951-57b4-47f3-b337-b3468e7c6637" (UID: "d1b7e951-57b4-47f3-b337-b3468e7c6637"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 08:10:56 crc kubenswrapper[4612]: I0227 08:10:56.474567 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ed43d122-6dcc-4bce-9ed1-184dcb96647c-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "ed43d122-6dcc-4bce-9ed1-184dcb96647c" (UID: "ed43d122-6dcc-4bce-9ed1-184dcb96647c"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 08:10:56 crc kubenswrapper[4612]: I0227 08:10:56.475235 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a5162728-e496-4f30-9efb-a411e8e52ba4-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "a5162728-e496-4f30-9efb-a411e8e52ba4" (UID: "a5162728-e496-4f30-9efb-a411e8e52ba4"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 08:10:56 crc kubenswrapper[4612]: I0227 08:10:56.475915 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e63e127a-2950-4a4c-a194-668b5142adb8-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "e63e127a-2950-4a4c-a194-668b5142adb8" (UID: "e63e127a-2950-4a4c-a194-668b5142adb8"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 08:10:56 crc kubenswrapper[4612]: I0227 08:10:56.479062 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5c30c00d-ed94-4980-a218-a2e3b95bd8eb-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "5c30c00d-ed94-4980-a218-a2e3b95bd8eb" (UID: "5c30c00d-ed94-4980-a218-a2e3b95bd8eb"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 08:10:56 crc kubenswrapper[4612]: I0227 08:10:56.479152 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a5162728-e496-4f30-9efb-a411e8e52ba4-kube-api-access-m546h" (OuterVolumeSpecName: "kube-api-access-m546h") pod "a5162728-e496-4f30-9efb-a411e8e52ba4" (UID: "a5162728-e496-4f30-9efb-a411e8e52ba4"). InnerVolumeSpecName "kube-api-access-m546h". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 08:10:56 crc kubenswrapper[4612]: I0227 08:10:56.479208 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5c30c00d-ed94-4980-a218-a2e3b95bd8eb-kube-api-access-mgqhr" (OuterVolumeSpecName: "kube-api-access-mgqhr") pod "5c30c00d-ed94-4980-a218-a2e3b95bd8eb" (UID: "5c30c00d-ed94-4980-a218-a2e3b95bd8eb"). InnerVolumeSpecName "kube-api-access-mgqhr". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 08:10:56 crc kubenswrapper[4612]: I0227 08:10:56.479690 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e63e127a-2950-4a4c-a194-668b5142adb8-kube-api-access-x2vlv" (OuterVolumeSpecName: "kube-api-access-x2vlv") pod "e63e127a-2950-4a4c-a194-668b5142adb8" (UID: "e63e127a-2950-4a4c-a194-668b5142adb8"). InnerVolumeSpecName "kube-api-access-x2vlv". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 08:10:56 crc kubenswrapper[4612]: I0227 08:10:56.480474 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d1b7e951-57b4-47f3-b337-b3468e7c6637-kube-api-access-kzdkv" (OuterVolumeSpecName: "kube-api-access-kzdkv") pod "d1b7e951-57b4-47f3-b337-b3468e7c6637" (UID: "d1b7e951-57b4-47f3-b337-b3468e7c6637"). InnerVolumeSpecName "kube-api-access-kzdkv". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 08:10:56 crc kubenswrapper[4612]: I0227 08:10:56.483809 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ed43d122-6dcc-4bce-9ed1-184dcb96647c-kube-api-access-8krrv" (OuterVolumeSpecName: "kube-api-access-8krrv") pod "ed43d122-6dcc-4bce-9ed1-184dcb96647c" (UID: "ed43d122-6dcc-4bce-9ed1-184dcb96647c"). InnerVolumeSpecName "kube-api-access-8krrv". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 08:10:56 crc kubenswrapper[4612]: I0227 08:10:56.543088 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-pqvhc" event={"ID":"5c30c00d-ed94-4980-a218-a2e3b95bd8eb","Type":"ContainerDied","Data":"72f46a1607606dd0aaa5a24ee872882b463e59b7d0f973888e5d048e09f7a541"} Feb 27 08:10:56 crc kubenswrapper[4612]: I0227 08:10:56.543130 4612 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="72f46a1607606dd0aaa5a24ee872882b463e59b7d0f973888e5d048e09f7a541" Feb 27 08:10:56 crc kubenswrapper[4612]: I0227 08:10:56.543189 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-pqvhc" Feb 27 08:10:56 crc kubenswrapper[4612]: I0227 08:10:56.549045 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-36be-account-create-update-7grtc" event={"ID":"e63e127a-2950-4a4c-a194-668b5142adb8","Type":"ContainerDied","Data":"840e146d38a8fb95b33ef622a441eb43aa0420e18a17d68bcc902fe4051d0651"} Feb 27 08:10:56 crc kubenswrapper[4612]: I0227 08:10:56.549184 4612 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="840e146d38a8fb95b33ef622a441eb43aa0420e18a17d68bcc902fe4051d0651" Feb 27 08:10:56 crc kubenswrapper[4612]: I0227 08:10:56.549243 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-36be-account-create-update-7grtc" Feb 27 08:10:56 crc kubenswrapper[4612]: I0227 08:10:56.553023 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-c7w6b" event={"ID":"06a51ad9-c1fb-452b-9c2b-62e4d2c80266","Type":"ContainerDied","Data":"a74a6ae364b9adc5f3d373f3c65f93353162f32546bccca3fe1f2c1206591088"} Feb 27 08:10:56 crc kubenswrapper[4612]: I0227 08:10:56.553055 4612 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a74a6ae364b9adc5f3d373f3c65f93353162f32546bccca3fe1f2c1206591088" Feb 27 08:10:56 crc kubenswrapper[4612]: I0227 08:10:56.553096 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-c7w6b" Feb 27 08:10:56 crc kubenswrapper[4612]: I0227 08:10:56.554849 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-5f7d-account-create-update-fptbd" event={"ID":"ed43d122-6dcc-4bce-9ed1-184dcb96647c","Type":"ContainerDied","Data":"3f035dc833d2fed58f08924c97c7fe7b4ddf5a74892934d4fa7cd390dfe5763d"} Feb 27 08:10:56 crc kubenswrapper[4612]: I0227 08:10:56.554872 4612 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3f035dc833d2fed58f08924c97c7fe7b4ddf5a74892934d4fa7cd390dfe5763d" Feb 27 08:10:56 crc kubenswrapper[4612]: I0227 08:10:56.554911 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-5f7d-account-create-update-fptbd" Feb 27 08:10:56 crc kubenswrapper[4612]: I0227 08:10:56.556631 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-2ac7-account-create-update-hnvdn" Feb 27 08:10:56 crc kubenswrapper[4612]: I0227 08:10:56.556660 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-2ac7-account-create-update-hnvdn" event={"ID":"d1b7e951-57b4-47f3-b337-b3468e7c6637","Type":"ContainerDied","Data":"7ce27fd45040abdd7bde16a6fb28d92c377217b2d0f61511643b36ac566e1f03"} Feb 27 08:10:56 crc kubenswrapper[4612]: I0227 08:10:56.556726 4612 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7ce27fd45040abdd7bde16a6fb28d92c377217b2d0f61511643b36ac566e1f03" Feb 27 08:10:56 crc kubenswrapper[4612]: I0227 08:10:56.568527 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-tfl6p" event={"ID":"a5162728-e496-4f30-9efb-a411e8e52ba4","Type":"ContainerDied","Data":"4e9319ce04de4731fd5845c2868786a7476ed6adc59f015bc3c618818b366cd6"} Feb 27 08:10:56 crc kubenswrapper[4612]: I0227 08:10:56.568569 4612 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4e9319ce04de4731fd5845c2868786a7476ed6adc59f015bc3c618818b366cd6" Feb 27 08:10:56 crc kubenswrapper[4612]: I0227 08:10:56.568621 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-tfl6p" Feb 27 08:10:56 crc kubenswrapper[4612]: I0227 08:10:56.576434 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mgqhr\" (UniqueName: \"kubernetes.io/projected/5c30c00d-ed94-4980-a218-a2e3b95bd8eb-kube-api-access-mgqhr\") on node \"crc\" DevicePath \"\"" Feb 27 08:10:56 crc kubenswrapper[4612]: I0227 08:10:56.576460 4612 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ed43d122-6dcc-4bce-9ed1-184dcb96647c-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 27 08:10:56 crc kubenswrapper[4612]: I0227 08:10:56.576499 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2vlv\" (UniqueName: \"kubernetes.io/projected/e63e127a-2950-4a4c-a194-668b5142adb8-kube-api-access-x2vlv\") on node \"crc\" DevicePath \"\"" Feb 27 08:10:56 crc kubenswrapper[4612]: I0227 08:10:56.576509 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8krrv\" (UniqueName: \"kubernetes.io/projected/ed43d122-6dcc-4bce-9ed1-184dcb96647c-kube-api-access-8krrv\") on node \"crc\" DevicePath \"\"" Feb 27 08:10:56 crc kubenswrapper[4612]: I0227 08:10:56.576518 4612 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e63e127a-2950-4a4c-a194-668b5142adb8-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 27 08:10:56 crc kubenswrapper[4612]: I0227 08:10:56.576528 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m546h\" (UniqueName: \"kubernetes.io/projected/a5162728-e496-4f30-9efb-a411e8e52ba4-kube-api-access-m546h\") on node \"crc\" DevicePath \"\"" Feb 27 08:10:56 crc kubenswrapper[4612]: I0227 08:10:56.576537 4612 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5c30c00d-ed94-4980-a218-a2e3b95bd8eb-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 27 08:10:56 crc kubenswrapper[4612]: I0227 08:10:56.576546 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kzdkv\" (UniqueName: \"kubernetes.io/projected/d1b7e951-57b4-47f3-b337-b3468e7c6637-kube-api-access-kzdkv\") on node \"crc\" DevicePath \"\"" Feb 27 08:10:56 crc kubenswrapper[4612]: I0227 08:10:56.576554 4612 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d1b7e951-57b4-47f3-b337-b3468e7c6637-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 27 08:10:56 crc kubenswrapper[4612]: I0227 08:10:56.576563 4612 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a5162728-e496-4f30-9efb-a411e8e52ba4-operator-scripts\") on node \"crc\" DevicePath \"\"" Feb 27 08:10:57 crc kubenswrapper[4612]: I0227 08:10:57.621021 4612 generic.go:334] "Generic (PLEG): container finished" podID="9a05b2bc-7c21-412d-b7ba-26e77517f60d" containerID="c9c7680af8a800caac979e94217d6e08c6a44613c883dc27baa42950bf08b3c4" exitCode=0 Feb 27 08:10:57 crc kubenswrapper[4612]: I0227 08:10:57.621332 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9a05b2bc-7c21-412d-b7ba-26e77517f60d","Type":"ContainerDied","Data":"c9c7680af8a800caac979e94217d6e08c6a44613c883dc27baa42950bf08b3c4"} Feb 27 08:10:57 crc kubenswrapper[4612]: I0227 08:10:57.742299 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 27 08:10:57 crc kubenswrapper[4612]: I0227 08:10:57.903644 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9a05b2bc-7c21-412d-b7ba-26e77517f60d-combined-ca-bundle\") pod \"9a05b2bc-7c21-412d-b7ba-26e77517f60d\" (UID: \"9a05b2bc-7c21-412d-b7ba-26e77517f60d\") " Feb 27 08:10:57 crc kubenswrapper[4612]: I0227 08:10:57.903900 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9a05b2bc-7c21-412d-b7ba-26e77517f60d-run-httpd\") pod \"9a05b2bc-7c21-412d-b7ba-26e77517f60d\" (UID: \"9a05b2bc-7c21-412d-b7ba-26e77517f60d\") " Feb 27 08:10:57 crc kubenswrapper[4612]: I0227 08:10:57.904031 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9a05b2bc-7c21-412d-b7ba-26e77517f60d-scripts\") pod \"9a05b2bc-7c21-412d-b7ba-26e77517f60d\" (UID: \"9a05b2bc-7c21-412d-b7ba-26e77517f60d\") " Feb 27 08:10:57 crc kubenswrapper[4612]: I0227 08:10:57.904171 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9a05b2bc-7c21-412d-b7ba-26e77517f60d-sg-core-conf-yaml\") pod \"9a05b2bc-7c21-412d-b7ba-26e77517f60d\" (UID: \"9a05b2bc-7c21-412d-b7ba-26e77517f60d\") " Feb 27 08:10:57 crc kubenswrapper[4612]: I0227 08:10:57.904314 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hnh2l\" (UniqueName: \"kubernetes.io/projected/9a05b2bc-7c21-412d-b7ba-26e77517f60d-kube-api-access-hnh2l\") pod \"9a05b2bc-7c21-412d-b7ba-26e77517f60d\" (UID: \"9a05b2bc-7c21-412d-b7ba-26e77517f60d\") " Feb 27 08:10:57 crc kubenswrapper[4612]: I0227 08:10:57.904402 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9a05b2bc-7c21-412d-b7ba-26e77517f60d-log-httpd\") pod \"9a05b2bc-7c21-412d-b7ba-26e77517f60d\" (UID: \"9a05b2bc-7c21-412d-b7ba-26e77517f60d\") " Feb 27 08:10:57 crc kubenswrapper[4612]: I0227 08:10:57.904551 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9a05b2bc-7c21-412d-b7ba-26e77517f60d-config-data\") pod \"9a05b2bc-7c21-412d-b7ba-26e77517f60d\" (UID: \"9a05b2bc-7c21-412d-b7ba-26e77517f60d\") " Feb 27 08:10:57 crc kubenswrapper[4612]: I0227 08:10:57.904755 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9a05b2bc-7c21-412d-b7ba-26e77517f60d-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "9a05b2bc-7c21-412d-b7ba-26e77517f60d" (UID: "9a05b2bc-7c21-412d-b7ba-26e77517f60d"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 08:10:57 crc kubenswrapper[4612]: I0227 08:10:57.904946 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9a05b2bc-7c21-412d-b7ba-26e77517f60d-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "9a05b2bc-7c21-412d-b7ba-26e77517f60d" (UID: "9a05b2bc-7c21-412d-b7ba-26e77517f60d"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 08:10:57 crc kubenswrapper[4612]: I0227 08:10:57.905939 4612 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9a05b2bc-7c21-412d-b7ba-26e77517f60d-run-httpd\") on node \"crc\" DevicePath \"\"" Feb 27 08:10:57 crc kubenswrapper[4612]: I0227 08:10:57.906169 4612 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9a05b2bc-7c21-412d-b7ba-26e77517f60d-log-httpd\") on node \"crc\" DevicePath \"\"" Feb 27 08:10:57 crc kubenswrapper[4612]: I0227 08:10:57.912871 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9a05b2bc-7c21-412d-b7ba-26e77517f60d-kube-api-access-hnh2l" (OuterVolumeSpecName: "kube-api-access-hnh2l") pod "9a05b2bc-7c21-412d-b7ba-26e77517f60d" (UID: "9a05b2bc-7c21-412d-b7ba-26e77517f60d"). InnerVolumeSpecName "kube-api-access-hnh2l". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 08:10:57 crc kubenswrapper[4612]: I0227 08:10:57.926618 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9a05b2bc-7c21-412d-b7ba-26e77517f60d-scripts" (OuterVolumeSpecName: "scripts") pod "9a05b2bc-7c21-412d-b7ba-26e77517f60d" (UID: "9a05b2bc-7c21-412d-b7ba-26e77517f60d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:10:57 crc kubenswrapper[4612]: I0227 08:10:57.931944 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9a05b2bc-7c21-412d-b7ba-26e77517f60d-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "9a05b2bc-7c21-412d-b7ba-26e77517f60d" (UID: "9a05b2bc-7c21-412d-b7ba-26e77517f60d"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:10:58 crc kubenswrapper[4612]: I0227 08:10:58.008082 4612 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9a05b2bc-7c21-412d-b7ba-26e77517f60d-scripts\") on node \"crc\" DevicePath \"\"" Feb 27 08:10:58 crc kubenswrapper[4612]: I0227 08:10:58.008105 4612 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9a05b2bc-7c21-412d-b7ba-26e77517f60d-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Feb 27 08:10:58 crc kubenswrapper[4612]: I0227 08:10:58.008115 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hnh2l\" (UniqueName: \"kubernetes.io/projected/9a05b2bc-7c21-412d-b7ba-26e77517f60d-kube-api-access-hnh2l\") on node \"crc\" DevicePath \"\"" Feb 27 08:10:58 crc kubenswrapper[4612]: I0227 08:10:58.009009 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9a05b2bc-7c21-412d-b7ba-26e77517f60d-config-data" (OuterVolumeSpecName: "config-data") pod "9a05b2bc-7c21-412d-b7ba-26e77517f60d" (UID: "9a05b2bc-7c21-412d-b7ba-26e77517f60d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:10:58 crc kubenswrapper[4612]: I0227 08:10:58.071157 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9a05b2bc-7c21-412d-b7ba-26e77517f60d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9a05b2bc-7c21-412d-b7ba-26e77517f60d" (UID: "9a05b2bc-7c21-412d-b7ba-26e77517f60d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:10:58 crc kubenswrapper[4612]: I0227 08:10:58.110542 4612 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9a05b2bc-7c21-412d-b7ba-26e77517f60d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 08:10:58 crc kubenswrapper[4612]: I0227 08:10:58.110575 4612 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9a05b2bc-7c21-412d-b7ba-26e77517f60d-config-data\") on node \"crc\" DevicePath \"\"" Feb 27 08:10:58 crc kubenswrapper[4612]: I0227 08:10:58.655825 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9a05b2bc-7c21-412d-b7ba-26e77517f60d","Type":"ContainerDied","Data":"f7d6c1da1dc0352307fccccf8c1045c0c7a5161657111f79d8d8dfab4d214fa1"} Feb 27 08:10:58 crc kubenswrapper[4612]: I0227 08:10:58.655908 4612 scope.go:117] "RemoveContainer" containerID="7b9eafefccc9df5427ba8a3a5ccd2e2d5026ab9bbca43a0ff2af18a3f38ffd5b" Feb 27 08:10:58 crc kubenswrapper[4612]: I0227 08:10:58.656016 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 27 08:10:58 crc kubenswrapper[4612]: I0227 08:10:58.694564 4612 scope.go:117] "RemoveContainer" containerID="371987e64261489b75417093a2cfe845de762e34592593234573c534290c186b" Feb 27 08:10:58 crc kubenswrapper[4612]: I0227 08:10:58.722160 4612 scope.go:117] "RemoveContainer" containerID="031c23bd61e46974556b76e828661e3ce104da5c8221a415ba11063a8de5e6af" Feb 27 08:10:58 crc kubenswrapper[4612]: I0227 08:10:58.722277 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 27 08:10:58 crc kubenswrapper[4612]: I0227 08:10:58.753494 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Feb 27 08:10:58 crc kubenswrapper[4612]: I0227 08:10:58.756979 4612 scope.go:117] "RemoveContainer" containerID="c9c7680af8a800caac979e94217d6e08c6a44613c883dc27baa42950bf08b3c4" Feb 27 08:10:58 crc kubenswrapper[4612]: I0227 08:10:58.765861 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Feb 27 08:10:58 crc kubenswrapper[4612]: E0227 08:10:58.766428 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="06a51ad9-c1fb-452b-9c2b-62e4d2c80266" containerName="mariadb-database-create" Feb 27 08:10:58 crc kubenswrapper[4612]: I0227 08:10:58.766500 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="06a51ad9-c1fb-452b-9c2b-62e4d2c80266" containerName="mariadb-database-create" Feb 27 08:10:58 crc kubenswrapper[4612]: E0227 08:10:58.766566 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a5162728-e496-4f30-9efb-a411e8e52ba4" containerName="mariadb-database-create" Feb 27 08:10:58 crc kubenswrapper[4612]: I0227 08:10:58.766624 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="a5162728-e496-4f30-9efb-a411e8e52ba4" containerName="mariadb-database-create" Feb 27 08:10:58 crc kubenswrapper[4612]: E0227 08:10:58.766694 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a05b2bc-7c21-412d-b7ba-26e77517f60d" containerName="ceilometer-notification-agent" Feb 27 08:10:58 crc kubenswrapper[4612]: I0227 08:10:58.766774 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a05b2bc-7c21-412d-b7ba-26e77517f60d" containerName="ceilometer-notification-agent" Feb 27 08:10:58 crc kubenswrapper[4612]: E0227 08:10:58.766832 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a05b2bc-7c21-412d-b7ba-26e77517f60d" containerName="sg-core" Feb 27 08:10:58 crc kubenswrapper[4612]: I0227 08:10:58.766890 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a05b2bc-7c21-412d-b7ba-26e77517f60d" containerName="sg-core" Feb 27 08:10:58 crc kubenswrapper[4612]: E0227 08:10:58.766944 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a05b2bc-7c21-412d-b7ba-26e77517f60d" containerName="proxy-httpd" Feb 27 08:10:58 crc kubenswrapper[4612]: I0227 08:10:58.766991 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a05b2bc-7c21-412d-b7ba-26e77517f60d" containerName="proxy-httpd" Feb 27 08:10:58 crc kubenswrapper[4612]: E0227 08:10:58.767054 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ed43d122-6dcc-4bce-9ed1-184dcb96647c" containerName="mariadb-account-create-update" Feb 27 08:10:58 crc kubenswrapper[4612]: I0227 08:10:58.767118 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="ed43d122-6dcc-4bce-9ed1-184dcb96647c" containerName="mariadb-account-create-update" Feb 27 08:10:58 crc kubenswrapper[4612]: E0227 08:10:58.767191 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a05b2bc-7c21-412d-b7ba-26e77517f60d" containerName="ceilometer-central-agent" Feb 27 08:10:58 crc kubenswrapper[4612]: I0227 08:10:58.767260 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a05b2bc-7c21-412d-b7ba-26e77517f60d" containerName="ceilometer-central-agent" Feb 27 08:10:58 crc kubenswrapper[4612]: E0227 08:10:58.767329 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e63e127a-2950-4a4c-a194-668b5142adb8" containerName="mariadb-account-create-update" Feb 27 08:10:58 crc kubenswrapper[4612]: I0227 08:10:58.767418 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="e63e127a-2950-4a4c-a194-668b5142adb8" containerName="mariadb-account-create-update" Feb 27 08:10:58 crc kubenswrapper[4612]: E0227 08:10:58.767479 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1b7e951-57b4-47f3-b337-b3468e7c6637" containerName="mariadb-account-create-update" Feb 27 08:10:58 crc kubenswrapper[4612]: I0227 08:10:58.767530 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1b7e951-57b4-47f3-b337-b3468e7c6637" containerName="mariadb-account-create-update" Feb 27 08:10:58 crc kubenswrapper[4612]: E0227 08:10:58.767595 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c30c00d-ed94-4980-a218-a2e3b95bd8eb" containerName="mariadb-database-create" Feb 27 08:10:58 crc kubenswrapper[4612]: I0227 08:10:58.767648 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c30c00d-ed94-4980-a218-a2e3b95bd8eb" containerName="mariadb-database-create" Feb 27 08:10:58 crc kubenswrapper[4612]: I0227 08:10:58.767902 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="a5162728-e496-4f30-9efb-a411e8e52ba4" containerName="mariadb-database-create" Feb 27 08:10:58 crc kubenswrapper[4612]: I0227 08:10:58.768003 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="9a05b2bc-7c21-412d-b7ba-26e77517f60d" containerName="sg-core" Feb 27 08:10:58 crc kubenswrapper[4612]: I0227 08:10:58.768069 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="e63e127a-2950-4a4c-a194-668b5142adb8" containerName="mariadb-account-create-update" Feb 27 08:10:58 crc kubenswrapper[4612]: I0227 08:10:58.768124 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="5c30c00d-ed94-4980-a218-a2e3b95bd8eb" containerName="mariadb-database-create" Feb 27 08:10:58 crc kubenswrapper[4612]: I0227 08:10:58.768187 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="9a05b2bc-7c21-412d-b7ba-26e77517f60d" containerName="ceilometer-notification-agent" Feb 27 08:10:58 crc kubenswrapper[4612]: I0227 08:10:58.768264 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="9a05b2bc-7c21-412d-b7ba-26e77517f60d" containerName="proxy-httpd" Feb 27 08:10:58 crc kubenswrapper[4612]: I0227 08:10:58.768339 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="9a05b2bc-7c21-412d-b7ba-26e77517f60d" containerName="ceilometer-central-agent" Feb 27 08:10:58 crc kubenswrapper[4612]: I0227 08:10:58.768401 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="06a51ad9-c1fb-452b-9c2b-62e4d2c80266" containerName="mariadb-database-create" Feb 27 08:10:58 crc kubenswrapper[4612]: I0227 08:10:58.768455 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="d1b7e951-57b4-47f3-b337-b3468e7c6637" containerName="mariadb-account-create-update" Feb 27 08:10:58 crc kubenswrapper[4612]: I0227 08:10:58.768527 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="ed43d122-6dcc-4bce-9ed1-184dcb96647c" containerName="mariadb-account-create-update" Feb 27 08:10:58 crc kubenswrapper[4612]: I0227 08:10:58.770259 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 27 08:10:58 crc kubenswrapper[4612]: I0227 08:10:58.779988 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 27 08:10:58 crc kubenswrapper[4612]: I0227 08:10:58.783265 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Feb 27 08:10:58 crc kubenswrapper[4612]: I0227 08:10:58.783470 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Feb 27 08:10:58 crc kubenswrapper[4612]: I0227 08:10:58.783589 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Feb 27 08:10:58 crc kubenswrapper[4612]: I0227 08:10:58.822714 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rrlc4\" (UniqueName: \"kubernetes.io/projected/a72550b1-f880-4d07-b097-fe298c8265b1-kube-api-access-rrlc4\") pod \"ceilometer-0\" (UID: \"a72550b1-f880-4d07-b097-fe298c8265b1\") " pod="openstack/ceilometer-0" Feb 27 08:10:58 crc kubenswrapper[4612]: I0227 08:10:58.822770 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a72550b1-f880-4d07-b097-fe298c8265b1-log-httpd\") pod \"ceilometer-0\" (UID: \"a72550b1-f880-4d07-b097-fe298c8265b1\") " pod="openstack/ceilometer-0" Feb 27 08:10:58 crc kubenswrapper[4612]: I0227 08:10:58.822793 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a72550b1-f880-4d07-b097-fe298c8265b1-run-httpd\") pod \"ceilometer-0\" (UID: \"a72550b1-f880-4d07-b097-fe298c8265b1\") " pod="openstack/ceilometer-0" Feb 27 08:10:58 crc kubenswrapper[4612]: I0227 08:10:58.822824 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a72550b1-f880-4d07-b097-fe298c8265b1-config-data\") pod \"ceilometer-0\" (UID: \"a72550b1-f880-4d07-b097-fe298c8265b1\") " pod="openstack/ceilometer-0" Feb 27 08:10:58 crc kubenswrapper[4612]: I0227 08:10:58.822961 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a72550b1-f880-4d07-b097-fe298c8265b1-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a72550b1-f880-4d07-b097-fe298c8265b1\") " pod="openstack/ceilometer-0" Feb 27 08:10:58 crc kubenswrapper[4612]: I0227 08:10:58.822999 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/a72550b1-f880-4d07-b097-fe298c8265b1-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"a72550b1-f880-4d07-b097-fe298c8265b1\") " pod="openstack/ceilometer-0" Feb 27 08:10:58 crc kubenswrapper[4612]: I0227 08:10:58.823101 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a72550b1-f880-4d07-b097-fe298c8265b1-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a72550b1-f880-4d07-b097-fe298c8265b1\") " pod="openstack/ceilometer-0" Feb 27 08:10:58 crc kubenswrapper[4612]: I0227 08:10:58.823205 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a72550b1-f880-4d07-b097-fe298c8265b1-scripts\") pod \"ceilometer-0\" (UID: \"a72550b1-f880-4d07-b097-fe298c8265b1\") " pod="openstack/ceilometer-0" Feb 27 08:10:58 crc kubenswrapper[4612]: I0227 08:10:58.864747 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9a05b2bc-7c21-412d-b7ba-26e77517f60d" path="/var/lib/kubelet/pods/9a05b2bc-7c21-412d-b7ba-26e77517f60d/volumes" Feb 27 08:10:58 crc kubenswrapper[4612]: I0227 08:10:58.924223 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rrlc4\" (UniqueName: \"kubernetes.io/projected/a72550b1-f880-4d07-b097-fe298c8265b1-kube-api-access-rrlc4\") pod \"ceilometer-0\" (UID: \"a72550b1-f880-4d07-b097-fe298c8265b1\") " pod="openstack/ceilometer-0" Feb 27 08:10:58 crc kubenswrapper[4612]: I0227 08:10:58.924265 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a72550b1-f880-4d07-b097-fe298c8265b1-log-httpd\") pod \"ceilometer-0\" (UID: \"a72550b1-f880-4d07-b097-fe298c8265b1\") " pod="openstack/ceilometer-0" Feb 27 08:10:58 crc kubenswrapper[4612]: I0227 08:10:58.924286 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a72550b1-f880-4d07-b097-fe298c8265b1-run-httpd\") pod \"ceilometer-0\" (UID: \"a72550b1-f880-4d07-b097-fe298c8265b1\") " pod="openstack/ceilometer-0" Feb 27 08:10:58 crc kubenswrapper[4612]: I0227 08:10:58.924316 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a72550b1-f880-4d07-b097-fe298c8265b1-config-data\") pod \"ceilometer-0\" (UID: \"a72550b1-f880-4d07-b097-fe298c8265b1\") " pod="openstack/ceilometer-0" Feb 27 08:10:58 crc kubenswrapper[4612]: I0227 08:10:58.924369 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a72550b1-f880-4d07-b097-fe298c8265b1-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a72550b1-f880-4d07-b097-fe298c8265b1\") " pod="openstack/ceilometer-0" Feb 27 08:10:58 crc kubenswrapper[4612]: I0227 08:10:58.924383 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/a72550b1-f880-4d07-b097-fe298c8265b1-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"a72550b1-f880-4d07-b097-fe298c8265b1\") " pod="openstack/ceilometer-0" Feb 27 08:10:58 crc kubenswrapper[4612]: I0227 08:10:58.924416 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a72550b1-f880-4d07-b097-fe298c8265b1-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a72550b1-f880-4d07-b097-fe298c8265b1\") " pod="openstack/ceilometer-0" Feb 27 08:10:58 crc kubenswrapper[4612]: I0227 08:10:58.924452 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a72550b1-f880-4d07-b097-fe298c8265b1-scripts\") pod \"ceilometer-0\" (UID: \"a72550b1-f880-4d07-b097-fe298c8265b1\") " pod="openstack/ceilometer-0" Feb 27 08:10:58 crc kubenswrapper[4612]: I0227 08:10:58.925176 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a72550b1-f880-4d07-b097-fe298c8265b1-run-httpd\") pod \"ceilometer-0\" (UID: \"a72550b1-f880-4d07-b097-fe298c8265b1\") " pod="openstack/ceilometer-0" Feb 27 08:10:58 crc kubenswrapper[4612]: I0227 08:10:58.925342 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a72550b1-f880-4d07-b097-fe298c8265b1-log-httpd\") pod \"ceilometer-0\" (UID: \"a72550b1-f880-4d07-b097-fe298c8265b1\") " pod="openstack/ceilometer-0" Feb 27 08:10:58 crc kubenswrapper[4612]: I0227 08:10:58.930391 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a72550b1-f880-4d07-b097-fe298c8265b1-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a72550b1-f880-4d07-b097-fe298c8265b1\") " pod="openstack/ceilometer-0" Feb 27 08:10:58 crc kubenswrapper[4612]: I0227 08:10:58.930483 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/a72550b1-f880-4d07-b097-fe298c8265b1-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"a72550b1-f880-4d07-b097-fe298c8265b1\") " pod="openstack/ceilometer-0" Feb 27 08:10:58 crc kubenswrapper[4612]: I0227 08:10:58.932229 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a72550b1-f880-4d07-b097-fe298c8265b1-config-data\") pod \"ceilometer-0\" (UID: \"a72550b1-f880-4d07-b097-fe298c8265b1\") " pod="openstack/ceilometer-0" Feb 27 08:10:58 crc kubenswrapper[4612]: I0227 08:10:58.934668 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a72550b1-f880-4d07-b097-fe298c8265b1-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a72550b1-f880-4d07-b097-fe298c8265b1\") " pod="openstack/ceilometer-0" Feb 27 08:10:58 crc kubenswrapper[4612]: I0227 08:10:58.937348 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 27 08:10:58 crc kubenswrapper[4612]: E0227 08:10:58.938251 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[kube-api-access-rrlc4 scripts], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/ceilometer-0" podUID="a72550b1-f880-4d07-b097-fe298c8265b1" Feb 27 08:10:58 crc kubenswrapper[4612]: I0227 08:10:58.946048 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rrlc4\" (UniqueName: \"kubernetes.io/projected/a72550b1-f880-4d07-b097-fe298c8265b1-kube-api-access-rrlc4\") pod \"ceilometer-0\" (UID: \"a72550b1-f880-4d07-b097-fe298c8265b1\") " pod="openstack/ceilometer-0" Feb 27 08:10:58 crc kubenswrapper[4612]: I0227 08:10:58.956451 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a72550b1-f880-4d07-b097-fe298c8265b1-scripts\") pod \"ceilometer-0\" (UID: \"a72550b1-f880-4d07-b097-fe298c8265b1\") " pod="openstack/ceilometer-0" Feb 27 08:10:59 crc kubenswrapper[4612]: I0227 08:10:59.106861 4612 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-6b985547d4-mkfpv" Feb 27 08:10:59 crc kubenswrapper[4612]: I0227 08:10:59.106913 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-6b985547d4-mkfpv" Feb 27 08:10:59 crc kubenswrapper[4612]: I0227 08:10:59.108589 4612 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-6b985547d4-mkfpv" podUID="76009243-16b3-4652-abfe-a6cdba363724" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.152:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.152:8443: connect: connection refused" Feb 27 08:10:59 crc kubenswrapper[4612]: I0227 08:10:59.271519 4612 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-77948db5bb-97qh6" Feb 27 08:10:59 crc kubenswrapper[4612]: I0227 08:10:59.272022 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-77948db5bb-97qh6" Feb 27 08:10:59 crc kubenswrapper[4612]: I0227 08:10:59.272668 4612 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-77948db5bb-97qh6" podUID="b72b541d-4661-44a7-a121-c2a8aef6db11" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.153:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.153:8443: connect: connection refused" Feb 27 08:10:59 crc kubenswrapper[4612]: I0227 08:10:59.665178 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 27 08:10:59 crc kubenswrapper[4612]: I0227 08:10:59.674826 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 27 08:10:59 crc kubenswrapper[4612]: I0227 08:10:59.739146 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/a72550b1-f880-4d07-b097-fe298c8265b1-ceilometer-tls-certs\") pod \"a72550b1-f880-4d07-b097-fe298c8265b1\" (UID: \"a72550b1-f880-4d07-b097-fe298c8265b1\") " Feb 27 08:10:59 crc kubenswrapper[4612]: I0227 08:10:59.739217 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a72550b1-f880-4d07-b097-fe298c8265b1-run-httpd\") pod \"a72550b1-f880-4d07-b097-fe298c8265b1\" (UID: \"a72550b1-f880-4d07-b097-fe298c8265b1\") " Feb 27 08:10:59 crc kubenswrapper[4612]: I0227 08:10:59.739265 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rrlc4\" (UniqueName: \"kubernetes.io/projected/a72550b1-f880-4d07-b097-fe298c8265b1-kube-api-access-rrlc4\") pod \"a72550b1-f880-4d07-b097-fe298c8265b1\" (UID: \"a72550b1-f880-4d07-b097-fe298c8265b1\") " Feb 27 08:10:59 crc kubenswrapper[4612]: I0227 08:10:59.739297 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a72550b1-f880-4d07-b097-fe298c8265b1-sg-core-conf-yaml\") pod \"a72550b1-f880-4d07-b097-fe298c8265b1\" (UID: \"a72550b1-f880-4d07-b097-fe298c8265b1\") " Feb 27 08:10:59 crc kubenswrapper[4612]: I0227 08:10:59.739355 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a72550b1-f880-4d07-b097-fe298c8265b1-config-data\") pod \"a72550b1-f880-4d07-b097-fe298c8265b1\" (UID: \"a72550b1-f880-4d07-b097-fe298c8265b1\") " Feb 27 08:10:59 crc kubenswrapper[4612]: I0227 08:10:59.739399 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a72550b1-f880-4d07-b097-fe298c8265b1-log-httpd\") pod \"a72550b1-f880-4d07-b097-fe298c8265b1\" (UID: \"a72550b1-f880-4d07-b097-fe298c8265b1\") " Feb 27 08:10:59 crc kubenswrapper[4612]: I0227 08:10:59.739467 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a72550b1-f880-4d07-b097-fe298c8265b1-scripts\") pod \"a72550b1-f880-4d07-b097-fe298c8265b1\" (UID: \"a72550b1-f880-4d07-b097-fe298c8265b1\") " Feb 27 08:10:59 crc kubenswrapper[4612]: I0227 08:10:59.739483 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a72550b1-f880-4d07-b097-fe298c8265b1-combined-ca-bundle\") pod \"a72550b1-f880-4d07-b097-fe298c8265b1\" (UID: \"a72550b1-f880-4d07-b097-fe298c8265b1\") " Feb 27 08:10:59 crc kubenswrapper[4612]: I0227 08:10:59.740480 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a72550b1-f880-4d07-b097-fe298c8265b1-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "a72550b1-f880-4d07-b097-fe298c8265b1" (UID: "a72550b1-f880-4d07-b097-fe298c8265b1"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 08:10:59 crc kubenswrapper[4612]: I0227 08:10:59.740770 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a72550b1-f880-4d07-b097-fe298c8265b1-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "a72550b1-f880-4d07-b097-fe298c8265b1" (UID: "a72550b1-f880-4d07-b097-fe298c8265b1"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 08:10:59 crc kubenswrapper[4612]: I0227 08:10:59.746839 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a72550b1-f880-4d07-b097-fe298c8265b1-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "a72550b1-f880-4d07-b097-fe298c8265b1" (UID: "a72550b1-f880-4d07-b097-fe298c8265b1"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:10:59 crc kubenswrapper[4612]: I0227 08:10:59.758874 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a72550b1-f880-4d07-b097-fe298c8265b1-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "a72550b1-f880-4d07-b097-fe298c8265b1" (UID: "a72550b1-f880-4d07-b097-fe298c8265b1"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:10:59 crc kubenswrapper[4612]: I0227 08:10:59.764080 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a72550b1-f880-4d07-b097-fe298c8265b1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a72550b1-f880-4d07-b097-fe298c8265b1" (UID: "a72550b1-f880-4d07-b097-fe298c8265b1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:10:59 crc kubenswrapper[4612]: I0227 08:10:59.765835 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a72550b1-f880-4d07-b097-fe298c8265b1-config-data" (OuterVolumeSpecName: "config-data") pod "a72550b1-f880-4d07-b097-fe298c8265b1" (UID: "a72550b1-f880-4d07-b097-fe298c8265b1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:10:59 crc kubenswrapper[4612]: I0227 08:10:59.768873 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a72550b1-f880-4d07-b097-fe298c8265b1-kube-api-access-rrlc4" (OuterVolumeSpecName: "kube-api-access-rrlc4") pod "a72550b1-f880-4d07-b097-fe298c8265b1" (UID: "a72550b1-f880-4d07-b097-fe298c8265b1"). InnerVolumeSpecName "kube-api-access-rrlc4". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 08:10:59 crc kubenswrapper[4612]: I0227 08:10:59.768947 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a72550b1-f880-4d07-b097-fe298c8265b1-scripts" (OuterVolumeSpecName: "scripts") pod "a72550b1-f880-4d07-b097-fe298c8265b1" (UID: "a72550b1-f880-4d07-b097-fe298c8265b1"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:10:59 crc kubenswrapper[4612]: I0227 08:10:59.772337 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Feb 27 08:10:59 crc kubenswrapper[4612]: I0227 08:10:59.840657 4612 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a72550b1-f880-4d07-b097-fe298c8265b1-config-data\") on node \"crc\" DevicePath \"\"" Feb 27 08:10:59 crc kubenswrapper[4612]: I0227 08:10:59.840683 4612 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a72550b1-f880-4d07-b097-fe298c8265b1-log-httpd\") on node \"crc\" DevicePath \"\"" Feb 27 08:10:59 crc kubenswrapper[4612]: I0227 08:10:59.840694 4612 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a72550b1-f880-4d07-b097-fe298c8265b1-scripts\") on node \"crc\" DevicePath \"\"" Feb 27 08:10:59 crc kubenswrapper[4612]: I0227 08:10:59.840798 4612 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a72550b1-f880-4d07-b097-fe298c8265b1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 08:10:59 crc kubenswrapper[4612]: I0227 08:10:59.840925 4612 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/a72550b1-f880-4d07-b097-fe298c8265b1-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 27 08:10:59 crc kubenswrapper[4612]: I0227 08:10:59.840950 4612 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a72550b1-f880-4d07-b097-fe298c8265b1-run-httpd\") on node \"crc\" DevicePath \"\"" Feb 27 08:10:59 crc kubenswrapper[4612]: I0227 08:10:59.840961 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rrlc4\" (UniqueName: \"kubernetes.io/projected/a72550b1-f880-4d07-b097-fe298c8265b1-kube-api-access-rrlc4\") on node \"crc\" DevicePath \"\"" Feb 27 08:10:59 crc kubenswrapper[4612]: I0227 08:10:59.840972 4612 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a72550b1-f880-4d07-b097-fe298c8265b1-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Feb 27 08:11:00 crc kubenswrapper[4612]: I0227 08:11:00.673278 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 27 08:11:00 crc kubenswrapper[4612]: I0227 08:11:00.735471 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 27 08:11:00 crc kubenswrapper[4612]: I0227 08:11:00.794769 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Feb 27 08:11:00 crc kubenswrapper[4612]: I0227 08:11:00.831777 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Feb 27 08:11:00 crc kubenswrapper[4612]: I0227 08:11:00.834102 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 27 08:11:00 crc kubenswrapper[4612]: I0227 08:11:00.841471 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Feb 27 08:11:00 crc kubenswrapper[4612]: I0227 08:11:00.841733 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Feb 27 08:11:00 crc kubenswrapper[4612]: I0227 08:11:00.841915 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Feb 27 08:11:00 crc kubenswrapper[4612]: I0227 08:11:00.888292 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a72550b1-f880-4d07-b097-fe298c8265b1" path="/var/lib/kubelet/pods/a72550b1-f880-4d07-b097-fe298c8265b1/volumes" Feb 27 08:11:00 crc kubenswrapper[4612]: I0227 08:11:00.888671 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 27 08:11:00 crc kubenswrapper[4612]: I0227 08:11:00.966903 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bbac0dc6-01b5-4211-94b0-9d5bcebd9cc1-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"bbac0dc6-01b5-4211-94b0-9d5bcebd9cc1\") " pod="openstack/ceilometer-0" Feb 27 08:11:00 crc kubenswrapper[4612]: I0227 08:11:00.966947 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ll4pt\" (UniqueName: \"kubernetes.io/projected/bbac0dc6-01b5-4211-94b0-9d5bcebd9cc1-kube-api-access-ll4pt\") pod \"ceilometer-0\" (UID: \"bbac0dc6-01b5-4211-94b0-9d5bcebd9cc1\") " pod="openstack/ceilometer-0" Feb 27 08:11:00 crc kubenswrapper[4612]: I0227 08:11:00.966975 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/bbac0dc6-01b5-4211-94b0-9d5bcebd9cc1-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"bbac0dc6-01b5-4211-94b0-9d5bcebd9cc1\") " pod="openstack/ceilometer-0" Feb 27 08:11:00 crc kubenswrapper[4612]: I0227 08:11:00.967045 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bbac0dc6-01b5-4211-94b0-9d5bcebd9cc1-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"bbac0dc6-01b5-4211-94b0-9d5bcebd9cc1\") " pod="openstack/ceilometer-0" Feb 27 08:11:00 crc kubenswrapper[4612]: I0227 08:11:00.967765 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bbac0dc6-01b5-4211-94b0-9d5bcebd9cc1-scripts\") pod \"ceilometer-0\" (UID: \"bbac0dc6-01b5-4211-94b0-9d5bcebd9cc1\") " pod="openstack/ceilometer-0" Feb 27 08:11:00 crc kubenswrapper[4612]: I0227 08:11:00.967937 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bbac0dc6-01b5-4211-94b0-9d5bcebd9cc1-run-httpd\") pod \"ceilometer-0\" (UID: \"bbac0dc6-01b5-4211-94b0-9d5bcebd9cc1\") " pod="openstack/ceilometer-0" Feb 27 08:11:00 crc kubenswrapper[4612]: I0227 08:11:00.967972 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bbac0dc6-01b5-4211-94b0-9d5bcebd9cc1-log-httpd\") pod \"ceilometer-0\" (UID: \"bbac0dc6-01b5-4211-94b0-9d5bcebd9cc1\") " pod="openstack/ceilometer-0" Feb 27 08:11:00 crc kubenswrapper[4612]: I0227 08:11:00.968048 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bbac0dc6-01b5-4211-94b0-9d5bcebd9cc1-config-data\") pod \"ceilometer-0\" (UID: \"bbac0dc6-01b5-4211-94b0-9d5bcebd9cc1\") " pod="openstack/ceilometer-0" Feb 27 08:11:01 crc kubenswrapper[4612]: I0227 08:11:01.069655 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bbac0dc6-01b5-4211-94b0-9d5bcebd9cc1-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"bbac0dc6-01b5-4211-94b0-9d5bcebd9cc1\") " pod="openstack/ceilometer-0" Feb 27 08:11:01 crc kubenswrapper[4612]: I0227 08:11:01.070444 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bbac0dc6-01b5-4211-94b0-9d5bcebd9cc1-scripts\") pod \"ceilometer-0\" (UID: \"bbac0dc6-01b5-4211-94b0-9d5bcebd9cc1\") " pod="openstack/ceilometer-0" Feb 27 08:11:01 crc kubenswrapper[4612]: I0227 08:11:01.070523 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bbac0dc6-01b5-4211-94b0-9d5bcebd9cc1-run-httpd\") pod \"ceilometer-0\" (UID: \"bbac0dc6-01b5-4211-94b0-9d5bcebd9cc1\") " pod="openstack/ceilometer-0" Feb 27 08:11:01 crc kubenswrapper[4612]: I0227 08:11:01.070545 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bbac0dc6-01b5-4211-94b0-9d5bcebd9cc1-log-httpd\") pod \"ceilometer-0\" (UID: \"bbac0dc6-01b5-4211-94b0-9d5bcebd9cc1\") " pod="openstack/ceilometer-0" Feb 27 08:11:01 crc kubenswrapper[4612]: I0227 08:11:01.070588 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bbac0dc6-01b5-4211-94b0-9d5bcebd9cc1-config-data\") pod \"ceilometer-0\" (UID: \"bbac0dc6-01b5-4211-94b0-9d5bcebd9cc1\") " pod="openstack/ceilometer-0" Feb 27 08:11:01 crc kubenswrapper[4612]: I0227 08:11:01.070612 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bbac0dc6-01b5-4211-94b0-9d5bcebd9cc1-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"bbac0dc6-01b5-4211-94b0-9d5bcebd9cc1\") " pod="openstack/ceilometer-0" Feb 27 08:11:01 crc kubenswrapper[4612]: I0227 08:11:01.070632 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ll4pt\" (UniqueName: \"kubernetes.io/projected/bbac0dc6-01b5-4211-94b0-9d5bcebd9cc1-kube-api-access-ll4pt\") pod \"ceilometer-0\" (UID: \"bbac0dc6-01b5-4211-94b0-9d5bcebd9cc1\") " pod="openstack/ceilometer-0" Feb 27 08:11:01 crc kubenswrapper[4612]: I0227 08:11:01.070653 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/bbac0dc6-01b5-4211-94b0-9d5bcebd9cc1-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"bbac0dc6-01b5-4211-94b0-9d5bcebd9cc1\") " pod="openstack/ceilometer-0" Feb 27 08:11:01 crc kubenswrapper[4612]: I0227 08:11:01.071217 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bbac0dc6-01b5-4211-94b0-9d5bcebd9cc1-run-httpd\") pod \"ceilometer-0\" (UID: \"bbac0dc6-01b5-4211-94b0-9d5bcebd9cc1\") " pod="openstack/ceilometer-0" Feb 27 08:11:01 crc kubenswrapper[4612]: I0227 08:11:01.071876 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bbac0dc6-01b5-4211-94b0-9d5bcebd9cc1-log-httpd\") pod \"ceilometer-0\" (UID: \"bbac0dc6-01b5-4211-94b0-9d5bcebd9cc1\") " pod="openstack/ceilometer-0" Feb 27 08:11:01 crc kubenswrapper[4612]: I0227 08:11:01.077396 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/bbac0dc6-01b5-4211-94b0-9d5bcebd9cc1-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"bbac0dc6-01b5-4211-94b0-9d5bcebd9cc1\") " pod="openstack/ceilometer-0" Feb 27 08:11:01 crc kubenswrapper[4612]: I0227 08:11:01.077677 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bbac0dc6-01b5-4211-94b0-9d5bcebd9cc1-config-data\") pod \"ceilometer-0\" (UID: \"bbac0dc6-01b5-4211-94b0-9d5bcebd9cc1\") " pod="openstack/ceilometer-0" Feb 27 08:11:01 crc kubenswrapper[4612]: I0227 08:11:01.079240 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bbac0dc6-01b5-4211-94b0-9d5bcebd9cc1-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"bbac0dc6-01b5-4211-94b0-9d5bcebd9cc1\") " pod="openstack/ceilometer-0" Feb 27 08:11:01 crc kubenswrapper[4612]: I0227 08:11:01.079641 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bbac0dc6-01b5-4211-94b0-9d5bcebd9cc1-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"bbac0dc6-01b5-4211-94b0-9d5bcebd9cc1\") " pod="openstack/ceilometer-0" Feb 27 08:11:01 crc kubenswrapper[4612]: I0227 08:11:01.080265 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bbac0dc6-01b5-4211-94b0-9d5bcebd9cc1-scripts\") pod \"ceilometer-0\" (UID: \"bbac0dc6-01b5-4211-94b0-9d5bcebd9cc1\") " pod="openstack/ceilometer-0" Feb 27 08:11:01 crc kubenswrapper[4612]: I0227 08:11:01.089011 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ll4pt\" (UniqueName: \"kubernetes.io/projected/bbac0dc6-01b5-4211-94b0-9d5bcebd9cc1-kube-api-access-ll4pt\") pod \"ceilometer-0\" (UID: \"bbac0dc6-01b5-4211-94b0-9d5bcebd9cc1\") " pod="openstack/ceilometer-0" Feb 27 08:11:01 crc kubenswrapper[4612]: I0227 08:11:01.155836 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 27 08:11:01 crc kubenswrapper[4612]: I0227 08:11:01.722322 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 27 08:11:01 crc kubenswrapper[4612]: I0227 08:11:01.777360 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-z67kr"] Feb 27 08:11:01 crc kubenswrapper[4612]: I0227 08:11:01.788017 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-z67kr" Feb 27 08:11:01 crc kubenswrapper[4612]: I0227 08:11:01.797219 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-9qmk6" Feb 27 08:11:01 crc kubenswrapper[4612]: I0227 08:11:01.797388 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Feb 27 08:11:01 crc kubenswrapper[4612]: I0227 08:11:01.797495 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Feb 27 08:11:01 crc kubenswrapper[4612]: I0227 08:11:01.814995 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-z67kr"] Feb 27 08:11:01 crc kubenswrapper[4612]: I0227 08:11:01.893586 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3d52b072-aded-480f-ae46-d89c9be0c9f8-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-z67kr\" (UID: \"3d52b072-aded-480f-ae46-d89c9be0c9f8\") " pod="openstack/nova-cell0-conductor-db-sync-z67kr" Feb 27 08:11:01 crc kubenswrapper[4612]: I0227 08:11:01.893633 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gw2x6\" (UniqueName: \"kubernetes.io/projected/3d52b072-aded-480f-ae46-d89c9be0c9f8-kube-api-access-gw2x6\") pod \"nova-cell0-conductor-db-sync-z67kr\" (UID: \"3d52b072-aded-480f-ae46-d89c9be0c9f8\") " pod="openstack/nova-cell0-conductor-db-sync-z67kr" Feb 27 08:11:01 crc kubenswrapper[4612]: I0227 08:11:01.893657 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3d52b072-aded-480f-ae46-d89c9be0c9f8-scripts\") pod \"nova-cell0-conductor-db-sync-z67kr\" (UID: \"3d52b072-aded-480f-ae46-d89c9be0c9f8\") " pod="openstack/nova-cell0-conductor-db-sync-z67kr" Feb 27 08:11:01 crc kubenswrapper[4612]: I0227 08:11:01.893823 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3d52b072-aded-480f-ae46-d89c9be0c9f8-config-data\") pod \"nova-cell0-conductor-db-sync-z67kr\" (UID: \"3d52b072-aded-480f-ae46-d89c9be0c9f8\") " pod="openstack/nova-cell0-conductor-db-sync-z67kr" Feb 27 08:11:01 crc kubenswrapper[4612]: I0227 08:11:01.995227 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3d52b072-aded-480f-ae46-d89c9be0c9f8-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-z67kr\" (UID: \"3d52b072-aded-480f-ae46-d89c9be0c9f8\") " pod="openstack/nova-cell0-conductor-db-sync-z67kr" Feb 27 08:11:01 crc kubenswrapper[4612]: I0227 08:11:01.995279 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gw2x6\" (UniqueName: \"kubernetes.io/projected/3d52b072-aded-480f-ae46-d89c9be0c9f8-kube-api-access-gw2x6\") pod \"nova-cell0-conductor-db-sync-z67kr\" (UID: \"3d52b072-aded-480f-ae46-d89c9be0c9f8\") " pod="openstack/nova-cell0-conductor-db-sync-z67kr" Feb 27 08:11:01 crc kubenswrapper[4612]: I0227 08:11:01.995302 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3d52b072-aded-480f-ae46-d89c9be0c9f8-scripts\") pod \"nova-cell0-conductor-db-sync-z67kr\" (UID: \"3d52b072-aded-480f-ae46-d89c9be0c9f8\") " pod="openstack/nova-cell0-conductor-db-sync-z67kr" Feb 27 08:11:01 crc kubenswrapper[4612]: I0227 08:11:01.995420 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3d52b072-aded-480f-ae46-d89c9be0c9f8-config-data\") pod \"nova-cell0-conductor-db-sync-z67kr\" (UID: \"3d52b072-aded-480f-ae46-d89c9be0c9f8\") " pod="openstack/nova-cell0-conductor-db-sync-z67kr" Feb 27 08:11:02 crc kubenswrapper[4612]: I0227 08:11:02.001015 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3d52b072-aded-480f-ae46-d89c9be0c9f8-scripts\") pod \"nova-cell0-conductor-db-sync-z67kr\" (UID: \"3d52b072-aded-480f-ae46-d89c9be0c9f8\") " pod="openstack/nova-cell0-conductor-db-sync-z67kr" Feb 27 08:11:02 crc kubenswrapper[4612]: I0227 08:11:02.002477 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3d52b072-aded-480f-ae46-d89c9be0c9f8-config-data\") pod \"nova-cell0-conductor-db-sync-z67kr\" (UID: \"3d52b072-aded-480f-ae46-d89c9be0c9f8\") " pod="openstack/nova-cell0-conductor-db-sync-z67kr" Feb 27 08:11:02 crc kubenswrapper[4612]: I0227 08:11:02.003270 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3d52b072-aded-480f-ae46-d89c9be0c9f8-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-z67kr\" (UID: \"3d52b072-aded-480f-ae46-d89c9be0c9f8\") " pod="openstack/nova-cell0-conductor-db-sync-z67kr" Feb 27 08:11:02 crc kubenswrapper[4612]: I0227 08:11:02.020215 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gw2x6\" (UniqueName: \"kubernetes.io/projected/3d52b072-aded-480f-ae46-d89c9be0c9f8-kube-api-access-gw2x6\") pod \"nova-cell0-conductor-db-sync-z67kr\" (UID: \"3d52b072-aded-480f-ae46-d89c9be0c9f8\") " pod="openstack/nova-cell0-conductor-db-sync-z67kr" Feb 27 08:11:02 crc kubenswrapper[4612]: I0227 08:11:02.114268 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-z67kr" Feb 27 08:11:02 crc kubenswrapper[4612]: I0227 08:11:02.624400 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 27 08:11:02 crc kubenswrapper[4612]: I0227 08:11:02.695788 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bbac0dc6-01b5-4211-94b0-9d5bcebd9cc1","Type":"ContainerStarted","Data":"b66a95e205ddfd70faa5b402d5655ad70db606ef3ee22f1865fffe56be8ed95a"} Feb 27 08:11:02 crc kubenswrapper[4612]: I0227 08:11:02.695832 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bbac0dc6-01b5-4211-94b0-9d5bcebd9cc1","Type":"ContainerStarted","Data":"9f1dd0ccd92fc13b12c7659b8fa079855a4374d0efb72f04382e46676c35f730"} Feb 27 08:11:02 crc kubenswrapper[4612]: I0227 08:11:02.704436 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-z67kr"] Feb 27 08:11:03 crc kubenswrapper[4612]: I0227 08:11:03.024775 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Feb 27 08:11:03 crc kubenswrapper[4612]: I0227 08:11:03.718413 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bbac0dc6-01b5-4211-94b0-9d5bcebd9cc1","Type":"ContainerStarted","Data":"6b62d31b4a918b5c83e01dcd5b3b469fd7cc15867557f458bb2e371820cd2c57"} Feb 27 08:11:03 crc kubenswrapper[4612]: I0227 08:11:03.729145 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-z67kr" event={"ID":"3d52b072-aded-480f-ae46-d89c9be0c9f8","Type":"ContainerStarted","Data":"425959e9762745eaa61d56db0138f50190e189c59350c8b4a043bbb6bcbd8271"} Feb 27 08:11:04 crc kubenswrapper[4612]: I0227 08:11:04.747160 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bbac0dc6-01b5-4211-94b0-9d5bcebd9cc1","Type":"ContainerStarted","Data":"ba79e5fe70c18c20497563a68988c8690081e0f0ba75488b280b88ca5b83d16e"} Feb 27 08:11:06 crc kubenswrapper[4612]: I0227 08:11:06.801530 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bbac0dc6-01b5-4211-94b0-9d5bcebd9cc1","Type":"ContainerStarted","Data":"2d06ff36815d20fb5029475648cc84007b2ac9385d41e9a28806917db2bd7829"} Feb 27 08:11:06 crc kubenswrapper[4612]: I0227 08:11:06.802002 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Feb 27 08:11:06 crc kubenswrapper[4612]: I0227 08:11:06.802035 4612 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="bbac0dc6-01b5-4211-94b0-9d5bcebd9cc1" containerName="proxy-httpd" containerID="cri-o://2d06ff36815d20fb5029475648cc84007b2ac9385d41e9a28806917db2bd7829" gracePeriod=30 Feb 27 08:11:06 crc kubenswrapper[4612]: I0227 08:11:06.802284 4612 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="bbac0dc6-01b5-4211-94b0-9d5bcebd9cc1" containerName="sg-core" containerID="cri-o://ba79e5fe70c18c20497563a68988c8690081e0f0ba75488b280b88ca5b83d16e" gracePeriod=30 Feb 27 08:11:06 crc kubenswrapper[4612]: I0227 08:11:06.802320 4612 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="bbac0dc6-01b5-4211-94b0-9d5bcebd9cc1" containerName="ceilometer-notification-agent" containerID="cri-o://6b62d31b4a918b5c83e01dcd5b3b469fd7cc15867557f458bb2e371820cd2c57" gracePeriod=30 Feb 27 08:11:06 crc kubenswrapper[4612]: I0227 08:11:06.802468 4612 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="bbac0dc6-01b5-4211-94b0-9d5bcebd9cc1" containerName="ceilometer-central-agent" containerID="cri-o://b66a95e205ddfd70faa5b402d5655ad70db606ef3ee22f1865fffe56be8ed95a" gracePeriod=30 Feb 27 08:11:06 crc kubenswrapper[4612]: I0227 08:11:06.842831 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.653899975 podStartE2EDuration="6.842809352s" podCreationTimestamp="2026-02-27 08:11:00 +0000 UTC" firstStartedPulling="2026-02-27 08:11:01.745888409 +0000 UTC m=+1319.599818407" lastFinishedPulling="2026-02-27 08:11:05.934797786 +0000 UTC m=+1323.788727784" observedRunningTime="2026-02-27 08:11:06.830987493 +0000 UTC m=+1324.684917491" watchObservedRunningTime="2026-02-27 08:11:06.842809352 +0000 UTC m=+1324.696739350" Feb 27 08:11:07 crc kubenswrapper[4612]: I0227 08:11:07.838509 4612 generic.go:334] "Generic (PLEG): container finished" podID="bbac0dc6-01b5-4211-94b0-9d5bcebd9cc1" containerID="2d06ff36815d20fb5029475648cc84007b2ac9385d41e9a28806917db2bd7829" exitCode=0 Feb 27 08:11:07 crc kubenswrapper[4612]: I0227 08:11:07.838845 4612 generic.go:334] "Generic (PLEG): container finished" podID="bbac0dc6-01b5-4211-94b0-9d5bcebd9cc1" containerID="ba79e5fe70c18c20497563a68988c8690081e0f0ba75488b280b88ca5b83d16e" exitCode=2 Feb 27 08:11:07 crc kubenswrapper[4612]: I0227 08:11:07.838853 4612 generic.go:334] "Generic (PLEG): container finished" podID="bbac0dc6-01b5-4211-94b0-9d5bcebd9cc1" containerID="6b62d31b4a918b5c83e01dcd5b3b469fd7cc15867557f458bb2e371820cd2c57" exitCode=0 Feb 27 08:11:07 crc kubenswrapper[4612]: I0227 08:11:07.838652 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bbac0dc6-01b5-4211-94b0-9d5bcebd9cc1","Type":"ContainerDied","Data":"2d06ff36815d20fb5029475648cc84007b2ac9385d41e9a28806917db2bd7829"} Feb 27 08:11:07 crc kubenswrapper[4612]: I0227 08:11:07.838882 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bbac0dc6-01b5-4211-94b0-9d5bcebd9cc1","Type":"ContainerDied","Data":"ba79e5fe70c18c20497563a68988c8690081e0f0ba75488b280b88ca5b83d16e"} Feb 27 08:11:07 crc kubenswrapper[4612]: I0227 08:11:07.838892 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bbac0dc6-01b5-4211-94b0-9d5bcebd9cc1","Type":"ContainerDied","Data":"6b62d31b4a918b5c83e01dcd5b3b469fd7cc15867557f458bb2e371820cd2c57"} Feb 27 08:11:09 crc kubenswrapper[4612]: I0227 08:11:09.109394 4612 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-6b985547d4-mkfpv" podUID="76009243-16b3-4652-abfe-a6cdba363724" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.152:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.152:8443: connect: connection refused" Feb 27 08:11:09 crc kubenswrapper[4612]: I0227 08:11:09.273976 4612 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-77948db5bb-97qh6" podUID="b72b541d-4661-44a7-a121-c2a8aef6db11" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.153:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.153:8443: connect: connection refused" Feb 27 08:11:13 crc kubenswrapper[4612]: I0227 08:11:13.758799 4612 scope.go:117] "RemoveContainer" containerID="98c86f4e5a52fa64298df451c4bba1d93fe8e52894afdbdbddb9b042118fb3f1" Feb 27 08:11:14 crc kubenswrapper[4612]: I0227 08:11:14.585116 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 27 08:11:14 crc kubenswrapper[4612]: I0227 08:11:14.753836 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bbac0dc6-01b5-4211-94b0-9d5bcebd9cc1-run-httpd\") pod \"bbac0dc6-01b5-4211-94b0-9d5bcebd9cc1\" (UID: \"bbac0dc6-01b5-4211-94b0-9d5bcebd9cc1\") " Feb 27 08:11:14 crc kubenswrapper[4612]: I0227 08:11:14.753878 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bbac0dc6-01b5-4211-94b0-9d5bcebd9cc1-log-httpd\") pod \"bbac0dc6-01b5-4211-94b0-9d5bcebd9cc1\" (UID: \"bbac0dc6-01b5-4211-94b0-9d5bcebd9cc1\") " Feb 27 08:11:14 crc kubenswrapper[4612]: I0227 08:11:14.754161 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bbac0dc6-01b5-4211-94b0-9d5bcebd9cc1-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "bbac0dc6-01b5-4211-94b0-9d5bcebd9cc1" (UID: "bbac0dc6-01b5-4211-94b0-9d5bcebd9cc1"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 08:11:14 crc kubenswrapper[4612]: I0227 08:11:14.754276 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bbac0dc6-01b5-4211-94b0-9d5bcebd9cc1-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "bbac0dc6-01b5-4211-94b0-9d5bcebd9cc1" (UID: "bbac0dc6-01b5-4211-94b0-9d5bcebd9cc1"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 08:11:14 crc kubenswrapper[4612]: I0227 08:11:14.754315 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bbac0dc6-01b5-4211-94b0-9d5bcebd9cc1-combined-ca-bundle\") pod \"bbac0dc6-01b5-4211-94b0-9d5bcebd9cc1\" (UID: \"bbac0dc6-01b5-4211-94b0-9d5bcebd9cc1\") " Feb 27 08:11:14 crc kubenswrapper[4612]: I0227 08:11:14.754711 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ll4pt\" (UniqueName: \"kubernetes.io/projected/bbac0dc6-01b5-4211-94b0-9d5bcebd9cc1-kube-api-access-ll4pt\") pod \"bbac0dc6-01b5-4211-94b0-9d5bcebd9cc1\" (UID: \"bbac0dc6-01b5-4211-94b0-9d5bcebd9cc1\") " Feb 27 08:11:14 crc kubenswrapper[4612]: I0227 08:11:14.755222 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bbac0dc6-01b5-4211-94b0-9d5bcebd9cc1-config-data\") pod \"bbac0dc6-01b5-4211-94b0-9d5bcebd9cc1\" (UID: \"bbac0dc6-01b5-4211-94b0-9d5bcebd9cc1\") " Feb 27 08:11:14 crc kubenswrapper[4612]: I0227 08:11:14.755276 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/bbac0dc6-01b5-4211-94b0-9d5bcebd9cc1-ceilometer-tls-certs\") pod \"bbac0dc6-01b5-4211-94b0-9d5bcebd9cc1\" (UID: \"bbac0dc6-01b5-4211-94b0-9d5bcebd9cc1\") " Feb 27 08:11:14 crc kubenswrapper[4612]: I0227 08:11:14.755324 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bbac0dc6-01b5-4211-94b0-9d5bcebd9cc1-sg-core-conf-yaml\") pod \"bbac0dc6-01b5-4211-94b0-9d5bcebd9cc1\" (UID: \"bbac0dc6-01b5-4211-94b0-9d5bcebd9cc1\") " Feb 27 08:11:14 crc kubenswrapper[4612]: I0227 08:11:14.755359 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bbac0dc6-01b5-4211-94b0-9d5bcebd9cc1-scripts\") pod \"bbac0dc6-01b5-4211-94b0-9d5bcebd9cc1\" (UID: \"bbac0dc6-01b5-4211-94b0-9d5bcebd9cc1\") " Feb 27 08:11:14 crc kubenswrapper[4612]: I0227 08:11:14.755644 4612 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bbac0dc6-01b5-4211-94b0-9d5bcebd9cc1-run-httpd\") on node \"crc\" DevicePath \"\"" Feb 27 08:11:14 crc kubenswrapper[4612]: I0227 08:11:14.755658 4612 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bbac0dc6-01b5-4211-94b0-9d5bcebd9cc1-log-httpd\") on node \"crc\" DevicePath \"\"" Feb 27 08:11:14 crc kubenswrapper[4612]: I0227 08:11:14.762809 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bbac0dc6-01b5-4211-94b0-9d5bcebd9cc1-scripts" (OuterVolumeSpecName: "scripts") pod "bbac0dc6-01b5-4211-94b0-9d5bcebd9cc1" (UID: "bbac0dc6-01b5-4211-94b0-9d5bcebd9cc1"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:11:14 crc kubenswrapper[4612]: I0227 08:11:14.762851 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bbac0dc6-01b5-4211-94b0-9d5bcebd9cc1-kube-api-access-ll4pt" (OuterVolumeSpecName: "kube-api-access-ll4pt") pod "bbac0dc6-01b5-4211-94b0-9d5bcebd9cc1" (UID: "bbac0dc6-01b5-4211-94b0-9d5bcebd9cc1"). InnerVolumeSpecName "kube-api-access-ll4pt". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 08:11:14 crc kubenswrapper[4612]: I0227 08:11:14.793643 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bbac0dc6-01b5-4211-94b0-9d5bcebd9cc1-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "bbac0dc6-01b5-4211-94b0-9d5bcebd9cc1" (UID: "bbac0dc6-01b5-4211-94b0-9d5bcebd9cc1"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:11:14 crc kubenswrapper[4612]: I0227 08:11:14.830999 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bbac0dc6-01b5-4211-94b0-9d5bcebd9cc1-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "bbac0dc6-01b5-4211-94b0-9d5bcebd9cc1" (UID: "bbac0dc6-01b5-4211-94b0-9d5bcebd9cc1"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:11:14 crc kubenswrapper[4612]: I0227 08:11:14.846792 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bbac0dc6-01b5-4211-94b0-9d5bcebd9cc1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bbac0dc6-01b5-4211-94b0-9d5bcebd9cc1" (UID: "bbac0dc6-01b5-4211-94b0-9d5bcebd9cc1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:11:14 crc kubenswrapper[4612]: I0227 08:11:14.872632 4612 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bbac0dc6-01b5-4211-94b0-9d5bcebd9cc1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 08:11:14 crc kubenswrapper[4612]: I0227 08:11:14.873618 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ll4pt\" (UniqueName: \"kubernetes.io/projected/bbac0dc6-01b5-4211-94b0-9d5bcebd9cc1-kube-api-access-ll4pt\") on node \"crc\" DevicePath \"\"" Feb 27 08:11:14 crc kubenswrapper[4612]: I0227 08:11:14.873631 4612 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/bbac0dc6-01b5-4211-94b0-9d5bcebd9cc1-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 27 08:11:14 crc kubenswrapper[4612]: I0227 08:11:14.873640 4612 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bbac0dc6-01b5-4211-94b0-9d5bcebd9cc1-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Feb 27 08:11:14 crc kubenswrapper[4612]: I0227 08:11:14.873650 4612 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bbac0dc6-01b5-4211-94b0-9d5bcebd9cc1-scripts\") on node \"crc\" DevicePath \"\"" Feb 27 08:11:14 crc kubenswrapper[4612]: I0227 08:11:14.900781 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bbac0dc6-01b5-4211-94b0-9d5bcebd9cc1-config-data" (OuterVolumeSpecName: "config-data") pod "bbac0dc6-01b5-4211-94b0-9d5bcebd9cc1" (UID: "bbac0dc6-01b5-4211-94b0-9d5bcebd9cc1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:11:14 crc kubenswrapper[4612]: I0227 08:11:14.911244 4612 generic.go:334] "Generic (PLEG): container finished" podID="bbac0dc6-01b5-4211-94b0-9d5bcebd9cc1" containerID="b66a95e205ddfd70faa5b402d5655ad70db606ef3ee22f1865fffe56be8ed95a" exitCode=0 Feb 27 08:11:14 crc kubenswrapper[4612]: I0227 08:11:14.911382 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 27 08:11:14 crc kubenswrapper[4612]: I0227 08:11:14.944542 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-z67kr" podStartSLOduration=2.787828049 podStartE2EDuration="13.944524068s" podCreationTimestamp="2026-02-27 08:11:01 +0000 UTC" firstStartedPulling="2026-02-27 08:11:02.719297263 +0000 UTC m=+1320.573227251" lastFinishedPulling="2026-02-27 08:11:13.875993272 +0000 UTC m=+1331.729923270" observedRunningTime="2026-02-27 08:11:14.932991347 +0000 UTC m=+1332.786921345" watchObservedRunningTime="2026-02-27 08:11:14.944524068 +0000 UTC m=+1332.798454066" Feb 27 08:11:14 crc kubenswrapper[4612]: I0227 08:11:14.975162 4612 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bbac0dc6-01b5-4211-94b0-9d5bcebd9cc1-config-data\") on node \"crc\" DevicePath \"\"" Feb 27 08:11:14 crc kubenswrapper[4612]: I0227 08:11:14.994162 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bbac0dc6-01b5-4211-94b0-9d5bcebd9cc1","Type":"ContainerDied","Data":"b66a95e205ddfd70faa5b402d5655ad70db606ef3ee22f1865fffe56be8ed95a"} Feb 27 08:11:14 crc kubenswrapper[4612]: I0227 08:11:14.994207 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bbac0dc6-01b5-4211-94b0-9d5bcebd9cc1","Type":"ContainerDied","Data":"9f1dd0ccd92fc13b12c7659b8fa079855a4374d0efb72f04382e46676c35f730"} Feb 27 08:11:14 crc kubenswrapper[4612]: I0227 08:11:14.994218 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-z67kr" event={"ID":"3d52b072-aded-480f-ae46-d89c9be0c9f8","Type":"ContainerStarted","Data":"73e8d5c17c8a1763106d32f761ba1f2030f5bf13fe92371e47c559d774a2b505"} Feb 27 08:11:14 crc kubenswrapper[4612]: I0227 08:11:14.994246 4612 scope.go:117] "RemoveContainer" containerID="2d06ff36815d20fb5029475648cc84007b2ac9385d41e9a28806917db2bd7829" Feb 27 08:11:15 crc kubenswrapper[4612]: I0227 08:11:15.020019 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 27 08:11:15 crc kubenswrapper[4612]: I0227 08:11:15.021095 4612 scope.go:117] "RemoveContainer" containerID="ba79e5fe70c18c20497563a68988c8690081e0f0ba75488b280b88ca5b83d16e" Feb 27 08:11:15 crc kubenswrapper[4612]: I0227 08:11:15.041196 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Feb 27 08:11:15 crc kubenswrapper[4612]: I0227 08:11:15.053983 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Feb 27 08:11:15 crc kubenswrapper[4612]: E0227 08:11:15.054440 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bbac0dc6-01b5-4211-94b0-9d5bcebd9cc1" containerName="ceilometer-notification-agent" Feb 27 08:11:15 crc kubenswrapper[4612]: I0227 08:11:15.054457 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="bbac0dc6-01b5-4211-94b0-9d5bcebd9cc1" containerName="ceilometer-notification-agent" Feb 27 08:11:15 crc kubenswrapper[4612]: E0227 08:11:15.054475 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bbac0dc6-01b5-4211-94b0-9d5bcebd9cc1" containerName="ceilometer-central-agent" Feb 27 08:11:15 crc kubenswrapper[4612]: I0227 08:11:15.054482 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="bbac0dc6-01b5-4211-94b0-9d5bcebd9cc1" containerName="ceilometer-central-agent" Feb 27 08:11:15 crc kubenswrapper[4612]: E0227 08:11:15.054500 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bbac0dc6-01b5-4211-94b0-9d5bcebd9cc1" containerName="proxy-httpd" Feb 27 08:11:15 crc kubenswrapper[4612]: I0227 08:11:15.054506 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="bbac0dc6-01b5-4211-94b0-9d5bcebd9cc1" containerName="proxy-httpd" Feb 27 08:11:15 crc kubenswrapper[4612]: E0227 08:11:15.054520 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bbac0dc6-01b5-4211-94b0-9d5bcebd9cc1" containerName="sg-core" Feb 27 08:11:15 crc kubenswrapper[4612]: I0227 08:11:15.054528 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="bbac0dc6-01b5-4211-94b0-9d5bcebd9cc1" containerName="sg-core" Feb 27 08:11:15 crc kubenswrapper[4612]: I0227 08:11:15.054728 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="bbac0dc6-01b5-4211-94b0-9d5bcebd9cc1" containerName="ceilometer-notification-agent" Feb 27 08:11:15 crc kubenswrapper[4612]: I0227 08:11:15.054748 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="bbac0dc6-01b5-4211-94b0-9d5bcebd9cc1" containerName="proxy-httpd" Feb 27 08:11:15 crc kubenswrapper[4612]: I0227 08:11:15.054761 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="bbac0dc6-01b5-4211-94b0-9d5bcebd9cc1" containerName="ceilometer-central-agent" Feb 27 08:11:15 crc kubenswrapper[4612]: I0227 08:11:15.054774 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="bbac0dc6-01b5-4211-94b0-9d5bcebd9cc1" containerName="sg-core" Feb 27 08:11:15 crc kubenswrapper[4612]: I0227 08:11:15.055764 4612 scope.go:117] "RemoveContainer" containerID="6b62d31b4a918b5c83e01dcd5b3b469fd7cc15867557f458bb2e371820cd2c57" Feb 27 08:11:15 crc kubenswrapper[4612]: I0227 08:11:15.056463 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 27 08:11:15 crc kubenswrapper[4612]: I0227 08:11:15.060920 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Feb 27 08:11:15 crc kubenswrapper[4612]: I0227 08:11:15.061090 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Feb 27 08:11:15 crc kubenswrapper[4612]: I0227 08:11:15.061219 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Feb 27 08:11:15 crc kubenswrapper[4612]: I0227 08:11:15.082514 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 27 08:11:15 crc kubenswrapper[4612]: I0227 08:11:15.105304 4612 scope.go:117] "RemoveContainer" containerID="b66a95e205ddfd70faa5b402d5655ad70db606ef3ee22f1865fffe56be8ed95a" Feb 27 08:11:15 crc kubenswrapper[4612]: I0227 08:11:15.124924 4612 scope.go:117] "RemoveContainer" containerID="2d06ff36815d20fb5029475648cc84007b2ac9385d41e9a28806917db2bd7829" Feb 27 08:11:15 crc kubenswrapper[4612]: E0227 08:11:15.125570 4612 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2d06ff36815d20fb5029475648cc84007b2ac9385d41e9a28806917db2bd7829\": container with ID starting with 2d06ff36815d20fb5029475648cc84007b2ac9385d41e9a28806917db2bd7829 not found: ID does not exist" containerID="2d06ff36815d20fb5029475648cc84007b2ac9385d41e9a28806917db2bd7829" Feb 27 08:11:15 crc kubenswrapper[4612]: I0227 08:11:15.125600 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2d06ff36815d20fb5029475648cc84007b2ac9385d41e9a28806917db2bd7829"} err="failed to get container status \"2d06ff36815d20fb5029475648cc84007b2ac9385d41e9a28806917db2bd7829\": rpc error: code = NotFound desc = could not find container \"2d06ff36815d20fb5029475648cc84007b2ac9385d41e9a28806917db2bd7829\": container with ID starting with 2d06ff36815d20fb5029475648cc84007b2ac9385d41e9a28806917db2bd7829 not found: ID does not exist" Feb 27 08:11:15 crc kubenswrapper[4612]: I0227 08:11:15.125621 4612 scope.go:117] "RemoveContainer" containerID="ba79e5fe70c18c20497563a68988c8690081e0f0ba75488b280b88ca5b83d16e" Feb 27 08:11:15 crc kubenswrapper[4612]: E0227 08:11:15.126124 4612 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ba79e5fe70c18c20497563a68988c8690081e0f0ba75488b280b88ca5b83d16e\": container with ID starting with ba79e5fe70c18c20497563a68988c8690081e0f0ba75488b280b88ca5b83d16e not found: ID does not exist" containerID="ba79e5fe70c18c20497563a68988c8690081e0f0ba75488b280b88ca5b83d16e" Feb 27 08:11:15 crc kubenswrapper[4612]: I0227 08:11:15.126187 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ba79e5fe70c18c20497563a68988c8690081e0f0ba75488b280b88ca5b83d16e"} err="failed to get container status \"ba79e5fe70c18c20497563a68988c8690081e0f0ba75488b280b88ca5b83d16e\": rpc error: code = NotFound desc = could not find container \"ba79e5fe70c18c20497563a68988c8690081e0f0ba75488b280b88ca5b83d16e\": container with ID starting with ba79e5fe70c18c20497563a68988c8690081e0f0ba75488b280b88ca5b83d16e not found: ID does not exist" Feb 27 08:11:15 crc kubenswrapper[4612]: I0227 08:11:15.126208 4612 scope.go:117] "RemoveContainer" containerID="6b62d31b4a918b5c83e01dcd5b3b469fd7cc15867557f458bb2e371820cd2c57" Feb 27 08:11:15 crc kubenswrapper[4612]: E0227 08:11:15.126445 4612 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6b62d31b4a918b5c83e01dcd5b3b469fd7cc15867557f458bb2e371820cd2c57\": container with ID starting with 6b62d31b4a918b5c83e01dcd5b3b469fd7cc15867557f458bb2e371820cd2c57 not found: ID does not exist" containerID="6b62d31b4a918b5c83e01dcd5b3b469fd7cc15867557f458bb2e371820cd2c57" Feb 27 08:11:15 crc kubenswrapper[4612]: I0227 08:11:15.126470 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6b62d31b4a918b5c83e01dcd5b3b469fd7cc15867557f458bb2e371820cd2c57"} err="failed to get container status \"6b62d31b4a918b5c83e01dcd5b3b469fd7cc15867557f458bb2e371820cd2c57\": rpc error: code = NotFound desc = could not find container \"6b62d31b4a918b5c83e01dcd5b3b469fd7cc15867557f458bb2e371820cd2c57\": container with ID starting with 6b62d31b4a918b5c83e01dcd5b3b469fd7cc15867557f458bb2e371820cd2c57 not found: ID does not exist" Feb 27 08:11:15 crc kubenswrapper[4612]: I0227 08:11:15.126488 4612 scope.go:117] "RemoveContainer" containerID="b66a95e205ddfd70faa5b402d5655ad70db606ef3ee22f1865fffe56be8ed95a" Feb 27 08:11:15 crc kubenswrapper[4612]: E0227 08:11:15.126903 4612 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b66a95e205ddfd70faa5b402d5655ad70db606ef3ee22f1865fffe56be8ed95a\": container with ID starting with b66a95e205ddfd70faa5b402d5655ad70db606ef3ee22f1865fffe56be8ed95a not found: ID does not exist" containerID="b66a95e205ddfd70faa5b402d5655ad70db606ef3ee22f1865fffe56be8ed95a" Feb 27 08:11:15 crc kubenswrapper[4612]: I0227 08:11:15.126929 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b66a95e205ddfd70faa5b402d5655ad70db606ef3ee22f1865fffe56be8ed95a"} err="failed to get container status \"b66a95e205ddfd70faa5b402d5655ad70db606ef3ee22f1865fffe56be8ed95a\": rpc error: code = NotFound desc = could not find container \"b66a95e205ddfd70faa5b402d5655ad70db606ef3ee22f1865fffe56be8ed95a\": container with ID starting with b66a95e205ddfd70faa5b402d5655ad70db606ef3ee22f1865fffe56be8ed95a not found: ID does not exist" Feb 27 08:11:15 crc kubenswrapper[4612]: I0227 08:11:15.177913 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/1f1e76b7-ca21-4a0b-ac75-0ff917106aa2-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"1f1e76b7-ca21-4a0b-ac75-0ff917106aa2\") " pod="openstack/ceilometer-0" Feb 27 08:11:15 crc kubenswrapper[4612]: I0227 08:11:15.177974 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1f1e76b7-ca21-4a0b-ac75-0ff917106aa2-run-httpd\") pod \"ceilometer-0\" (UID: \"1f1e76b7-ca21-4a0b-ac75-0ff917106aa2\") " pod="openstack/ceilometer-0" Feb 27 08:11:15 crc kubenswrapper[4612]: I0227 08:11:15.177997 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1f1e76b7-ca21-4a0b-ac75-0ff917106aa2-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"1f1e76b7-ca21-4a0b-ac75-0ff917106aa2\") " pod="openstack/ceilometer-0" Feb 27 08:11:15 crc kubenswrapper[4612]: I0227 08:11:15.178190 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-br4rw\" (UniqueName: \"kubernetes.io/projected/1f1e76b7-ca21-4a0b-ac75-0ff917106aa2-kube-api-access-br4rw\") pod \"ceilometer-0\" (UID: \"1f1e76b7-ca21-4a0b-ac75-0ff917106aa2\") " pod="openstack/ceilometer-0" Feb 27 08:11:15 crc kubenswrapper[4612]: I0227 08:11:15.178243 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1f1e76b7-ca21-4a0b-ac75-0ff917106aa2-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"1f1e76b7-ca21-4a0b-ac75-0ff917106aa2\") " pod="openstack/ceilometer-0" Feb 27 08:11:15 crc kubenswrapper[4612]: I0227 08:11:15.178285 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1f1e76b7-ca21-4a0b-ac75-0ff917106aa2-scripts\") pod \"ceilometer-0\" (UID: \"1f1e76b7-ca21-4a0b-ac75-0ff917106aa2\") " pod="openstack/ceilometer-0" Feb 27 08:11:15 crc kubenswrapper[4612]: I0227 08:11:15.178309 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1f1e76b7-ca21-4a0b-ac75-0ff917106aa2-log-httpd\") pod \"ceilometer-0\" (UID: \"1f1e76b7-ca21-4a0b-ac75-0ff917106aa2\") " pod="openstack/ceilometer-0" Feb 27 08:11:15 crc kubenswrapper[4612]: I0227 08:11:15.178347 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1f1e76b7-ca21-4a0b-ac75-0ff917106aa2-config-data\") pod \"ceilometer-0\" (UID: \"1f1e76b7-ca21-4a0b-ac75-0ff917106aa2\") " pod="openstack/ceilometer-0" Feb 27 08:11:15 crc kubenswrapper[4612]: I0227 08:11:15.280077 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-br4rw\" (UniqueName: \"kubernetes.io/projected/1f1e76b7-ca21-4a0b-ac75-0ff917106aa2-kube-api-access-br4rw\") pod \"ceilometer-0\" (UID: \"1f1e76b7-ca21-4a0b-ac75-0ff917106aa2\") " pod="openstack/ceilometer-0" Feb 27 08:11:15 crc kubenswrapper[4612]: I0227 08:11:15.280354 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1f1e76b7-ca21-4a0b-ac75-0ff917106aa2-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"1f1e76b7-ca21-4a0b-ac75-0ff917106aa2\") " pod="openstack/ceilometer-0" Feb 27 08:11:15 crc kubenswrapper[4612]: I0227 08:11:15.280481 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1f1e76b7-ca21-4a0b-ac75-0ff917106aa2-scripts\") pod \"ceilometer-0\" (UID: \"1f1e76b7-ca21-4a0b-ac75-0ff917106aa2\") " pod="openstack/ceilometer-0" Feb 27 08:11:15 crc kubenswrapper[4612]: I0227 08:11:15.280559 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1f1e76b7-ca21-4a0b-ac75-0ff917106aa2-log-httpd\") pod \"ceilometer-0\" (UID: \"1f1e76b7-ca21-4a0b-ac75-0ff917106aa2\") " pod="openstack/ceilometer-0" Feb 27 08:11:15 crc kubenswrapper[4612]: I0227 08:11:15.280649 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1f1e76b7-ca21-4a0b-ac75-0ff917106aa2-config-data\") pod \"ceilometer-0\" (UID: \"1f1e76b7-ca21-4a0b-ac75-0ff917106aa2\") " pod="openstack/ceilometer-0" Feb 27 08:11:15 crc kubenswrapper[4612]: I0227 08:11:15.280766 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/1f1e76b7-ca21-4a0b-ac75-0ff917106aa2-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"1f1e76b7-ca21-4a0b-ac75-0ff917106aa2\") " pod="openstack/ceilometer-0" Feb 27 08:11:15 crc kubenswrapper[4612]: I0227 08:11:15.280849 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1f1e76b7-ca21-4a0b-ac75-0ff917106aa2-run-httpd\") pod \"ceilometer-0\" (UID: \"1f1e76b7-ca21-4a0b-ac75-0ff917106aa2\") " pod="openstack/ceilometer-0" Feb 27 08:11:15 crc kubenswrapper[4612]: I0227 08:11:15.280926 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1f1e76b7-ca21-4a0b-ac75-0ff917106aa2-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"1f1e76b7-ca21-4a0b-ac75-0ff917106aa2\") " pod="openstack/ceilometer-0" Feb 27 08:11:15 crc kubenswrapper[4612]: I0227 08:11:15.282441 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1f1e76b7-ca21-4a0b-ac75-0ff917106aa2-log-httpd\") pod \"ceilometer-0\" (UID: \"1f1e76b7-ca21-4a0b-ac75-0ff917106aa2\") " pod="openstack/ceilometer-0" Feb 27 08:11:15 crc kubenswrapper[4612]: I0227 08:11:15.282488 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1f1e76b7-ca21-4a0b-ac75-0ff917106aa2-run-httpd\") pod \"ceilometer-0\" (UID: \"1f1e76b7-ca21-4a0b-ac75-0ff917106aa2\") " pod="openstack/ceilometer-0" Feb 27 08:11:15 crc kubenswrapper[4612]: I0227 08:11:15.286296 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1f1e76b7-ca21-4a0b-ac75-0ff917106aa2-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"1f1e76b7-ca21-4a0b-ac75-0ff917106aa2\") " pod="openstack/ceilometer-0" Feb 27 08:11:15 crc kubenswrapper[4612]: I0227 08:11:15.296058 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1f1e76b7-ca21-4a0b-ac75-0ff917106aa2-config-data\") pod \"ceilometer-0\" (UID: \"1f1e76b7-ca21-4a0b-ac75-0ff917106aa2\") " pod="openstack/ceilometer-0" Feb 27 08:11:15 crc kubenswrapper[4612]: I0227 08:11:15.296760 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/1f1e76b7-ca21-4a0b-ac75-0ff917106aa2-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"1f1e76b7-ca21-4a0b-ac75-0ff917106aa2\") " pod="openstack/ceilometer-0" Feb 27 08:11:15 crc kubenswrapper[4612]: I0227 08:11:15.297907 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1f1e76b7-ca21-4a0b-ac75-0ff917106aa2-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"1f1e76b7-ca21-4a0b-ac75-0ff917106aa2\") " pod="openstack/ceilometer-0" Feb 27 08:11:15 crc kubenswrapper[4612]: I0227 08:11:15.298478 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1f1e76b7-ca21-4a0b-ac75-0ff917106aa2-scripts\") pod \"ceilometer-0\" (UID: \"1f1e76b7-ca21-4a0b-ac75-0ff917106aa2\") " pod="openstack/ceilometer-0" Feb 27 08:11:15 crc kubenswrapper[4612]: I0227 08:11:15.304357 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-br4rw\" (UniqueName: \"kubernetes.io/projected/1f1e76b7-ca21-4a0b-ac75-0ff917106aa2-kube-api-access-br4rw\") pod \"ceilometer-0\" (UID: \"1f1e76b7-ca21-4a0b-ac75-0ff917106aa2\") " pod="openstack/ceilometer-0" Feb 27 08:11:15 crc kubenswrapper[4612]: I0227 08:11:15.375999 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 27 08:11:15 crc kubenswrapper[4612]: I0227 08:11:15.934763 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 27 08:11:16 crc kubenswrapper[4612]: I0227 08:11:16.027158 4612 patch_prober.go:28] interesting pod/machine-config-daemon-924vb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 27 08:11:16 crc kubenswrapper[4612]: I0227 08:11:16.027216 4612 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 27 08:11:16 crc kubenswrapper[4612]: I0227 08:11:16.862124 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bbac0dc6-01b5-4211-94b0-9d5bcebd9cc1" path="/var/lib/kubelet/pods/bbac0dc6-01b5-4211-94b0-9d5bcebd9cc1/volumes" Feb 27 08:11:16 crc kubenswrapper[4612]: I0227 08:11:16.933708 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1f1e76b7-ca21-4a0b-ac75-0ff917106aa2","Type":"ContainerStarted","Data":"540c359bbafca25714b4f9d4d4571dcf4ca593ccb6e817557d126798c24dea03"} Feb 27 08:11:16 crc kubenswrapper[4612]: I0227 08:11:16.933757 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1f1e76b7-ca21-4a0b-ac75-0ff917106aa2","Type":"ContainerStarted","Data":"2394db91ebe749930bf8729ac23fa1fd1853550f6085d41b3c3943152c74f190"} Feb 27 08:11:17 crc kubenswrapper[4612]: I0227 08:11:17.491360 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 27 08:11:17 crc kubenswrapper[4612]: I0227 08:11:17.492620 4612 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="16f91699-dd9d-4e71-8ff2-2bb767b3655c" containerName="glance-log" containerID="cri-o://6539429dfe0f2fb854db7c705771db9d5e7a1a793224317e2b3dee4af19646b5" gracePeriod=30 Feb 27 08:11:17 crc kubenswrapper[4612]: I0227 08:11:17.492727 4612 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="16f91699-dd9d-4e71-8ff2-2bb767b3655c" containerName="glance-httpd" containerID="cri-o://ddc49c3fd91b29cf3ba40b6f8e5717f9961e8e1840b755bb2cd2acf23ab56475" gracePeriod=30 Feb 27 08:11:17 crc kubenswrapper[4612]: I0227 08:11:17.942272 4612 generic.go:334] "Generic (PLEG): container finished" podID="16f91699-dd9d-4e71-8ff2-2bb767b3655c" containerID="6539429dfe0f2fb854db7c705771db9d5e7a1a793224317e2b3dee4af19646b5" exitCode=143 Feb 27 08:11:17 crc kubenswrapper[4612]: I0227 08:11:17.942365 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"16f91699-dd9d-4e71-8ff2-2bb767b3655c","Type":"ContainerDied","Data":"6539429dfe0f2fb854db7c705771db9d5e7a1a793224317e2b3dee4af19646b5"} Feb 27 08:11:17 crc kubenswrapper[4612]: I0227 08:11:17.944645 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1f1e76b7-ca21-4a0b-ac75-0ff917106aa2","Type":"ContainerStarted","Data":"27daefaeecf254b2c7057d0bdc634e5c6eac899815b470415df04a22acaf432d"} Feb 27 08:11:18 crc kubenswrapper[4612]: I0227 08:11:18.955209 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1f1e76b7-ca21-4a0b-ac75-0ff917106aa2","Type":"ContainerStarted","Data":"affd5d783f5ae025ebd8c67d5813022c368a25b81b112427191ecceb399717e9"} Feb 27 08:11:19 crc kubenswrapper[4612]: I0227 08:11:19.049054 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 27 08:11:19 crc kubenswrapper[4612]: I0227 08:11:19.049276 4612 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="f3bb8dc2-0c27-4a7a-baa1-1bfd8916f920" containerName="glance-log" containerID="cri-o://5238c7525d8aa4bdbb238a8f27a1c89df731af309999f71d7cc75733c681e871" gracePeriod=30 Feb 27 08:11:19 crc kubenswrapper[4612]: I0227 08:11:19.049392 4612 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="f3bb8dc2-0c27-4a7a-baa1-1bfd8916f920" containerName="glance-httpd" containerID="cri-o://c698bb30368f5b284e472e8eeb1597a14d477233b727140ada2c89a5d8044522" gracePeriod=30 Feb 27 08:11:19 crc kubenswrapper[4612]: I0227 08:11:19.970497 4612 generic.go:334] "Generic (PLEG): container finished" podID="f3bb8dc2-0c27-4a7a-baa1-1bfd8916f920" containerID="5238c7525d8aa4bdbb238a8f27a1c89df731af309999f71d7cc75733c681e871" exitCode=143 Feb 27 08:11:19 crc kubenswrapper[4612]: I0227 08:11:19.970781 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f3bb8dc2-0c27-4a7a-baa1-1bfd8916f920","Type":"ContainerDied","Data":"5238c7525d8aa4bdbb238a8f27a1c89df731af309999f71d7cc75733c681e871"} Feb 27 08:11:20 crc kubenswrapper[4612]: I0227 08:11:20.811710 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 27 08:11:20 crc kubenswrapper[4612]: I0227 08:11:20.980622 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1f1e76b7-ca21-4a0b-ac75-0ff917106aa2","Type":"ContainerStarted","Data":"32e856f1f5d26d4f0b991fa4f1943f3de4404572c4bd114be26e59f84691c594"} Feb 27 08:11:21 crc kubenswrapper[4612]: I0227 08:11:21.008293 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.04360056 podStartE2EDuration="6.008273478s" podCreationTimestamp="2026-02-27 08:11:15 +0000 UTC" firstStartedPulling="2026-02-27 08:11:15.940560502 +0000 UTC m=+1333.794490500" lastFinishedPulling="2026-02-27 08:11:19.90523342 +0000 UTC m=+1337.759163418" observedRunningTime="2026-02-27 08:11:21.001970016 +0000 UTC m=+1338.855900014" watchObservedRunningTime="2026-02-27 08:11:21.008273478 +0000 UTC m=+1338.862203476" Feb 27 08:11:21 crc kubenswrapper[4612]: I0227 08:11:21.987883 4612 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="1f1e76b7-ca21-4a0b-ac75-0ff917106aa2" containerName="ceilometer-central-agent" containerID="cri-o://540c359bbafca25714b4f9d4d4571dcf4ca593ccb6e817557d126798c24dea03" gracePeriod=30 Feb 27 08:11:21 crc kubenswrapper[4612]: I0227 08:11:21.988171 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Feb 27 08:11:21 crc kubenswrapper[4612]: I0227 08:11:21.988418 4612 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="1f1e76b7-ca21-4a0b-ac75-0ff917106aa2" containerName="proxy-httpd" containerID="cri-o://32e856f1f5d26d4f0b991fa4f1943f3de4404572c4bd114be26e59f84691c594" gracePeriod=30 Feb 27 08:11:21 crc kubenswrapper[4612]: I0227 08:11:21.988467 4612 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="1f1e76b7-ca21-4a0b-ac75-0ff917106aa2" containerName="sg-core" containerID="cri-o://affd5d783f5ae025ebd8c67d5813022c368a25b81b112427191ecceb399717e9" gracePeriod=30 Feb 27 08:11:21 crc kubenswrapper[4612]: I0227 08:11:21.988500 4612 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="1f1e76b7-ca21-4a0b-ac75-0ff917106aa2" containerName="ceilometer-notification-agent" containerID="cri-o://27daefaeecf254b2c7057d0bdc634e5c6eac899815b470415df04a22acaf432d" gracePeriod=30 Feb 27 08:11:22 crc kubenswrapper[4612]: I0227 08:11:22.134229 4612 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-6b985547d4-mkfpv" Feb 27 08:11:22 crc kubenswrapper[4612]: I0227 08:11:22.152955 4612 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-77948db5bb-97qh6" Feb 27 08:11:22 crc kubenswrapper[4612]: I0227 08:11:22.477599 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 27 08:11:22 crc kubenswrapper[4612]: I0227 08:11:22.652282 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lbkk7\" (UniqueName: \"kubernetes.io/projected/16f91699-dd9d-4e71-8ff2-2bb767b3655c-kube-api-access-lbkk7\") pod \"16f91699-dd9d-4e71-8ff2-2bb767b3655c\" (UID: \"16f91699-dd9d-4e71-8ff2-2bb767b3655c\") " Feb 27 08:11:22 crc kubenswrapper[4612]: I0227 08:11:22.652323 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/16f91699-dd9d-4e71-8ff2-2bb767b3655c-config-data\") pod \"16f91699-dd9d-4e71-8ff2-2bb767b3655c\" (UID: \"16f91699-dd9d-4e71-8ff2-2bb767b3655c\") " Feb 27 08:11:22 crc kubenswrapper[4612]: I0227 08:11:22.652388 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/16f91699-dd9d-4e71-8ff2-2bb767b3655c-combined-ca-bundle\") pod \"16f91699-dd9d-4e71-8ff2-2bb767b3655c\" (UID: \"16f91699-dd9d-4e71-8ff2-2bb767b3655c\") " Feb 27 08:11:22 crc kubenswrapper[4612]: I0227 08:11:22.652417 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"16f91699-dd9d-4e71-8ff2-2bb767b3655c\" (UID: \"16f91699-dd9d-4e71-8ff2-2bb767b3655c\") " Feb 27 08:11:22 crc kubenswrapper[4612]: I0227 08:11:22.652437 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/16f91699-dd9d-4e71-8ff2-2bb767b3655c-scripts\") pod \"16f91699-dd9d-4e71-8ff2-2bb767b3655c\" (UID: \"16f91699-dd9d-4e71-8ff2-2bb767b3655c\") " Feb 27 08:11:22 crc kubenswrapper[4612]: I0227 08:11:22.652486 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/16f91699-dd9d-4e71-8ff2-2bb767b3655c-httpd-run\") pod \"16f91699-dd9d-4e71-8ff2-2bb767b3655c\" (UID: \"16f91699-dd9d-4e71-8ff2-2bb767b3655c\") " Feb 27 08:11:22 crc kubenswrapper[4612]: I0227 08:11:22.652503 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/16f91699-dd9d-4e71-8ff2-2bb767b3655c-public-tls-certs\") pod \"16f91699-dd9d-4e71-8ff2-2bb767b3655c\" (UID: \"16f91699-dd9d-4e71-8ff2-2bb767b3655c\") " Feb 27 08:11:22 crc kubenswrapper[4612]: I0227 08:11:22.652525 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/16f91699-dd9d-4e71-8ff2-2bb767b3655c-logs\") pod \"16f91699-dd9d-4e71-8ff2-2bb767b3655c\" (UID: \"16f91699-dd9d-4e71-8ff2-2bb767b3655c\") " Feb 27 08:11:22 crc kubenswrapper[4612]: I0227 08:11:22.652943 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/16f91699-dd9d-4e71-8ff2-2bb767b3655c-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "16f91699-dd9d-4e71-8ff2-2bb767b3655c" (UID: "16f91699-dd9d-4e71-8ff2-2bb767b3655c"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 08:11:22 crc kubenswrapper[4612]: I0227 08:11:22.653039 4612 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/16f91699-dd9d-4e71-8ff2-2bb767b3655c-httpd-run\") on node \"crc\" DevicePath \"\"" Feb 27 08:11:22 crc kubenswrapper[4612]: I0227 08:11:22.653245 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/16f91699-dd9d-4e71-8ff2-2bb767b3655c-logs" (OuterVolumeSpecName: "logs") pod "16f91699-dd9d-4e71-8ff2-2bb767b3655c" (UID: "16f91699-dd9d-4e71-8ff2-2bb767b3655c"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 08:11:22 crc kubenswrapper[4612]: I0227 08:11:22.679030 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/16f91699-dd9d-4e71-8ff2-2bb767b3655c-scripts" (OuterVolumeSpecName: "scripts") pod "16f91699-dd9d-4e71-8ff2-2bb767b3655c" (UID: "16f91699-dd9d-4e71-8ff2-2bb767b3655c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:11:22 crc kubenswrapper[4612]: I0227 08:11:22.687391 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/16f91699-dd9d-4e71-8ff2-2bb767b3655c-kube-api-access-lbkk7" (OuterVolumeSpecName: "kube-api-access-lbkk7") pod "16f91699-dd9d-4e71-8ff2-2bb767b3655c" (UID: "16f91699-dd9d-4e71-8ff2-2bb767b3655c"). InnerVolumeSpecName "kube-api-access-lbkk7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 08:11:22 crc kubenswrapper[4612]: I0227 08:11:22.696041 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage03-crc" (OuterVolumeSpecName: "glance") pod "16f91699-dd9d-4e71-8ff2-2bb767b3655c" (UID: "16f91699-dd9d-4e71-8ff2-2bb767b3655c"). InnerVolumeSpecName "local-storage03-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Feb 27 08:11:22 crc kubenswrapper[4612]: I0227 08:11:22.716951 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/16f91699-dd9d-4e71-8ff2-2bb767b3655c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "16f91699-dd9d-4e71-8ff2-2bb767b3655c" (UID: "16f91699-dd9d-4e71-8ff2-2bb767b3655c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:11:22 crc kubenswrapper[4612]: I0227 08:11:22.743297 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/16f91699-dd9d-4e71-8ff2-2bb767b3655c-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "16f91699-dd9d-4e71-8ff2-2bb767b3655c" (UID: "16f91699-dd9d-4e71-8ff2-2bb767b3655c"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:11:22 crc kubenswrapper[4612]: I0227 08:11:22.745903 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/16f91699-dd9d-4e71-8ff2-2bb767b3655c-config-data" (OuterVolumeSpecName: "config-data") pod "16f91699-dd9d-4e71-8ff2-2bb767b3655c" (UID: "16f91699-dd9d-4e71-8ff2-2bb767b3655c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:11:22 crc kubenswrapper[4612]: I0227 08:11:22.755112 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lbkk7\" (UniqueName: \"kubernetes.io/projected/16f91699-dd9d-4e71-8ff2-2bb767b3655c-kube-api-access-lbkk7\") on node \"crc\" DevicePath \"\"" Feb 27 08:11:22 crc kubenswrapper[4612]: I0227 08:11:22.755147 4612 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/16f91699-dd9d-4e71-8ff2-2bb767b3655c-config-data\") on node \"crc\" DevicePath \"\"" Feb 27 08:11:22 crc kubenswrapper[4612]: I0227 08:11:22.755159 4612 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/16f91699-dd9d-4e71-8ff2-2bb767b3655c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 08:11:22 crc kubenswrapper[4612]: I0227 08:11:22.755186 4612 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" " Feb 27 08:11:22 crc kubenswrapper[4612]: I0227 08:11:22.755196 4612 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/16f91699-dd9d-4e71-8ff2-2bb767b3655c-scripts\") on node \"crc\" DevicePath \"\"" Feb 27 08:11:22 crc kubenswrapper[4612]: I0227 08:11:22.755204 4612 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/16f91699-dd9d-4e71-8ff2-2bb767b3655c-public-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 27 08:11:22 crc kubenswrapper[4612]: I0227 08:11:22.755213 4612 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/16f91699-dd9d-4e71-8ff2-2bb767b3655c-logs\") on node \"crc\" DevicePath \"\"" Feb 27 08:11:22 crc kubenswrapper[4612]: I0227 08:11:22.789873 4612 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage03-crc" (UniqueName: "kubernetes.io/local-volume/local-storage03-crc") on node "crc" Feb 27 08:11:22 crc kubenswrapper[4612]: I0227 08:11:22.838256 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 27 08:11:22 crc kubenswrapper[4612]: I0227 08:11:22.859297 4612 reconciler_common.go:293] "Volume detached for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" DevicePath \"\"" Feb 27 08:11:22 crc kubenswrapper[4612]: I0227 08:11:22.959962 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f3bb8dc2-0c27-4a7a-baa1-1bfd8916f920-scripts\") pod \"f3bb8dc2-0c27-4a7a-baa1-1bfd8916f920\" (UID: \"f3bb8dc2-0c27-4a7a-baa1-1bfd8916f920\") " Feb 27 08:11:22 crc kubenswrapper[4612]: I0227 08:11:22.960067 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kqqmt\" (UniqueName: \"kubernetes.io/projected/f3bb8dc2-0c27-4a7a-baa1-1bfd8916f920-kube-api-access-kqqmt\") pod \"f3bb8dc2-0c27-4a7a-baa1-1bfd8916f920\" (UID: \"f3bb8dc2-0c27-4a7a-baa1-1bfd8916f920\") " Feb 27 08:11:22 crc kubenswrapper[4612]: I0227 08:11:22.960109 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f3bb8dc2-0c27-4a7a-baa1-1bfd8916f920-logs\") pod \"f3bb8dc2-0c27-4a7a-baa1-1bfd8916f920\" (UID: \"f3bb8dc2-0c27-4a7a-baa1-1bfd8916f920\") " Feb 27 08:11:22 crc kubenswrapper[4612]: I0227 08:11:22.960190 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f3bb8dc2-0c27-4a7a-baa1-1bfd8916f920-httpd-run\") pod \"f3bb8dc2-0c27-4a7a-baa1-1bfd8916f920\" (UID: \"f3bb8dc2-0c27-4a7a-baa1-1bfd8916f920\") " Feb 27 08:11:22 crc kubenswrapper[4612]: I0227 08:11:22.960223 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f3bb8dc2-0c27-4a7a-baa1-1bfd8916f920-config-data\") pod \"f3bb8dc2-0c27-4a7a-baa1-1bfd8916f920\" (UID: \"f3bb8dc2-0c27-4a7a-baa1-1bfd8916f920\") " Feb 27 08:11:22 crc kubenswrapper[4612]: I0227 08:11:22.960272 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f3bb8dc2-0c27-4a7a-baa1-1bfd8916f920-internal-tls-certs\") pod \"f3bb8dc2-0c27-4a7a-baa1-1bfd8916f920\" (UID: \"f3bb8dc2-0c27-4a7a-baa1-1bfd8916f920\") " Feb 27 08:11:22 crc kubenswrapper[4612]: I0227 08:11:22.960301 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3bb8dc2-0c27-4a7a-baa1-1bfd8916f920-combined-ca-bundle\") pod \"f3bb8dc2-0c27-4a7a-baa1-1bfd8916f920\" (UID: \"f3bb8dc2-0c27-4a7a-baa1-1bfd8916f920\") " Feb 27 08:11:22 crc kubenswrapper[4612]: I0227 08:11:22.960346 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"f3bb8dc2-0c27-4a7a-baa1-1bfd8916f920\" (UID: \"f3bb8dc2-0c27-4a7a-baa1-1bfd8916f920\") " Feb 27 08:11:22 crc kubenswrapper[4612]: I0227 08:11:22.961654 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f3bb8dc2-0c27-4a7a-baa1-1bfd8916f920-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "f3bb8dc2-0c27-4a7a-baa1-1bfd8916f920" (UID: "f3bb8dc2-0c27-4a7a-baa1-1bfd8916f920"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 08:11:22 crc kubenswrapper[4612]: I0227 08:11:22.962940 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f3bb8dc2-0c27-4a7a-baa1-1bfd8916f920-logs" (OuterVolumeSpecName: "logs") pod "f3bb8dc2-0c27-4a7a-baa1-1bfd8916f920" (UID: "f3bb8dc2-0c27-4a7a-baa1-1bfd8916f920"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 08:11:22 crc kubenswrapper[4612]: I0227 08:11:22.969487 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f3bb8dc2-0c27-4a7a-baa1-1bfd8916f920-scripts" (OuterVolumeSpecName: "scripts") pod "f3bb8dc2-0c27-4a7a-baa1-1bfd8916f920" (UID: "f3bb8dc2-0c27-4a7a-baa1-1bfd8916f920"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:11:22 crc kubenswrapper[4612]: I0227 08:11:22.973072 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f3bb8dc2-0c27-4a7a-baa1-1bfd8916f920-kube-api-access-kqqmt" (OuterVolumeSpecName: "kube-api-access-kqqmt") pod "f3bb8dc2-0c27-4a7a-baa1-1bfd8916f920" (UID: "f3bb8dc2-0c27-4a7a-baa1-1bfd8916f920"). InnerVolumeSpecName "kube-api-access-kqqmt". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 08:11:22 crc kubenswrapper[4612]: I0227 08:11:22.973229 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage04-crc" (OuterVolumeSpecName: "glance") pod "f3bb8dc2-0c27-4a7a-baa1-1bfd8916f920" (UID: "f3bb8dc2-0c27-4a7a-baa1-1bfd8916f920"). InnerVolumeSpecName "local-storage04-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Feb 27 08:11:23 crc kubenswrapper[4612]: I0227 08:11:23.001572 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f3bb8dc2-0c27-4a7a-baa1-1bfd8916f920-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f3bb8dc2-0c27-4a7a-baa1-1bfd8916f920" (UID: "f3bb8dc2-0c27-4a7a-baa1-1bfd8916f920"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:11:23 crc kubenswrapper[4612]: I0227 08:11:23.002822 4612 generic.go:334] "Generic (PLEG): container finished" podID="16f91699-dd9d-4e71-8ff2-2bb767b3655c" containerID="ddc49c3fd91b29cf3ba40b6f8e5717f9961e8e1840b755bb2cd2acf23ab56475" exitCode=0 Feb 27 08:11:23 crc kubenswrapper[4612]: I0227 08:11:23.002883 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"16f91699-dd9d-4e71-8ff2-2bb767b3655c","Type":"ContainerDied","Data":"ddc49c3fd91b29cf3ba40b6f8e5717f9961e8e1840b755bb2cd2acf23ab56475"} Feb 27 08:11:23 crc kubenswrapper[4612]: I0227 08:11:23.002912 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"16f91699-dd9d-4e71-8ff2-2bb767b3655c","Type":"ContainerDied","Data":"53130813b4018d05f1ac9e30b5ca1a3ee60f45537c457fde1fe6161d6a333c54"} Feb 27 08:11:23 crc kubenswrapper[4612]: I0227 08:11:23.002928 4612 scope.go:117] "RemoveContainer" containerID="ddc49c3fd91b29cf3ba40b6f8e5717f9961e8e1840b755bb2cd2acf23ab56475" Feb 27 08:11:23 crc kubenswrapper[4612]: I0227 08:11:23.003046 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 27 08:11:23 crc kubenswrapper[4612]: I0227 08:11:23.010560 4612 generic.go:334] "Generic (PLEG): container finished" podID="f3bb8dc2-0c27-4a7a-baa1-1bfd8916f920" containerID="c698bb30368f5b284e472e8eeb1597a14d477233b727140ada2c89a5d8044522" exitCode=0 Feb 27 08:11:23 crc kubenswrapper[4612]: I0227 08:11:23.010610 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f3bb8dc2-0c27-4a7a-baa1-1bfd8916f920","Type":"ContainerDied","Data":"c698bb30368f5b284e472e8eeb1597a14d477233b727140ada2c89a5d8044522"} Feb 27 08:11:23 crc kubenswrapper[4612]: I0227 08:11:23.010634 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f3bb8dc2-0c27-4a7a-baa1-1bfd8916f920","Type":"ContainerDied","Data":"061c95d3248b23c6fb1255c379f096d63ba92d34e7ca9cb254e6fffd9b2e9904"} Feb 27 08:11:23 crc kubenswrapper[4612]: I0227 08:11:23.010740 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 27 08:11:23 crc kubenswrapper[4612]: I0227 08:11:23.030039 4612 generic.go:334] "Generic (PLEG): container finished" podID="1f1e76b7-ca21-4a0b-ac75-0ff917106aa2" containerID="32e856f1f5d26d4f0b991fa4f1943f3de4404572c4bd114be26e59f84691c594" exitCode=0 Feb 27 08:11:23 crc kubenswrapper[4612]: I0227 08:11:23.030758 4612 generic.go:334] "Generic (PLEG): container finished" podID="1f1e76b7-ca21-4a0b-ac75-0ff917106aa2" containerID="affd5d783f5ae025ebd8c67d5813022c368a25b81b112427191ecceb399717e9" exitCode=2 Feb 27 08:11:23 crc kubenswrapper[4612]: I0227 08:11:23.030860 4612 generic.go:334] "Generic (PLEG): container finished" podID="1f1e76b7-ca21-4a0b-ac75-0ff917106aa2" containerID="27daefaeecf254b2c7057d0bdc634e5c6eac899815b470415df04a22acaf432d" exitCode=0 Feb 27 08:11:23 crc kubenswrapper[4612]: I0227 08:11:23.031035 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1f1e76b7-ca21-4a0b-ac75-0ff917106aa2","Type":"ContainerDied","Data":"32e856f1f5d26d4f0b991fa4f1943f3de4404572c4bd114be26e59f84691c594"} Feb 27 08:11:23 crc kubenswrapper[4612]: I0227 08:11:23.031180 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1f1e76b7-ca21-4a0b-ac75-0ff917106aa2","Type":"ContainerDied","Data":"affd5d783f5ae025ebd8c67d5813022c368a25b81b112427191ecceb399717e9"} Feb 27 08:11:23 crc kubenswrapper[4612]: I0227 08:11:23.031307 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1f1e76b7-ca21-4a0b-ac75-0ff917106aa2","Type":"ContainerDied","Data":"27daefaeecf254b2c7057d0bdc634e5c6eac899815b470415df04a22acaf432d"} Feb 27 08:11:23 crc kubenswrapper[4612]: I0227 08:11:23.043185 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f3bb8dc2-0c27-4a7a-baa1-1bfd8916f920-config-data" (OuterVolumeSpecName: "config-data") pod "f3bb8dc2-0c27-4a7a-baa1-1bfd8916f920" (UID: "f3bb8dc2-0c27-4a7a-baa1-1bfd8916f920"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:11:23 crc kubenswrapper[4612]: I0227 08:11:23.046328 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f3bb8dc2-0c27-4a7a-baa1-1bfd8916f920-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "f3bb8dc2-0c27-4a7a-baa1-1bfd8916f920" (UID: "f3bb8dc2-0c27-4a7a-baa1-1bfd8916f920"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:11:23 crc kubenswrapper[4612]: I0227 08:11:23.061925 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kqqmt\" (UniqueName: \"kubernetes.io/projected/f3bb8dc2-0c27-4a7a-baa1-1bfd8916f920-kube-api-access-kqqmt\") on node \"crc\" DevicePath \"\"" Feb 27 08:11:23 crc kubenswrapper[4612]: I0227 08:11:23.061947 4612 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f3bb8dc2-0c27-4a7a-baa1-1bfd8916f920-logs\") on node \"crc\" DevicePath \"\"" Feb 27 08:11:23 crc kubenswrapper[4612]: I0227 08:11:23.061956 4612 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f3bb8dc2-0c27-4a7a-baa1-1bfd8916f920-httpd-run\") on node \"crc\" DevicePath \"\"" Feb 27 08:11:23 crc kubenswrapper[4612]: I0227 08:11:23.061965 4612 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f3bb8dc2-0c27-4a7a-baa1-1bfd8916f920-config-data\") on node \"crc\" DevicePath \"\"" Feb 27 08:11:23 crc kubenswrapper[4612]: I0227 08:11:23.061974 4612 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f3bb8dc2-0c27-4a7a-baa1-1bfd8916f920-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 27 08:11:23 crc kubenswrapper[4612]: I0227 08:11:23.061987 4612 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3bb8dc2-0c27-4a7a-baa1-1bfd8916f920-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 08:11:23 crc kubenswrapper[4612]: I0227 08:11:23.062035 4612 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" " Feb 27 08:11:23 crc kubenswrapper[4612]: I0227 08:11:23.062048 4612 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f3bb8dc2-0c27-4a7a-baa1-1bfd8916f920-scripts\") on node \"crc\" DevicePath \"\"" Feb 27 08:11:23 crc kubenswrapper[4612]: I0227 08:11:23.097383 4612 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage04-crc" (UniqueName: "kubernetes.io/local-volume/local-storage04-crc") on node "crc" Feb 27 08:11:23 crc kubenswrapper[4612]: I0227 08:11:23.140347 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 27 08:11:23 crc kubenswrapper[4612]: I0227 08:11:23.152396 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 27 08:11:23 crc kubenswrapper[4612]: I0227 08:11:23.163326 4612 reconciler_common.go:293] "Volume detached for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" DevicePath \"\"" Feb 27 08:11:23 crc kubenswrapper[4612]: I0227 08:11:23.167409 4612 scope.go:117] "RemoveContainer" containerID="6539429dfe0f2fb854db7c705771db9d5e7a1a793224317e2b3dee4af19646b5" Feb 27 08:11:23 crc kubenswrapper[4612]: I0227 08:11:23.176163 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Feb 27 08:11:23 crc kubenswrapper[4612]: E0227 08:11:23.176508 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f3bb8dc2-0c27-4a7a-baa1-1bfd8916f920" containerName="glance-log" Feb 27 08:11:23 crc kubenswrapper[4612]: I0227 08:11:23.176526 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="f3bb8dc2-0c27-4a7a-baa1-1bfd8916f920" containerName="glance-log" Feb 27 08:11:23 crc kubenswrapper[4612]: E0227 08:11:23.176540 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f3bb8dc2-0c27-4a7a-baa1-1bfd8916f920" containerName="glance-httpd" Feb 27 08:11:23 crc kubenswrapper[4612]: I0227 08:11:23.176546 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="f3bb8dc2-0c27-4a7a-baa1-1bfd8916f920" containerName="glance-httpd" Feb 27 08:11:23 crc kubenswrapper[4612]: E0227 08:11:23.176558 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="16f91699-dd9d-4e71-8ff2-2bb767b3655c" containerName="glance-httpd" Feb 27 08:11:23 crc kubenswrapper[4612]: I0227 08:11:23.176563 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="16f91699-dd9d-4e71-8ff2-2bb767b3655c" containerName="glance-httpd" Feb 27 08:11:23 crc kubenswrapper[4612]: E0227 08:11:23.176591 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="16f91699-dd9d-4e71-8ff2-2bb767b3655c" containerName="glance-log" Feb 27 08:11:23 crc kubenswrapper[4612]: I0227 08:11:23.176597 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="16f91699-dd9d-4e71-8ff2-2bb767b3655c" containerName="glance-log" Feb 27 08:11:23 crc kubenswrapper[4612]: I0227 08:11:23.177278 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="16f91699-dd9d-4e71-8ff2-2bb767b3655c" containerName="glance-httpd" Feb 27 08:11:23 crc kubenswrapper[4612]: I0227 08:11:23.177299 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="f3bb8dc2-0c27-4a7a-baa1-1bfd8916f920" containerName="glance-log" Feb 27 08:11:23 crc kubenswrapper[4612]: I0227 08:11:23.177313 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="16f91699-dd9d-4e71-8ff2-2bb767b3655c" containerName="glance-log" Feb 27 08:11:23 crc kubenswrapper[4612]: I0227 08:11:23.177321 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="f3bb8dc2-0c27-4a7a-baa1-1bfd8916f920" containerName="glance-httpd" Feb 27 08:11:23 crc kubenswrapper[4612]: I0227 08:11:23.179104 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 27 08:11:23 crc kubenswrapper[4612]: I0227 08:11:23.182489 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Feb 27 08:11:23 crc kubenswrapper[4612]: I0227 08:11:23.190343 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Feb 27 08:11:23 crc kubenswrapper[4612]: I0227 08:11:23.234763 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 27 08:11:23 crc kubenswrapper[4612]: I0227 08:11:23.241055 4612 scope.go:117] "RemoveContainer" containerID="ddc49c3fd91b29cf3ba40b6f8e5717f9961e8e1840b755bb2cd2acf23ab56475" Feb 27 08:11:23 crc kubenswrapper[4612]: E0227 08:11:23.241614 4612 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ddc49c3fd91b29cf3ba40b6f8e5717f9961e8e1840b755bb2cd2acf23ab56475\": container with ID starting with ddc49c3fd91b29cf3ba40b6f8e5717f9961e8e1840b755bb2cd2acf23ab56475 not found: ID does not exist" containerID="ddc49c3fd91b29cf3ba40b6f8e5717f9961e8e1840b755bb2cd2acf23ab56475" Feb 27 08:11:23 crc kubenswrapper[4612]: I0227 08:11:23.241659 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ddc49c3fd91b29cf3ba40b6f8e5717f9961e8e1840b755bb2cd2acf23ab56475"} err="failed to get container status \"ddc49c3fd91b29cf3ba40b6f8e5717f9961e8e1840b755bb2cd2acf23ab56475\": rpc error: code = NotFound desc = could not find container \"ddc49c3fd91b29cf3ba40b6f8e5717f9961e8e1840b755bb2cd2acf23ab56475\": container with ID starting with ddc49c3fd91b29cf3ba40b6f8e5717f9961e8e1840b755bb2cd2acf23ab56475 not found: ID does not exist" Feb 27 08:11:23 crc kubenswrapper[4612]: I0227 08:11:23.241684 4612 scope.go:117] "RemoveContainer" containerID="6539429dfe0f2fb854db7c705771db9d5e7a1a793224317e2b3dee4af19646b5" Feb 27 08:11:23 crc kubenswrapper[4612]: E0227 08:11:23.244290 4612 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6539429dfe0f2fb854db7c705771db9d5e7a1a793224317e2b3dee4af19646b5\": container with ID starting with 6539429dfe0f2fb854db7c705771db9d5e7a1a793224317e2b3dee4af19646b5 not found: ID does not exist" containerID="6539429dfe0f2fb854db7c705771db9d5e7a1a793224317e2b3dee4af19646b5" Feb 27 08:11:23 crc kubenswrapper[4612]: I0227 08:11:23.244332 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6539429dfe0f2fb854db7c705771db9d5e7a1a793224317e2b3dee4af19646b5"} err="failed to get container status \"6539429dfe0f2fb854db7c705771db9d5e7a1a793224317e2b3dee4af19646b5\": rpc error: code = NotFound desc = could not find container \"6539429dfe0f2fb854db7c705771db9d5e7a1a793224317e2b3dee4af19646b5\": container with ID starting with 6539429dfe0f2fb854db7c705771db9d5e7a1a793224317e2b3dee4af19646b5 not found: ID does not exist" Feb 27 08:11:23 crc kubenswrapper[4612]: I0227 08:11:23.244359 4612 scope.go:117] "RemoveContainer" containerID="c698bb30368f5b284e472e8eeb1597a14d477233b727140ada2c89a5d8044522" Feb 27 08:11:23 crc kubenswrapper[4612]: I0227 08:11:23.264711 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/58945c32-1660-410d-9a49-b0a38829f907-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"58945c32-1660-410d-9a49-b0a38829f907\") " pod="openstack/glance-default-external-api-0" Feb 27 08:11:23 crc kubenswrapper[4612]: I0227 08:11:23.264757 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/58945c32-1660-410d-9a49-b0a38829f907-logs\") pod \"glance-default-external-api-0\" (UID: \"58945c32-1660-410d-9a49-b0a38829f907\") " pod="openstack/glance-default-external-api-0" Feb 27 08:11:23 crc kubenswrapper[4612]: I0227 08:11:23.264810 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58945c32-1660-410d-9a49-b0a38829f907-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"58945c32-1660-410d-9a49-b0a38829f907\") " pod="openstack/glance-default-external-api-0" Feb 27 08:11:23 crc kubenswrapper[4612]: I0227 08:11:23.264892 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jvzwp\" (UniqueName: \"kubernetes.io/projected/58945c32-1660-410d-9a49-b0a38829f907-kube-api-access-jvzwp\") pod \"glance-default-external-api-0\" (UID: \"58945c32-1660-410d-9a49-b0a38829f907\") " pod="openstack/glance-default-external-api-0" Feb 27 08:11:23 crc kubenswrapper[4612]: I0227 08:11:23.264917 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-external-api-0\" (UID: \"58945c32-1660-410d-9a49-b0a38829f907\") " pod="openstack/glance-default-external-api-0" Feb 27 08:11:23 crc kubenswrapper[4612]: I0227 08:11:23.264949 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/58945c32-1660-410d-9a49-b0a38829f907-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"58945c32-1660-410d-9a49-b0a38829f907\") " pod="openstack/glance-default-external-api-0" Feb 27 08:11:23 crc kubenswrapper[4612]: I0227 08:11:23.264974 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/58945c32-1660-410d-9a49-b0a38829f907-config-data\") pod \"glance-default-external-api-0\" (UID: \"58945c32-1660-410d-9a49-b0a38829f907\") " pod="openstack/glance-default-external-api-0" Feb 27 08:11:23 crc kubenswrapper[4612]: I0227 08:11:23.264999 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/58945c32-1660-410d-9a49-b0a38829f907-scripts\") pod \"glance-default-external-api-0\" (UID: \"58945c32-1660-410d-9a49-b0a38829f907\") " pod="openstack/glance-default-external-api-0" Feb 27 08:11:23 crc kubenswrapper[4612]: I0227 08:11:23.272063 4612 scope.go:117] "RemoveContainer" containerID="5238c7525d8aa4bdbb238a8f27a1c89df731af309999f71d7cc75733c681e871" Feb 27 08:11:23 crc kubenswrapper[4612]: I0227 08:11:23.290462 4612 scope.go:117] "RemoveContainer" containerID="c698bb30368f5b284e472e8eeb1597a14d477233b727140ada2c89a5d8044522" Feb 27 08:11:23 crc kubenswrapper[4612]: E0227 08:11:23.292014 4612 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c698bb30368f5b284e472e8eeb1597a14d477233b727140ada2c89a5d8044522\": container with ID starting with c698bb30368f5b284e472e8eeb1597a14d477233b727140ada2c89a5d8044522 not found: ID does not exist" containerID="c698bb30368f5b284e472e8eeb1597a14d477233b727140ada2c89a5d8044522" Feb 27 08:11:23 crc kubenswrapper[4612]: I0227 08:11:23.292062 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c698bb30368f5b284e472e8eeb1597a14d477233b727140ada2c89a5d8044522"} err="failed to get container status \"c698bb30368f5b284e472e8eeb1597a14d477233b727140ada2c89a5d8044522\": rpc error: code = NotFound desc = could not find container \"c698bb30368f5b284e472e8eeb1597a14d477233b727140ada2c89a5d8044522\": container with ID starting with c698bb30368f5b284e472e8eeb1597a14d477233b727140ada2c89a5d8044522 not found: ID does not exist" Feb 27 08:11:23 crc kubenswrapper[4612]: I0227 08:11:23.292089 4612 scope.go:117] "RemoveContainer" containerID="5238c7525d8aa4bdbb238a8f27a1c89df731af309999f71d7cc75733c681e871" Feb 27 08:11:23 crc kubenswrapper[4612]: E0227 08:11:23.292384 4612 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5238c7525d8aa4bdbb238a8f27a1c89df731af309999f71d7cc75733c681e871\": container with ID starting with 5238c7525d8aa4bdbb238a8f27a1c89df731af309999f71d7cc75733c681e871 not found: ID does not exist" containerID="5238c7525d8aa4bdbb238a8f27a1c89df731af309999f71d7cc75733c681e871" Feb 27 08:11:23 crc kubenswrapper[4612]: I0227 08:11:23.292423 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5238c7525d8aa4bdbb238a8f27a1c89df731af309999f71d7cc75733c681e871"} err="failed to get container status \"5238c7525d8aa4bdbb238a8f27a1c89df731af309999f71d7cc75733c681e871\": rpc error: code = NotFound desc = could not find container \"5238c7525d8aa4bdbb238a8f27a1c89df731af309999f71d7cc75733c681e871\": container with ID starting with 5238c7525d8aa4bdbb238a8f27a1c89df731af309999f71d7cc75733c681e871 not found: ID does not exist" Feb 27 08:11:23 crc kubenswrapper[4612]: I0227 08:11:23.349766 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 27 08:11:23 crc kubenswrapper[4612]: I0227 08:11:23.357503 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 27 08:11:23 crc kubenswrapper[4612]: I0227 08:11:23.370520 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/58945c32-1660-410d-9a49-b0a38829f907-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"58945c32-1660-410d-9a49-b0a38829f907\") " pod="openstack/glance-default-external-api-0" Feb 27 08:11:23 crc kubenswrapper[4612]: I0227 08:11:23.370794 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/58945c32-1660-410d-9a49-b0a38829f907-config-data\") pod \"glance-default-external-api-0\" (UID: \"58945c32-1660-410d-9a49-b0a38829f907\") " pod="openstack/glance-default-external-api-0" Feb 27 08:11:23 crc kubenswrapper[4612]: I0227 08:11:23.370860 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/58945c32-1660-410d-9a49-b0a38829f907-scripts\") pod \"glance-default-external-api-0\" (UID: \"58945c32-1660-410d-9a49-b0a38829f907\") " pod="openstack/glance-default-external-api-0" Feb 27 08:11:23 crc kubenswrapper[4612]: I0227 08:11:23.371036 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/58945c32-1660-410d-9a49-b0a38829f907-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"58945c32-1660-410d-9a49-b0a38829f907\") " pod="openstack/glance-default-external-api-0" Feb 27 08:11:23 crc kubenswrapper[4612]: I0227 08:11:23.371073 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/58945c32-1660-410d-9a49-b0a38829f907-logs\") pod \"glance-default-external-api-0\" (UID: \"58945c32-1660-410d-9a49-b0a38829f907\") " pod="openstack/glance-default-external-api-0" Feb 27 08:11:23 crc kubenswrapper[4612]: I0227 08:11:23.371177 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58945c32-1660-410d-9a49-b0a38829f907-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"58945c32-1660-410d-9a49-b0a38829f907\") " pod="openstack/glance-default-external-api-0" Feb 27 08:11:23 crc kubenswrapper[4612]: I0227 08:11:23.371390 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jvzwp\" (UniqueName: \"kubernetes.io/projected/58945c32-1660-410d-9a49-b0a38829f907-kube-api-access-jvzwp\") pod \"glance-default-external-api-0\" (UID: \"58945c32-1660-410d-9a49-b0a38829f907\") " pod="openstack/glance-default-external-api-0" Feb 27 08:11:23 crc kubenswrapper[4612]: I0227 08:11:23.371534 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-external-api-0\" (UID: \"58945c32-1660-410d-9a49-b0a38829f907\") " pod="openstack/glance-default-external-api-0" Feb 27 08:11:23 crc kubenswrapper[4612]: I0227 08:11:23.376221 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/58945c32-1660-410d-9a49-b0a38829f907-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"58945c32-1660-410d-9a49-b0a38829f907\") " pod="openstack/glance-default-external-api-0" Feb 27 08:11:23 crc kubenswrapper[4612]: I0227 08:11:23.378128 4612 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-external-api-0\" (UID: \"58945c32-1660-410d-9a49-b0a38829f907\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/glance-default-external-api-0" Feb 27 08:11:23 crc kubenswrapper[4612]: I0227 08:11:23.387935 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/58945c32-1660-410d-9a49-b0a38829f907-logs\") pod \"glance-default-external-api-0\" (UID: \"58945c32-1660-410d-9a49-b0a38829f907\") " pod="openstack/glance-default-external-api-0" Feb 27 08:11:23 crc kubenswrapper[4612]: I0227 08:11:23.392307 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 27 08:11:23 crc kubenswrapper[4612]: I0227 08:11:23.395583 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58945c32-1660-410d-9a49-b0a38829f907-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"58945c32-1660-410d-9a49-b0a38829f907\") " pod="openstack/glance-default-external-api-0" Feb 27 08:11:23 crc kubenswrapper[4612]: I0227 08:11:23.397384 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/58945c32-1660-410d-9a49-b0a38829f907-config-data\") pod \"glance-default-external-api-0\" (UID: \"58945c32-1660-410d-9a49-b0a38829f907\") " pod="openstack/glance-default-external-api-0" Feb 27 08:11:23 crc kubenswrapper[4612]: I0227 08:11:23.403196 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 27 08:11:23 crc kubenswrapper[4612]: I0227 08:11:23.411654 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Feb 27 08:11:23 crc kubenswrapper[4612]: I0227 08:11:23.425398 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/58945c32-1660-410d-9a49-b0a38829f907-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"58945c32-1660-410d-9a49-b0a38829f907\") " pod="openstack/glance-default-external-api-0" Feb 27 08:11:23 crc kubenswrapper[4612]: I0227 08:11:23.426291 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jvzwp\" (UniqueName: \"kubernetes.io/projected/58945c32-1660-410d-9a49-b0a38829f907-kube-api-access-jvzwp\") pod \"glance-default-external-api-0\" (UID: \"58945c32-1660-410d-9a49-b0a38829f907\") " pod="openstack/glance-default-external-api-0" Feb 27 08:11:23 crc kubenswrapper[4612]: I0227 08:11:23.427824 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Feb 27 08:11:23 crc kubenswrapper[4612]: I0227 08:11:23.466837 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/58945c32-1660-410d-9a49-b0a38829f907-scripts\") pod \"glance-default-external-api-0\" (UID: \"58945c32-1660-410d-9a49-b0a38829f907\") " pod="openstack/glance-default-external-api-0" Feb 27 08:11:23 crc kubenswrapper[4612]: I0227 08:11:23.500295 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 27 08:11:23 crc kubenswrapper[4612]: I0227 08:11:23.516608 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-external-api-0\" (UID: \"58945c32-1660-410d-9a49-b0a38829f907\") " pod="openstack/glance-default-external-api-0" Feb 27 08:11:23 crc kubenswrapper[4612]: I0227 08:11:23.582763 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f4ea14ec-7119-4952-a548-8afa0ad086bb-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"f4ea14ec-7119-4952-a548-8afa0ad086bb\") " pod="openstack/glance-default-internal-api-0" Feb 27 08:11:23 crc kubenswrapper[4612]: I0227 08:11:23.582818 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"f4ea14ec-7119-4952-a548-8afa0ad086bb\") " pod="openstack/glance-default-internal-api-0" Feb 27 08:11:23 crc kubenswrapper[4612]: I0227 08:11:23.582910 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f4ea14ec-7119-4952-a548-8afa0ad086bb-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"f4ea14ec-7119-4952-a548-8afa0ad086bb\") " pod="openstack/glance-default-internal-api-0" Feb 27 08:11:23 crc kubenswrapper[4612]: I0227 08:11:23.582931 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f4ea14ec-7119-4952-a548-8afa0ad086bb-logs\") pod \"glance-default-internal-api-0\" (UID: \"f4ea14ec-7119-4952-a548-8afa0ad086bb\") " pod="openstack/glance-default-internal-api-0" Feb 27 08:11:23 crc kubenswrapper[4612]: I0227 08:11:23.583008 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f4ea14ec-7119-4952-a548-8afa0ad086bb-scripts\") pod \"glance-default-internal-api-0\" (UID: \"f4ea14ec-7119-4952-a548-8afa0ad086bb\") " pod="openstack/glance-default-internal-api-0" Feb 27 08:11:23 crc kubenswrapper[4612]: I0227 08:11:23.583067 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f4ea14ec-7119-4952-a548-8afa0ad086bb-config-data\") pod \"glance-default-internal-api-0\" (UID: \"f4ea14ec-7119-4952-a548-8afa0ad086bb\") " pod="openstack/glance-default-internal-api-0" Feb 27 08:11:23 crc kubenswrapper[4612]: I0227 08:11:23.583134 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-867p8\" (UniqueName: \"kubernetes.io/projected/f4ea14ec-7119-4952-a548-8afa0ad086bb-kube-api-access-867p8\") pod \"glance-default-internal-api-0\" (UID: \"f4ea14ec-7119-4952-a548-8afa0ad086bb\") " pod="openstack/glance-default-internal-api-0" Feb 27 08:11:23 crc kubenswrapper[4612]: I0227 08:11:23.583221 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f4ea14ec-7119-4952-a548-8afa0ad086bb-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"f4ea14ec-7119-4952-a548-8afa0ad086bb\") " pod="openstack/glance-default-internal-api-0" Feb 27 08:11:23 crc kubenswrapper[4612]: I0227 08:11:23.684294 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f4ea14ec-7119-4952-a548-8afa0ad086bb-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"f4ea14ec-7119-4952-a548-8afa0ad086bb\") " pod="openstack/glance-default-internal-api-0" Feb 27 08:11:23 crc kubenswrapper[4612]: I0227 08:11:23.684336 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"f4ea14ec-7119-4952-a548-8afa0ad086bb\") " pod="openstack/glance-default-internal-api-0" Feb 27 08:11:23 crc kubenswrapper[4612]: I0227 08:11:23.684380 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f4ea14ec-7119-4952-a548-8afa0ad086bb-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"f4ea14ec-7119-4952-a548-8afa0ad086bb\") " pod="openstack/glance-default-internal-api-0" Feb 27 08:11:23 crc kubenswrapper[4612]: I0227 08:11:23.684396 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f4ea14ec-7119-4952-a548-8afa0ad086bb-logs\") pod \"glance-default-internal-api-0\" (UID: \"f4ea14ec-7119-4952-a548-8afa0ad086bb\") " pod="openstack/glance-default-internal-api-0" Feb 27 08:11:23 crc kubenswrapper[4612]: I0227 08:11:23.684419 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f4ea14ec-7119-4952-a548-8afa0ad086bb-scripts\") pod \"glance-default-internal-api-0\" (UID: \"f4ea14ec-7119-4952-a548-8afa0ad086bb\") " pod="openstack/glance-default-internal-api-0" Feb 27 08:11:23 crc kubenswrapper[4612]: I0227 08:11:23.684754 4612 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"f4ea14ec-7119-4952-a548-8afa0ad086bb\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/glance-default-internal-api-0" Feb 27 08:11:23 crc kubenswrapper[4612]: I0227 08:11:23.684967 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f4ea14ec-7119-4952-a548-8afa0ad086bb-logs\") pod \"glance-default-internal-api-0\" (UID: \"f4ea14ec-7119-4952-a548-8afa0ad086bb\") " pod="openstack/glance-default-internal-api-0" Feb 27 08:11:23 crc kubenswrapper[4612]: I0227 08:11:23.685228 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f4ea14ec-7119-4952-a548-8afa0ad086bb-config-data\") pod \"glance-default-internal-api-0\" (UID: \"f4ea14ec-7119-4952-a548-8afa0ad086bb\") " pod="openstack/glance-default-internal-api-0" Feb 27 08:11:23 crc kubenswrapper[4612]: I0227 08:11:23.685296 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-867p8\" (UniqueName: \"kubernetes.io/projected/f4ea14ec-7119-4952-a548-8afa0ad086bb-kube-api-access-867p8\") pod \"glance-default-internal-api-0\" (UID: \"f4ea14ec-7119-4952-a548-8afa0ad086bb\") " pod="openstack/glance-default-internal-api-0" Feb 27 08:11:23 crc kubenswrapper[4612]: I0227 08:11:23.685715 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f4ea14ec-7119-4952-a548-8afa0ad086bb-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"f4ea14ec-7119-4952-a548-8afa0ad086bb\") " pod="openstack/glance-default-internal-api-0" Feb 27 08:11:23 crc kubenswrapper[4612]: I0227 08:11:23.685972 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f4ea14ec-7119-4952-a548-8afa0ad086bb-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"f4ea14ec-7119-4952-a548-8afa0ad086bb\") " pod="openstack/glance-default-internal-api-0" Feb 27 08:11:23 crc kubenswrapper[4612]: I0227 08:11:23.692955 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f4ea14ec-7119-4952-a548-8afa0ad086bb-config-data\") pod \"glance-default-internal-api-0\" (UID: \"f4ea14ec-7119-4952-a548-8afa0ad086bb\") " pod="openstack/glance-default-internal-api-0" Feb 27 08:11:23 crc kubenswrapper[4612]: I0227 08:11:23.693433 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f4ea14ec-7119-4952-a548-8afa0ad086bb-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"f4ea14ec-7119-4952-a548-8afa0ad086bb\") " pod="openstack/glance-default-internal-api-0" Feb 27 08:11:23 crc kubenswrapper[4612]: I0227 08:11:23.698463 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f4ea14ec-7119-4952-a548-8afa0ad086bb-scripts\") pod \"glance-default-internal-api-0\" (UID: \"f4ea14ec-7119-4952-a548-8afa0ad086bb\") " pod="openstack/glance-default-internal-api-0" Feb 27 08:11:23 crc kubenswrapper[4612]: I0227 08:11:23.705296 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-867p8\" (UniqueName: \"kubernetes.io/projected/f4ea14ec-7119-4952-a548-8afa0ad086bb-kube-api-access-867p8\") pod \"glance-default-internal-api-0\" (UID: \"f4ea14ec-7119-4952-a548-8afa0ad086bb\") " pod="openstack/glance-default-internal-api-0" Feb 27 08:11:23 crc kubenswrapper[4612]: I0227 08:11:23.708917 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f4ea14ec-7119-4952-a548-8afa0ad086bb-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"f4ea14ec-7119-4952-a548-8afa0ad086bb\") " pod="openstack/glance-default-internal-api-0" Feb 27 08:11:23 crc kubenswrapper[4612]: I0227 08:11:23.714272 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"f4ea14ec-7119-4952-a548-8afa0ad086bb\") " pod="openstack/glance-default-internal-api-0" Feb 27 08:11:23 crc kubenswrapper[4612]: I0227 08:11:23.803618 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Feb 27 08:11:23 crc kubenswrapper[4612]: I0227 08:11:23.865360 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Feb 27 08:11:24 crc kubenswrapper[4612]: I0227 08:11:24.458706 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Feb 27 08:11:24 crc kubenswrapper[4612]: I0227 08:11:24.811202 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-6b985547d4-mkfpv" Feb 27 08:11:24 crc kubenswrapper[4612]: I0227 08:11:24.880944 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="16f91699-dd9d-4e71-8ff2-2bb767b3655c" path="/var/lib/kubelet/pods/16f91699-dd9d-4e71-8ff2-2bb767b3655c/volumes" Feb 27 08:11:24 crc kubenswrapper[4612]: I0227 08:11:24.884565 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f3bb8dc2-0c27-4a7a-baa1-1bfd8916f920" path="/var/lib/kubelet/pods/f3bb8dc2-0c27-4a7a-baa1-1bfd8916f920/volumes" Feb 27 08:11:24 crc kubenswrapper[4612]: I0227 08:11:24.942314 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-77948db5bb-97qh6" Feb 27 08:11:25 crc kubenswrapper[4612]: I0227 08:11:25.037230 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-6b985547d4-mkfpv"] Feb 27 08:11:25 crc kubenswrapper[4612]: I0227 08:11:25.098815 4612 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-6b985547d4-mkfpv" podUID="76009243-16b3-4652-abfe-a6cdba363724" containerName="horizon-log" containerID="cri-o://deeb662d0bcfb1868b750b03f6073e2ff029f55cf5b6b403283389a06b1b3b17" gracePeriod=30 Feb 27 08:11:25 crc kubenswrapper[4612]: I0227 08:11:25.099078 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f4ea14ec-7119-4952-a548-8afa0ad086bb","Type":"ContainerStarted","Data":"93b24893ec011d57978da8f9a1725365ed8392184d6bc05a2c663bb0357200d5"} Feb 27 08:11:25 crc kubenswrapper[4612]: I0227 08:11:25.099330 4612 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-6b985547d4-mkfpv" podUID="76009243-16b3-4652-abfe-a6cdba363724" containerName="horizon" containerID="cri-o://36510ad0ab2650616b7e36d7449187c82f7e47b4ef8d4d802e348d602213aa5a" gracePeriod=30 Feb 27 08:11:25 crc kubenswrapper[4612]: I0227 08:11:25.217299 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Feb 27 08:11:25 crc kubenswrapper[4612]: W0227 08:11:25.223217 4612 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod58945c32_1660_410d_9a49_b0a38829f907.slice/crio-fedb41d0d51efab9d1891f4ceb978b144692af3dbd1e539d8eb7166b8898ab2d WatchSource:0}: Error finding container fedb41d0d51efab9d1891f4ceb978b144692af3dbd1e539d8eb7166b8898ab2d: Status 404 returned error can't find the container with id fedb41d0d51efab9d1891f4ceb978b144692af3dbd1e539d8eb7166b8898ab2d Feb 27 08:11:26 crc kubenswrapper[4612]: I0227 08:11:26.111587 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"58945c32-1660-410d-9a49-b0a38829f907","Type":"ContainerStarted","Data":"38e88dd01c8e3b7226fea0a81c3e8441626657edfb7f38da58d83e7f1d1ad6cb"} Feb 27 08:11:26 crc kubenswrapper[4612]: I0227 08:11:26.111857 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"58945c32-1660-410d-9a49-b0a38829f907","Type":"ContainerStarted","Data":"fedb41d0d51efab9d1891f4ceb978b144692af3dbd1e539d8eb7166b8898ab2d"} Feb 27 08:11:26 crc kubenswrapper[4612]: I0227 08:11:26.115011 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f4ea14ec-7119-4952-a548-8afa0ad086bb","Type":"ContainerStarted","Data":"9c11abe7397ba2a461769b87fb485c8da3bed96da551e68761982de9842d583a"} Feb 27 08:11:26 crc kubenswrapper[4612]: I0227 08:11:26.118460 4612 generic.go:334] "Generic (PLEG): container finished" podID="1f1e76b7-ca21-4a0b-ac75-0ff917106aa2" containerID="540c359bbafca25714b4f9d4d4571dcf4ca593ccb6e817557d126798c24dea03" exitCode=0 Feb 27 08:11:26 crc kubenswrapper[4612]: I0227 08:11:26.118488 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1f1e76b7-ca21-4a0b-ac75-0ff917106aa2","Type":"ContainerDied","Data":"540c359bbafca25714b4f9d4d4571dcf4ca593ccb6e817557d126798c24dea03"} Feb 27 08:11:26 crc kubenswrapper[4612]: I0227 08:11:26.415972 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 27 08:11:26 crc kubenswrapper[4612]: I0227 08:11:26.572400 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1f1e76b7-ca21-4a0b-ac75-0ff917106aa2-config-data\") pod \"1f1e76b7-ca21-4a0b-ac75-0ff917106aa2\" (UID: \"1f1e76b7-ca21-4a0b-ac75-0ff917106aa2\") " Feb 27 08:11:26 crc kubenswrapper[4612]: I0227 08:11:26.572619 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-br4rw\" (UniqueName: \"kubernetes.io/projected/1f1e76b7-ca21-4a0b-ac75-0ff917106aa2-kube-api-access-br4rw\") pod \"1f1e76b7-ca21-4a0b-ac75-0ff917106aa2\" (UID: \"1f1e76b7-ca21-4a0b-ac75-0ff917106aa2\") " Feb 27 08:11:26 crc kubenswrapper[4612]: I0227 08:11:26.572665 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1f1e76b7-ca21-4a0b-ac75-0ff917106aa2-run-httpd\") pod \"1f1e76b7-ca21-4a0b-ac75-0ff917106aa2\" (UID: \"1f1e76b7-ca21-4a0b-ac75-0ff917106aa2\") " Feb 27 08:11:26 crc kubenswrapper[4612]: I0227 08:11:26.572757 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/1f1e76b7-ca21-4a0b-ac75-0ff917106aa2-ceilometer-tls-certs\") pod \"1f1e76b7-ca21-4a0b-ac75-0ff917106aa2\" (UID: \"1f1e76b7-ca21-4a0b-ac75-0ff917106aa2\") " Feb 27 08:11:26 crc kubenswrapper[4612]: I0227 08:11:26.572816 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1f1e76b7-ca21-4a0b-ac75-0ff917106aa2-scripts\") pod \"1f1e76b7-ca21-4a0b-ac75-0ff917106aa2\" (UID: \"1f1e76b7-ca21-4a0b-ac75-0ff917106aa2\") " Feb 27 08:11:26 crc kubenswrapper[4612]: I0227 08:11:26.572859 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1f1e76b7-ca21-4a0b-ac75-0ff917106aa2-combined-ca-bundle\") pod \"1f1e76b7-ca21-4a0b-ac75-0ff917106aa2\" (UID: \"1f1e76b7-ca21-4a0b-ac75-0ff917106aa2\") " Feb 27 08:11:26 crc kubenswrapper[4612]: I0227 08:11:26.572901 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1f1e76b7-ca21-4a0b-ac75-0ff917106aa2-log-httpd\") pod \"1f1e76b7-ca21-4a0b-ac75-0ff917106aa2\" (UID: \"1f1e76b7-ca21-4a0b-ac75-0ff917106aa2\") " Feb 27 08:11:26 crc kubenswrapper[4612]: I0227 08:11:26.572949 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1f1e76b7-ca21-4a0b-ac75-0ff917106aa2-sg-core-conf-yaml\") pod \"1f1e76b7-ca21-4a0b-ac75-0ff917106aa2\" (UID: \"1f1e76b7-ca21-4a0b-ac75-0ff917106aa2\") " Feb 27 08:11:26 crc kubenswrapper[4612]: I0227 08:11:26.573592 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1f1e76b7-ca21-4a0b-ac75-0ff917106aa2-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "1f1e76b7-ca21-4a0b-ac75-0ff917106aa2" (UID: "1f1e76b7-ca21-4a0b-ac75-0ff917106aa2"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 08:11:26 crc kubenswrapper[4612]: I0227 08:11:26.574799 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1f1e76b7-ca21-4a0b-ac75-0ff917106aa2-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "1f1e76b7-ca21-4a0b-ac75-0ff917106aa2" (UID: "1f1e76b7-ca21-4a0b-ac75-0ff917106aa2"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 08:11:26 crc kubenswrapper[4612]: I0227 08:11:26.609993 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1f1e76b7-ca21-4a0b-ac75-0ff917106aa2-scripts" (OuterVolumeSpecName: "scripts") pod "1f1e76b7-ca21-4a0b-ac75-0ff917106aa2" (UID: "1f1e76b7-ca21-4a0b-ac75-0ff917106aa2"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:11:26 crc kubenswrapper[4612]: I0227 08:11:26.610437 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1f1e76b7-ca21-4a0b-ac75-0ff917106aa2-kube-api-access-br4rw" (OuterVolumeSpecName: "kube-api-access-br4rw") pod "1f1e76b7-ca21-4a0b-ac75-0ff917106aa2" (UID: "1f1e76b7-ca21-4a0b-ac75-0ff917106aa2"). InnerVolumeSpecName "kube-api-access-br4rw". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 08:11:26 crc kubenswrapper[4612]: I0227 08:11:26.660921 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1f1e76b7-ca21-4a0b-ac75-0ff917106aa2-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "1f1e76b7-ca21-4a0b-ac75-0ff917106aa2" (UID: "1f1e76b7-ca21-4a0b-ac75-0ff917106aa2"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:11:26 crc kubenswrapper[4612]: I0227 08:11:26.674086 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1f1e76b7-ca21-4a0b-ac75-0ff917106aa2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1f1e76b7-ca21-4a0b-ac75-0ff917106aa2" (UID: "1f1e76b7-ca21-4a0b-ac75-0ff917106aa2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:11:26 crc kubenswrapper[4612]: I0227 08:11:26.675275 4612 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1f1e76b7-ca21-4a0b-ac75-0ff917106aa2-scripts\") on node \"crc\" DevicePath \"\"" Feb 27 08:11:26 crc kubenswrapper[4612]: I0227 08:11:26.675311 4612 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1f1e76b7-ca21-4a0b-ac75-0ff917106aa2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 08:11:26 crc kubenswrapper[4612]: I0227 08:11:26.675321 4612 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1f1e76b7-ca21-4a0b-ac75-0ff917106aa2-log-httpd\") on node \"crc\" DevicePath \"\"" Feb 27 08:11:26 crc kubenswrapper[4612]: I0227 08:11:26.675331 4612 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1f1e76b7-ca21-4a0b-ac75-0ff917106aa2-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Feb 27 08:11:26 crc kubenswrapper[4612]: I0227 08:11:26.675340 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-br4rw\" (UniqueName: \"kubernetes.io/projected/1f1e76b7-ca21-4a0b-ac75-0ff917106aa2-kube-api-access-br4rw\") on node \"crc\" DevicePath \"\"" Feb 27 08:11:26 crc kubenswrapper[4612]: I0227 08:11:26.675349 4612 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1f1e76b7-ca21-4a0b-ac75-0ff917106aa2-run-httpd\") on node \"crc\" DevicePath \"\"" Feb 27 08:11:26 crc kubenswrapper[4612]: I0227 08:11:26.688153 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1f1e76b7-ca21-4a0b-ac75-0ff917106aa2-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "1f1e76b7-ca21-4a0b-ac75-0ff917106aa2" (UID: "1f1e76b7-ca21-4a0b-ac75-0ff917106aa2"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:11:26 crc kubenswrapper[4612]: I0227 08:11:26.695310 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1f1e76b7-ca21-4a0b-ac75-0ff917106aa2-config-data" (OuterVolumeSpecName: "config-data") pod "1f1e76b7-ca21-4a0b-ac75-0ff917106aa2" (UID: "1f1e76b7-ca21-4a0b-ac75-0ff917106aa2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:11:26 crc kubenswrapper[4612]: I0227 08:11:26.776774 4612 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1f1e76b7-ca21-4a0b-ac75-0ff917106aa2-config-data\") on node \"crc\" DevicePath \"\"" Feb 27 08:11:26 crc kubenswrapper[4612]: I0227 08:11:26.776811 4612 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/1f1e76b7-ca21-4a0b-ac75-0ff917106aa2-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 27 08:11:27 crc kubenswrapper[4612]: I0227 08:11:27.137386 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"58945c32-1660-410d-9a49-b0a38829f907","Type":"ContainerStarted","Data":"467b4b9032c5b95fcbed84b8f67a0d6bba546250eb6999de5017b454d54dc841"} Feb 27 08:11:27 crc kubenswrapper[4612]: I0227 08:11:27.140372 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"f4ea14ec-7119-4952-a548-8afa0ad086bb","Type":"ContainerStarted","Data":"2ea9ab13834aa17b3be2cc4f5a40054124cc462f726a7d98458a52a0c17f5c7b"} Feb 27 08:11:27 crc kubenswrapper[4612]: I0227 08:11:27.145305 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1f1e76b7-ca21-4a0b-ac75-0ff917106aa2","Type":"ContainerDied","Data":"2394db91ebe749930bf8729ac23fa1fd1853550f6085d41b3c3943152c74f190"} Feb 27 08:11:27 crc kubenswrapper[4612]: I0227 08:11:27.145345 4612 scope.go:117] "RemoveContainer" containerID="32e856f1f5d26d4f0b991fa4f1943f3de4404572c4bd114be26e59f84691c594" Feb 27 08:11:27 crc kubenswrapper[4612]: I0227 08:11:27.145492 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 27 08:11:27 crc kubenswrapper[4612]: I0227 08:11:27.185421 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=4.185399603 podStartE2EDuration="4.185399603s" podCreationTimestamp="2026-02-27 08:11:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 08:11:27.159685194 +0000 UTC m=+1345.013615192" watchObservedRunningTime="2026-02-27 08:11:27.185399603 +0000 UTC m=+1345.039329611" Feb 27 08:11:27 crc kubenswrapper[4612]: I0227 08:11:27.195559 4612 scope.go:117] "RemoveContainer" containerID="affd5d783f5ae025ebd8c67d5813022c368a25b81b112427191ecceb399717e9" Feb 27 08:11:27 crc kubenswrapper[4612]: I0227 08:11:27.206980 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=4.206958902 podStartE2EDuration="4.206958902s" podCreationTimestamp="2026-02-27 08:11:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 08:11:27.180523513 +0000 UTC m=+1345.034453511" watchObservedRunningTime="2026-02-27 08:11:27.206958902 +0000 UTC m=+1345.060888900" Feb 27 08:11:27 crc kubenswrapper[4612]: I0227 08:11:27.225460 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 27 08:11:27 crc kubenswrapper[4612]: I0227 08:11:27.231636 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Feb 27 08:11:27 crc kubenswrapper[4612]: I0227 08:11:27.234975 4612 scope.go:117] "RemoveContainer" containerID="27daefaeecf254b2c7057d0bdc634e5c6eac899815b470415df04a22acaf432d" Feb 27 08:11:27 crc kubenswrapper[4612]: I0227 08:11:27.243439 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Feb 27 08:11:27 crc kubenswrapper[4612]: E0227 08:11:27.243790 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1f1e76b7-ca21-4a0b-ac75-0ff917106aa2" containerName="ceilometer-central-agent" Feb 27 08:11:27 crc kubenswrapper[4612]: I0227 08:11:27.243806 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="1f1e76b7-ca21-4a0b-ac75-0ff917106aa2" containerName="ceilometer-central-agent" Feb 27 08:11:27 crc kubenswrapper[4612]: E0227 08:11:27.243820 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1f1e76b7-ca21-4a0b-ac75-0ff917106aa2" containerName="ceilometer-notification-agent" Feb 27 08:11:27 crc kubenswrapper[4612]: I0227 08:11:27.243826 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="1f1e76b7-ca21-4a0b-ac75-0ff917106aa2" containerName="ceilometer-notification-agent" Feb 27 08:11:27 crc kubenswrapper[4612]: E0227 08:11:27.243846 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1f1e76b7-ca21-4a0b-ac75-0ff917106aa2" containerName="sg-core" Feb 27 08:11:27 crc kubenswrapper[4612]: I0227 08:11:27.243852 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="1f1e76b7-ca21-4a0b-ac75-0ff917106aa2" containerName="sg-core" Feb 27 08:11:27 crc kubenswrapper[4612]: E0227 08:11:27.243862 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1f1e76b7-ca21-4a0b-ac75-0ff917106aa2" containerName="proxy-httpd" Feb 27 08:11:27 crc kubenswrapper[4612]: I0227 08:11:27.243867 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="1f1e76b7-ca21-4a0b-ac75-0ff917106aa2" containerName="proxy-httpd" Feb 27 08:11:27 crc kubenswrapper[4612]: I0227 08:11:27.244048 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="1f1e76b7-ca21-4a0b-ac75-0ff917106aa2" containerName="ceilometer-notification-agent" Feb 27 08:11:27 crc kubenswrapper[4612]: I0227 08:11:27.244068 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="1f1e76b7-ca21-4a0b-ac75-0ff917106aa2" containerName="ceilometer-central-agent" Feb 27 08:11:27 crc kubenswrapper[4612]: I0227 08:11:27.244080 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="1f1e76b7-ca21-4a0b-ac75-0ff917106aa2" containerName="sg-core" Feb 27 08:11:27 crc kubenswrapper[4612]: I0227 08:11:27.244089 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="1f1e76b7-ca21-4a0b-ac75-0ff917106aa2" containerName="proxy-httpd" Feb 27 08:11:27 crc kubenswrapper[4612]: I0227 08:11:27.248283 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 27 08:11:27 crc kubenswrapper[4612]: I0227 08:11:27.251809 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Feb 27 08:11:27 crc kubenswrapper[4612]: I0227 08:11:27.252044 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Feb 27 08:11:27 crc kubenswrapper[4612]: I0227 08:11:27.252209 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Feb 27 08:11:27 crc kubenswrapper[4612]: I0227 08:11:27.266455 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 27 08:11:27 crc kubenswrapper[4612]: I0227 08:11:27.304460 4612 scope.go:117] "RemoveContainer" containerID="540c359bbafca25714b4f9d4d4571dcf4ca593ccb6e817557d126798c24dea03" Feb 27 08:11:27 crc kubenswrapper[4612]: I0227 08:11:27.388358 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f6ca7ac0-4618-4a5b-af1a-7df5fa0e3701-run-httpd\") pod \"ceilometer-0\" (UID: \"f6ca7ac0-4618-4a5b-af1a-7df5fa0e3701\") " pod="openstack/ceilometer-0" Feb 27 08:11:27 crc kubenswrapper[4612]: I0227 08:11:27.388635 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f6ca7ac0-4618-4a5b-af1a-7df5fa0e3701-scripts\") pod \"ceilometer-0\" (UID: \"f6ca7ac0-4618-4a5b-af1a-7df5fa0e3701\") " pod="openstack/ceilometer-0" Feb 27 08:11:27 crc kubenswrapper[4612]: I0227 08:11:27.388655 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f6ca7ac0-4618-4a5b-af1a-7df5fa0e3701-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f6ca7ac0-4618-4a5b-af1a-7df5fa0e3701\") " pod="openstack/ceilometer-0" Feb 27 08:11:27 crc kubenswrapper[4612]: I0227 08:11:27.388697 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6ca7ac0-4618-4a5b-af1a-7df5fa0e3701-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f6ca7ac0-4618-4a5b-af1a-7df5fa0e3701\") " pod="openstack/ceilometer-0" Feb 27 08:11:27 crc kubenswrapper[4612]: I0227 08:11:27.388756 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f6ca7ac0-4618-4a5b-af1a-7df5fa0e3701-config-data\") pod \"ceilometer-0\" (UID: \"f6ca7ac0-4618-4a5b-af1a-7df5fa0e3701\") " pod="openstack/ceilometer-0" Feb 27 08:11:27 crc kubenswrapper[4612]: I0227 08:11:27.388776 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f6ca7ac0-4618-4a5b-af1a-7df5fa0e3701-log-httpd\") pod \"ceilometer-0\" (UID: \"f6ca7ac0-4618-4a5b-af1a-7df5fa0e3701\") " pod="openstack/ceilometer-0" Feb 27 08:11:27 crc kubenswrapper[4612]: I0227 08:11:27.388798 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/f6ca7ac0-4618-4a5b-af1a-7df5fa0e3701-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"f6ca7ac0-4618-4a5b-af1a-7df5fa0e3701\") " pod="openstack/ceilometer-0" Feb 27 08:11:27 crc kubenswrapper[4612]: I0227 08:11:27.388933 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z728g\" (UniqueName: \"kubernetes.io/projected/f6ca7ac0-4618-4a5b-af1a-7df5fa0e3701-kube-api-access-z728g\") pod \"ceilometer-0\" (UID: \"f6ca7ac0-4618-4a5b-af1a-7df5fa0e3701\") " pod="openstack/ceilometer-0" Feb 27 08:11:27 crc kubenswrapper[4612]: I0227 08:11:27.491159 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f6ca7ac0-4618-4a5b-af1a-7df5fa0e3701-run-httpd\") pod \"ceilometer-0\" (UID: \"f6ca7ac0-4618-4a5b-af1a-7df5fa0e3701\") " pod="openstack/ceilometer-0" Feb 27 08:11:27 crc kubenswrapper[4612]: I0227 08:11:27.491257 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f6ca7ac0-4618-4a5b-af1a-7df5fa0e3701-scripts\") pod \"ceilometer-0\" (UID: \"f6ca7ac0-4618-4a5b-af1a-7df5fa0e3701\") " pod="openstack/ceilometer-0" Feb 27 08:11:27 crc kubenswrapper[4612]: I0227 08:11:27.491277 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f6ca7ac0-4618-4a5b-af1a-7df5fa0e3701-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f6ca7ac0-4618-4a5b-af1a-7df5fa0e3701\") " pod="openstack/ceilometer-0" Feb 27 08:11:27 crc kubenswrapper[4612]: I0227 08:11:27.491323 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6ca7ac0-4618-4a5b-af1a-7df5fa0e3701-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f6ca7ac0-4618-4a5b-af1a-7df5fa0e3701\") " pod="openstack/ceilometer-0" Feb 27 08:11:27 crc kubenswrapper[4612]: I0227 08:11:27.491346 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f6ca7ac0-4618-4a5b-af1a-7df5fa0e3701-config-data\") pod \"ceilometer-0\" (UID: \"f6ca7ac0-4618-4a5b-af1a-7df5fa0e3701\") " pod="openstack/ceilometer-0" Feb 27 08:11:27 crc kubenswrapper[4612]: I0227 08:11:27.491369 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f6ca7ac0-4618-4a5b-af1a-7df5fa0e3701-log-httpd\") pod \"ceilometer-0\" (UID: \"f6ca7ac0-4618-4a5b-af1a-7df5fa0e3701\") " pod="openstack/ceilometer-0" Feb 27 08:11:27 crc kubenswrapper[4612]: I0227 08:11:27.491392 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/f6ca7ac0-4618-4a5b-af1a-7df5fa0e3701-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"f6ca7ac0-4618-4a5b-af1a-7df5fa0e3701\") " pod="openstack/ceilometer-0" Feb 27 08:11:27 crc kubenswrapper[4612]: I0227 08:11:27.491442 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z728g\" (UniqueName: \"kubernetes.io/projected/f6ca7ac0-4618-4a5b-af1a-7df5fa0e3701-kube-api-access-z728g\") pod \"ceilometer-0\" (UID: \"f6ca7ac0-4618-4a5b-af1a-7df5fa0e3701\") " pod="openstack/ceilometer-0" Feb 27 08:11:27 crc kubenswrapper[4612]: I0227 08:11:27.492942 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f6ca7ac0-4618-4a5b-af1a-7df5fa0e3701-log-httpd\") pod \"ceilometer-0\" (UID: \"f6ca7ac0-4618-4a5b-af1a-7df5fa0e3701\") " pod="openstack/ceilometer-0" Feb 27 08:11:27 crc kubenswrapper[4612]: I0227 08:11:27.494122 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f6ca7ac0-4618-4a5b-af1a-7df5fa0e3701-run-httpd\") pod \"ceilometer-0\" (UID: \"f6ca7ac0-4618-4a5b-af1a-7df5fa0e3701\") " pod="openstack/ceilometer-0" Feb 27 08:11:27 crc kubenswrapper[4612]: I0227 08:11:27.499792 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6ca7ac0-4618-4a5b-af1a-7df5fa0e3701-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f6ca7ac0-4618-4a5b-af1a-7df5fa0e3701\") " pod="openstack/ceilometer-0" Feb 27 08:11:27 crc kubenswrapper[4612]: I0227 08:11:27.506377 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f6ca7ac0-4618-4a5b-af1a-7df5fa0e3701-scripts\") pod \"ceilometer-0\" (UID: \"f6ca7ac0-4618-4a5b-af1a-7df5fa0e3701\") " pod="openstack/ceilometer-0" Feb 27 08:11:27 crc kubenswrapper[4612]: I0227 08:11:27.510797 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/f6ca7ac0-4618-4a5b-af1a-7df5fa0e3701-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"f6ca7ac0-4618-4a5b-af1a-7df5fa0e3701\") " pod="openstack/ceilometer-0" Feb 27 08:11:27 crc kubenswrapper[4612]: I0227 08:11:27.511891 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f6ca7ac0-4618-4a5b-af1a-7df5fa0e3701-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f6ca7ac0-4618-4a5b-af1a-7df5fa0e3701\") " pod="openstack/ceilometer-0" Feb 27 08:11:27 crc kubenswrapper[4612]: I0227 08:11:27.514509 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f6ca7ac0-4618-4a5b-af1a-7df5fa0e3701-config-data\") pod \"ceilometer-0\" (UID: \"f6ca7ac0-4618-4a5b-af1a-7df5fa0e3701\") " pod="openstack/ceilometer-0" Feb 27 08:11:27 crc kubenswrapper[4612]: I0227 08:11:27.515564 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z728g\" (UniqueName: \"kubernetes.io/projected/f6ca7ac0-4618-4a5b-af1a-7df5fa0e3701-kube-api-access-z728g\") pod \"ceilometer-0\" (UID: \"f6ca7ac0-4618-4a5b-af1a-7df5fa0e3701\") " pod="openstack/ceilometer-0" Feb 27 08:11:27 crc kubenswrapper[4612]: I0227 08:11:27.605184 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 27 08:11:28 crc kubenswrapper[4612]: W0227 08:11:28.086412 4612 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf6ca7ac0_4618_4a5b_af1a_7df5fa0e3701.slice/crio-df40f175d4e13aaf29051c5a619d63514d77a1c982646acb8391a1eed8cc83b5 WatchSource:0}: Error finding container df40f175d4e13aaf29051c5a619d63514d77a1c982646acb8391a1eed8cc83b5: Status 404 returned error can't find the container with id df40f175d4e13aaf29051c5a619d63514d77a1c982646acb8391a1eed8cc83b5 Feb 27 08:11:28 crc kubenswrapper[4612]: I0227 08:11:28.094813 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 27 08:11:28 crc kubenswrapper[4612]: I0227 08:11:28.174485 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f6ca7ac0-4618-4a5b-af1a-7df5fa0e3701","Type":"ContainerStarted","Data":"df40f175d4e13aaf29051c5a619d63514d77a1c982646acb8391a1eed8cc83b5"} Feb 27 08:11:28 crc kubenswrapper[4612]: I0227 08:11:28.865242 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1f1e76b7-ca21-4a0b-ac75-0ff917106aa2" path="/var/lib/kubelet/pods/1f1e76b7-ca21-4a0b-ac75-0ff917106aa2/volumes" Feb 27 08:11:29 crc kubenswrapper[4612]: I0227 08:11:29.108198 4612 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-6b985547d4-mkfpv" podUID="76009243-16b3-4652-abfe-a6cdba363724" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.152:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.152:8443: connect: connection refused" Feb 27 08:11:29 crc kubenswrapper[4612]: I0227 08:11:29.194047 4612 generic.go:334] "Generic (PLEG): container finished" podID="76009243-16b3-4652-abfe-a6cdba363724" containerID="36510ad0ab2650616b7e36d7449187c82f7e47b4ef8d4d802e348d602213aa5a" exitCode=0 Feb 27 08:11:29 crc kubenswrapper[4612]: I0227 08:11:29.194113 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6b985547d4-mkfpv" event={"ID":"76009243-16b3-4652-abfe-a6cdba363724","Type":"ContainerDied","Data":"36510ad0ab2650616b7e36d7449187c82f7e47b4ef8d4d802e348d602213aa5a"} Feb 27 08:11:29 crc kubenswrapper[4612]: I0227 08:11:29.194149 4612 scope.go:117] "RemoveContainer" containerID="a0baf671a66706916e8b02ab97ccdead951de20e3f62e64d913a8590aa359cb3" Feb 27 08:11:29 crc kubenswrapper[4612]: I0227 08:11:29.196367 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f6ca7ac0-4618-4a5b-af1a-7df5fa0e3701","Type":"ContainerStarted","Data":"dc0c3dd3d7b1c7ed0866ab1f050022d49c1a0f1c9020462c96b150e41116c369"} Feb 27 08:11:30 crc kubenswrapper[4612]: I0227 08:11:30.206441 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f6ca7ac0-4618-4a5b-af1a-7df5fa0e3701","Type":"ContainerStarted","Data":"01be7c2b40fa8ae2e3fc259c649bbe57b7ff91643eb9d4b0a6310dc040b990e8"} Feb 27 08:11:31 crc kubenswrapper[4612]: I0227 08:11:31.215769 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f6ca7ac0-4618-4a5b-af1a-7df5fa0e3701","Type":"ContainerStarted","Data":"f70a60fbef6664a97cb3a6aea70caf097b359a8295be9fd0b3ce433a45ffd9db"} Feb 27 08:11:32 crc kubenswrapper[4612]: I0227 08:11:32.231735 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f6ca7ac0-4618-4a5b-af1a-7df5fa0e3701","Type":"ContainerStarted","Data":"4ff36f9dedc579a142a58de2fc0e4ceb22548dad53a45b4ad10082fe2343907e"} Feb 27 08:11:32 crc kubenswrapper[4612]: I0227 08:11:32.232082 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Feb 27 08:11:32 crc kubenswrapper[4612]: I0227 08:11:32.234602 4612 generic.go:334] "Generic (PLEG): container finished" podID="3d52b072-aded-480f-ae46-d89c9be0c9f8" containerID="73e8d5c17c8a1763106d32f761ba1f2030f5bf13fe92371e47c559d774a2b505" exitCode=0 Feb 27 08:11:32 crc kubenswrapper[4612]: I0227 08:11:32.234625 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-z67kr" event={"ID":"3d52b072-aded-480f-ae46-d89c9be0c9f8","Type":"ContainerDied","Data":"73e8d5c17c8a1763106d32f761ba1f2030f5bf13fe92371e47c559d774a2b505"} Feb 27 08:11:32 crc kubenswrapper[4612]: I0227 08:11:32.271321 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.606467303 podStartE2EDuration="5.27129588s" podCreationTimestamp="2026-02-27 08:11:27 +0000 UTC" firstStartedPulling="2026-02-27 08:11:28.089819476 +0000 UTC m=+1345.943749474" lastFinishedPulling="2026-02-27 08:11:31.754648053 +0000 UTC m=+1349.608578051" observedRunningTime="2026-02-27 08:11:32.256226947 +0000 UTC m=+1350.110156985" watchObservedRunningTime="2026-02-27 08:11:32.27129588 +0000 UTC m=+1350.125225918" Feb 27 08:11:33 crc kubenswrapper[4612]: I0227 08:11:33.632218 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-z67kr" Feb 27 08:11:33 crc kubenswrapper[4612]: I0227 08:11:33.720762 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3d52b072-aded-480f-ae46-d89c9be0c9f8-scripts\") pod \"3d52b072-aded-480f-ae46-d89c9be0c9f8\" (UID: \"3d52b072-aded-480f-ae46-d89c9be0c9f8\") " Feb 27 08:11:33 crc kubenswrapper[4612]: I0227 08:11:33.720879 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gw2x6\" (UniqueName: \"kubernetes.io/projected/3d52b072-aded-480f-ae46-d89c9be0c9f8-kube-api-access-gw2x6\") pod \"3d52b072-aded-480f-ae46-d89c9be0c9f8\" (UID: \"3d52b072-aded-480f-ae46-d89c9be0c9f8\") " Feb 27 08:11:33 crc kubenswrapper[4612]: I0227 08:11:33.720956 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3d52b072-aded-480f-ae46-d89c9be0c9f8-combined-ca-bundle\") pod \"3d52b072-aded-480f-ae46-d89c9be0c9f8\" (UID: \"3d52b072-aded-480f-ae46-d89c9be0c9f8\") " Feb 27 08:11:33 crc kubenswrapper[4612]: I0227 08:11:33.721028 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3d52b072-aded-480f-ae46-d89c9be0c9f8-config-data\") pod \"3d52b072-aded-480f-ae46-d89c9be0c9f8\" (UID: \"3d52b072-aded-480f-ae46-d89c9be0c9f8\") " Feb 27 08:11:33 crc kubenswrapper[4612]: I0227 08:11:33.727925 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3d52b072-aded-480f-ae46-d89c9be0c9f8-kube-api-access-gw2x6" (OuterVolumeSpecName: "kube-api-access-gw2x6") pod "3d52b072-aded-480f-ae46-d89c9be0c9f8" (UID: "3d52b072-aded-480f-ae46-d89c9be0c9f8"). InnerVolumeSpecName "kube-api-access-gw2x6". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 08:11:33 crc kubenswrapper[4612]: I0227 08:11:33.745858 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3d52b072-aded-480f-ae46-d89c9be0c9f8-scripts" (OuterVolumeSpecName: "scripts") pod "3d52b072-aded-480f-ae46-d89c9be0c9f8" (UID: "3d52b072-aded-480f-ae46-d89c9be0c9f8"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:11:33 crc kubenswrapper[4612]: I0227 08:11:33.770400 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3d52b072-aded-480f-ae46-d89c9be0c9f8-config-data" (OuterVolumeSpecName: "config-data") pod "3d52b072-aded-480f-ae46-d89c9be0c9f8" (UID: "3d52b072-aded-480f-ae46-d89c9be0c9f8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:11:33 crc kubenswrapper[4612]: I0227 08:11:33.788842 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3d52b072-aded-480f-ae46-d89c9be0c9f8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3d52b072-aded-480f-ae46-d89c9be0c9f8" (UID: "3d52b072-aded-480f-ae46-d89c9be0c9f8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:11:33 crc kubenswrapper[4612]: I0227 08:11:33.804542 4612 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Feb 27 08:11:33 crc kubenswrapper[4612]: I0227 08:11:33.804585 4612 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Feb 27 08:11:33 crc kubenswrapper[4612]: I0227 08:11:33.823290 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gw2x6\" (UniqueName: \"kubernetes.io/projected/3d52b072-aded-480f-ae46-d89c9be0c9f8-kube-api-access-gw2x6\") on node \"crc\" DevicePath \"\"" Feb 27 08:11:33 crc kubenswrapper[4612]: I0227 08:11:33.823324 4612 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3d52b072-aded-480f-ae46-d89c9be0c9f8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 08:11:33 crc kubenswrapper[4612]: I0227 08:11:33.823334 4612 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3d52b072-aded-480f-ae46-d89c9be0c9f8-config-data\") on node \"crc\" DevicePath \"\"" Feb 27 08:11:33 crc kubenswrapper[4612]: I0227 08:11:33.823346 4612 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3d52b072-aded-480f-ae46-d89c9be0c9f8-scripts\") on node \"crc\" DevicePath \"\"" Feb 27 08:11:33 crc kubenswrapper[4612]: I0227 08:11:33.854352 4612 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Feb 27 08:11:33 crc kubenswrapper[4612]: I0227 08:11:33.854875 4612 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Feb 27 08:11:33 crc kubenswrapper[4612]: I0227 08:11:33.866136 4612 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Feb 27 08:11:33 crc kubenswrapper[4612]: I0227 08:11:33.866179 4612 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Feb 27 08:11:33 crc kubenswrapper[4612]: I0227 08:11:33.904016 4612 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Feb 27 08:11:33 crc kubenswrapper[4612]: I0227 08:11:33.910336 4612 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Feb 27 08:11:34 crc kubenswrapper[4612]: I0227 08:11:34.258111 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-z67kr" event={"ID":"3d52b072-aded-480f-ae46-d89c9be0c9f8","Type":"ContainerDied","Data":"425959e9762745eaa61d56db0138f50190e189c59350c8b4a043bbb6bcbd8271"} Feb 27 08:11:34 crc kubenswrapper[4612]: I0227 08:11:34.258168 4612 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="425959e9762745eaa61d56db0138f50190e189c59350c8b4a043bbb6bcbd8271" Feb 27 08:11:34 crc kubenswrapper[4612]: I0227 08:11:34.258729 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-z67kr" Feb 27 08:11:34 crc kubenswrapper[4612]: I0227 08:11:34.258749 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Feb 27 08:11:34 crc kubenswrapper[4612]: I0227 08:11:34.259958 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Feb 27 08:11:34 crc kubenswrapper[4612]: I0227 08:11:34.259990 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Feb 27 08:11:34 crc kubenswrapper[4612]: I0227 08:11:34.260017 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Feb 27 08:11:34 crc kubenswrapper[4612]: I0227 08:11:34.410560 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Feb 27 08:11:34 crc kubenswrapper[4612]: E0227 08:11:34.411005 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3d52b072-aded-480f-ae46-d89c9be0c9f8" containerName="nova-cell0-conductor-db-sync" Feb 27 08:11:34 crc kubenswrapper[4612]: I0227 08:11:34.411017 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="3d52b072-aded-480f-ae46-d89c9be0c9f8" containerName="nova-cell0-conductor-db-sync" Feb 27 08:11:34 crc kubenswrapper[4612]: I0227 08:11:34.411214 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="3d52b072-aded-480f-ae46-d89c9be0c9f8" containerName="nova-cell0-conductor-db-sync" Feb 27 08:11:34 crc kubenswrapper[4612]: I0227 08:11:34.411838 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Feb 27 08:11:34 crc kubenswrapper[4612]: I0227 08:11:34.417123 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Feb 27 08:11:34 crc kubenswrapper[4612]: I0227 08:11:34.417484 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-9qmk6" Feb 27 08:11:34 crc kubenswrapper[4612]: I0227 08:11:34.437141 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Feb 27 08:11:34 crc kubenswrapper[4612]: I0227 08:11:34.537889 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/602ab3aa-60d5-45ff-9890-647a6f8ffb6c-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"602ab3aa-60d5-45ff-9890-647a6f8ffb6c\") " pod="openstack/nova-cell0-conductor-0" Feb 27 08:11:34 crc kubenswrapper[4612]: I0227 08:11:34.538166 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jwkjs\" (UniqueName: \"kubernetes.io/projected/602ab3aa-60d5-45ff-9890-647a6f8ffb6c-kube-api-access-jwkjs\") pod \"nova-cell0-conductor-0\" (UID: \"602ab3aa-60d5-45ff-9890-647a6f8ffb6c\") " pod="openstack/nova-cell0-conductor-0" Feb 27 08:11:34 crc kubenswrapper[4612]: I0227 08:11:34.538341 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/602ab3aa-60d5-45ff-9890-647a6f8ffb6c-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"602ab3aa-60d5-45ff-9890-647a6f8ffb6c\") " pod="openstack/nova-cell0-conductor-0" Feb 27 08:11:34 crc kubenswrapper[4612]: I0227 08:11:34.640073 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/602ab3aa-60d5-45ff-9890-647a6f8ffb6c-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"602ab3aa-60d5-45ff-9890-647a6f8ffb6c\") " pod="openstack/nova-cell0-conductor-0" Feb 27 08:11:34 crc kubenswrapper[4612]: I0227 08:11:34.640141 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jwkjs\" (UniqueName: \"kubernetes.io/projected/602ab3aa-60d5-45ff-9890-647a6f8ffb6c-kube-api-access-jwkjs\") pod \"nova-cell0-conductor-0\" (UID: \"602ab3aa-60d5-45ff-9890-647a6f8ffb6c\") " pod="openstack/nova-cell0-conductor-0" Feb 27 08:11:34 crc kubenswrapper[4612]: I0227 08:11:34.640198 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/602ab3aa-60d5-45ff-9890-647a6f8ffb6c-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"602ab3aa-60d5-45ff-9890-647a6f8ffb6c\") " pod="openstack/nova-cell0-conductor-0" Feb 27 08:11:34 crc kubenswrapper[4612]: I0227 08:11:34.646406 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/602ab3aa-60d5-45ff-9890-647a6f8ffb6c-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"602ab3aa-60d5-45ff-9890-647a6f8ffb6c\") " pod="openstack/nova-cell0-conductor-0" Feb 27 08:11:34 crc kubenswrapper[4612]: I0227 08:11:34.651791 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/602ab3aa-60d5-45ff-9890-647a6f8ffb6c-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"602ab3aa-60d5-45ff-9890-647a6f8ffb6c\") " pod="openstack/nova-cell0-conductor-0" Feb 27 08:11:34 crc kubenswrapper[4612]: I0227 08:11:34.676240 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jwkjs\" (UniqueName: \"kubernetes.io/projected/602ab3aa-60d5-45ff-9890-647a6f8ffb6c-kube-api-access-jwkjs\") pod \"nova-cell0-conductor-0\" (UID: \"602ab3aa-60d5-45ff-9890-647a6f8ffb6c\") " pod="openstack/nova-cell0-conductor-0" Feb 27 08:11:34 crc kubenswrapper[4612]: I0227 08:11:34.755844 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Feb 27 08:11:35 crc kubenswrapper[4612]: I0227 08:11:35.279543 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Feb 27 08:11:35 crc kubenswrapper[4612]: W0227 08:11:35.281394 4612 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod602ab3aa_60d5_45ff_9890_647a6f8ffb6c.slice/crio-3ab28c34d5ac93f8fd707b103ae4cc0fcd320c3ad721659bc610ec40fa74ec8e WatchSource:0}: Error finding container 3ab28c34d5ac93f8fd707b103ae4cc0fcd320c3ad721659bc610ec40fa74ec8e: Status 404 returned error can't find the container with id 3ab28c34d5ac93f8fd707b103ae4cc0fcd320c3ad721659bc610ec40fa74ec8e Feb 27 08:11:36 crc kubenswrapper[4612]: I0227 08:11:36.274143 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"602ab3aa-60d5-45ff-9890-647a6f8ffb6c","Type":"ContainerStarted","Data":"6ddce433418ca0fbb673f434a25c182312280d63037c5bac52a92c2892b1b42e"} Feb 27 08:11:36 crc kubenswrapper[4612]: I0227 08:11:36.274724 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"602ab3aa-60d5-45ff-9890-647a6f8ffb6c","Type":"ContainerStarted","Data":"3ab28c34d5ac93f8fd707b103ae4cc0fcd320c3ad721659bc610ec40fa74ec8e"} Feb 27 08:11:36 crc kubenswrapper[4612]: I0227 08:11:36.274748 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Feb 27 08:11:36 crc kubenswrapper[4612]: I0227 08:11:36.274242 4612 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 27 08:11:36 crc kubenswrapper[4612]: I0227 08:11:36.274777 4612 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 27 08:11:36 crc kubenswrapper[4612]: I0227 08:11:36.274192 4612 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 27 08:11:36 crc kubenswrapper[4612]: I0227 08:11:36.275135 4612 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 27 08:11:36 crc kubenswrapper[4612]: I0227 08:11:36.303335 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.303313982 podStartE2EDuration="2.303313982s" podCreationTimestamp="2026-02-27 08:11:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 08:11:36.291176154 +0000 UTC m=+1354.145106152" watchObservedRunningTime="2026-02-27 08:11:36.303313982 +0000 UTC m=+1354.157243980" Feb 27 08:11:36 crc kubenswrapper[4612]: I0227 08:11:36.934380 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Feb 27 08:11:37 crc kubenswrapper[4612]: I0227 08:11:37.099438 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Feb 27 08:11:37 crc kubenswrapper[4612]: I0227 08:11:37.411361 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Feb 27 08:11:37 crc kubenswrapper[4612]: I0227 08:11:37.411461 4612 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 27 08:11:37 crc kubenswrapper[4612]: I0227 08:11:37.470011 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Feb 27 08:11:39 crc kubenswrapper[4612]: I0227 08:11:39.107609 4612 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-6b985547d4-mkfpv" podUID="76009243-16b3-4652-abfe-a6cdba363724" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.152:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.152:8443: connect: connection refused" Feb 27 08:11:44 crc kubenswrapper[4612]: I0227 08:11:44.797048 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Feb 27 08:11:45 crc kubenswrapper[4612]: I0227 08:11:45.291739 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-x5mhp"] Feb 27 08:11:45 crc kubenswrapper[4612]: I0227 08:11:45.293458 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-x5mhp" Feb 27 08:11:45 crc kubenswrapper[4612]: I0227 08:11:45.298425 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Feb 27 08:11:45 crc kubenswrapper[4612]: I0227 08:11:45.303738 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Feb 27 08:11:45 crc kubenswrapper[4612]: I0227 08:11:45.311621 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-x5mhp"] Feb 27 08:11:45 crc kubenswrapper[4612]: I0227 08:11:45.362337 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5d42ee9b-5073-4f40-8a3d-632914f14bbc-config-data\") pod \"nova-cell0-cell-mapping-x5mhp\" (UID: \"5d42ee9b-5073-4f40-8a3d-632914f14bbc\") " pod="openstack/nova-cell0-cell-mapping-x5mhp" Feb 27 08:11:45 crc kubenswrapper[4612]: I0227 08:11:45.362379 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dx6ks\" (UniqueName: \"kubernetes.io/projected/5d42ee9b-5073-4f40-8a3d-632914f14bbc-kube-api-access-dx6ks\") pod \"nova-cell0-cell-mapping-x5mhp\" (UID: \"5d42ee9b-5073-4f40-8a3d-632914f14bbc\") " pod="openstack/nova-cell0-cell-mapping-x5mhp" Feb 27 08:11:45 crc kubenswrapper[4612]: I0227 08:11:45.362398 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d42ee9b-5073-4f40-8a3d-632914f14bbc-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-x5mhp\" (UID: \"5d42ee9b-5073-4f40-8a3d-632914f14bbc\") " pod="openstack/nova-cell0-cell-mapping-x5mhp" Feb 27 08:11:45 crc kubenswrapper[4612]: I0227 08:11:45.362442 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5d42ee9b-5073-4f40-8a3d-632914f14bbc-scripts\") pod \"nova-cell0-cell-mapping-x5mhp\" (UID: \"5d42ee9b-5073-4f40-8a3d-632914f14bbc\") " pod="openstack/nova-cell0-cell-mapping-x5mhp" Feb 27 08:11:45 crc kubenswrapper[4612]: I0227 08:11:45.463769 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5d42ee9b-5073-4f40-8a3d-632914f14bbc-config-data\") pod \"nova-cell0-cell-mapping-x5mhp\" (UID: \"5d42ee9b-5073-4f40-8a3d-632914f14bbc\") " pod="openstack/nova-cell0-cell-mapping-x5mhp" Feb 27 08:11:45 crc kubenswrapper[4612]: I0227 08:11:45.463809 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dx6ks\" (UniqueName: \"kubernetes.io/projected/5d42ee9b-5073-4f40-8a3d-632914f14bbc-kube-api-access-dx6ks\") pod \"nova-cell0-cell-mapping-x5mhp\" (UID: \"5d42ee9b-5073-4f40-8a3d-632914f14bbc\") " pod="openstack/nova-cell0-cell-mapping-x5mhp" Feb 27 08:11:45 crc kubenswrapper[4612]: I0227 08:11:45.463827 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d42ee9b-5073-4f40-8a3d-632914f14bbc-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-x5mhp\" (UID: \"5d42ee9b-5073-4f40-8a3d-632914f14bbc\") " pod="openstack/nova-cell0-cell-mapping-x5mhp" Feb 27 08:11:45 crc kubenswrapper[4612]: I0227 08:11:45.463874 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5d42ee9b-5073-4f40-8a3d-632914f14bbc-scripts\") pod \"nova-cell0-cell-mapping-x5mhp\" (UID: \"5d42ee9b-5073-4f40-8a3d-632914f14bbc\") " pod="openstack/nova-cell0-cell-mapping-x5mhp" Feb 27 08:11:45 crc kubenswrapper[4612]: I0227 08:11:45.472853 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5d42ee9b-5073-4f40-8a3d-632914f14bbc-config-data\") pod \"nova-cell0-cell-mapping-x5mhp\" (UID: \"5d42ee9b-5073-4f40-8a3d-632914f14bbc\") " pod="openstack/nova-cell0-cell-mapping-x5mhp" Feb 27 08:11:45 crc kubenswrapper[4612]: I0227 08:11:45.477082 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5d42ee9b-5073-4f40-8a3d-632914f14bbc-scripts\") pod \"nova-cell0-cell-mapping-x5mhp\" (UID: \"5d42ee9b-5073-4f40-8a3d-632914f14bbc\") " pod="openstack/nova-cell0-cell-mapping-x5mhp" Feb 27 08:11:45 crc kubenswrapper[4612]: I0227 08:11:45.484427 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d42ee9b-5073-4f40-8a3d-632914f14bbc-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-x5mhp\" (UID: \"5d42ee9b-5073-4f40-8a3d-632914f14bbc\") " pod="openstack/nova-cell0-cell-mapping-x5mhp" Feb 27 08:11:45 crc kubenswrapper[4612]: I0227 08:11:45.495183 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dx6ks\" (UniqueName: \"kubernetes.io/projected/5d42ee9b-5073-4f40-8a3d-632914f14bbc-kube-api-access-dx6ks\") pod \"nova-cell0-cell-mapping-x5mhp\" (UID: \"5d42ee9b-5073-4f40-8a3d-632914f14bbc\") " pod="openstack/nova-cell0-cell-mapping-x5mhp" Feb 27 08:11:45 crc kubenswrapper[4612]: I0227 08:11:45.509778 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Feb 27 08:11:45 crc kubenswrapper[4612]: I0227 08:11:45.511271 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 27 08:11:45 crc kubenswrapper[4612]: I0227 08:11:45.513712 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Feb 27 08:11:45 crc kubenswrapper[4612]: I0227 08:11:45.585201 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Feb 27 08:11:45 crc kubenswrapper[4612]: I0227 08:11:45.634289 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-x5mhp" Feb 27 08:11:45 crc kubenswrapper[4612]: I0227 08:11:45.641625 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Feb 27 08:11:45 crc kubenswrapper[4612]: I0227 08:11:45.643298 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 27 08:11:45 crc kubenswrapper[4612]: I0227 08:11:45.659302 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Feb 27 08:11:45 crc kubenswrapper[4612]: I0227 08:11:45.674299 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Feb 27 08:11:45 crc kubenswrapper[4612]: I0227 08:11:45.677117 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0a4db194-fc59-4410-9782-698dd03f6efb-logs\") pod \"nova-api-0\" (UID: \"0a4db194-fc59-4410-9782-698dd03f6efb\") " pod="openstack/nova-api-0" Feb 27 08:11:45 crc kubenswrapper[4612]: I0227 08:11:45.677193 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a4db194-fc59-4410-9782-698dd03f6efb-config-data\") pod \"nova-api-0\" (UID: \"0a4db194-fc59-4410-9782-698dd03f6efb\") " pod="openstack/nova-api-0" Feb 27 08:11:45 crc kubenswrapper[4612]: I0227 08:11:45.677302 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a4db194-fc59-4410-9782-698dd03f6efb-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"0a4db194-fc59-4410-9782-698dd03f6efb\") " pod="openstack/nova-api-0" Feb 27 08:11:45 crc kubenswrapper[4612]: I0227 08:11:45.677438 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bh2p6\" (UniqueName: \"kubernetes.io/projected/0a4db194-fc59-4410-9782-698dd03f6efb-kube-api-access-bh2p6\") pod \"nova-api-0\" (UID: \"0a4db194-fc59-4410-9782-698dd03f6efb\") " pod="openstack/nova-api-0" Feb 27 08:11:45 crc kubenswrapper[4612]: I0227 08:11:45.737715 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Feb 27 08:11:45 crc kubenswrapper[4612]: I0227 08:11:45.740226 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 27 08:11:45 crc kubenswrapper[4612]: I0227 08:11:45.749009 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Feb 27 08:11:45 crc kubenswrapper[4612]: I0227 08:11:45.785516 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/348494e3-ef53-4083-866a-da64ad196ed0-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"348494e3-ef53-4083-866a-da64ad196ed0\") " pod="openstack/nova-metadata-0" Feb 27 08:11:45 crc kubenswrapper[4612]: I0227 08:11:45.785827 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-twk4f\" (UniqueName: \"kubernetes.io/projected/348494e3-ef53-4083-866a-da64ad196ed0-kube-api-access-twk4f\") pod \"nova-metadata-0\" (UID: \"348494e3-ef53-4083-866a-da64ad196ed0\") " pod="openstack/nova-metadata-0" Feb 27 08:11:45 crc kubenswrapper[4612]: I0227 08:11:45.785854 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bh2p6\" (UniqueName: \"kubernetes.io/projected/0a4db194-fc59-4410-9782-698dd03f6efb-kube-api-access-bh2p6\") pod \"nova-api-0\" (UID: \"0a4db194-fc59-4410-9782-698dd03f6efb\") " pod="openstack/nova-api-0" Feb 27 08:11:45 crc kubenswrapper[4612]: I0227 08:11:45.785903 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/348494e3-ef53-4083-866a-da64ad196ed0-config-data\") pod \"nova-metadata-0\" (UID: \"348494e3-ef53-4083-866a-da64ad196ed0\") " pod="openstack/nova-metadata-0" Feb 27 08:11:45 crc kubenswrapper[4612]: I0227 08:11:45.785953 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0a4db194-fc59-4410-9782-698dd03f6efb-logs\") pod \"nova-api-0\" (UID: \"0a4db194-fc59-4410-9782-698dd03f6efb\") " pod="openstack/nova-api-0" Feb 27 08:11:45 crc kubenswrapper[4612]: I0227 08:11:45.785983 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a4db194-fc59-4410-9782-698dd03f6efb-config-data\") pod \"nova-api-0\" (UID: \"0a4db194-fc59-4410-9782-698dd03f6efb\") " pod="openstack/nova-api-0" Feb 27 08:11:45 crc kubenswrapper[4612]: I0227 08:11:45.786018 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a4db194-fc59-4410-9782-698dd03f6efb-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"0a4db194-fc59-4410-9782-698dd03f6efb\") " pod="openstack/nova-api-0" Feb 27 08:11:45 crc kubenswrapper[4612]: I0227 08:11:45.786039 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/348494e3-ef53-4083-866a-da64ad196ed0-logs\") pod \"nova-metadata-0\" (UID: \"348494e3-ef53-4083-866a-da64ad196ed0\") " pod="openstack/nova-metadata-0" Feb 27 08:11:45 crc kubenswrapper[4612]: I0227 08:11:45.786587 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0a4db194-fc59-4410-9782-698dd03f6efb-logs\") pod \"nova-api-0\" (UID: \"0a4db194-fc59-4410-9782-698dd03f6efb\") " pod="openstack/nova-api-0" Feb 27 08:11:45 crc kubenswrapper[4612]: I0227 08:11:45.787640 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Feb 27 08:11:45 crc kubenswrapper[4612]: I0227 08:11:45.800182 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a4db194-fc59-4410-9782-698dd03f6efb-config-data\") pod \"nova-api-0\" (UID: \"0a4db194-fc59-4410-9782-698dd03f6efb\") " pod="openstack/nova-api-0" Feb 27 08:11:45 crc kubenswrapper[4612]: I0227 08:11:45.820540 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a4db194-fc59-4410-9782-698dd03f6efb-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"0a4db194-fc59-4410-9782-698dd03f6efb\") " pod="openstack/nova-api-0" Feb 27 08:11:45 crc kubenswrapper[4612]: I0227 08:11:45.853514 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bh2p6\" (UniqueName: \"kubernetes.io/projected/0a4db194-fc59-4410-9782-698dd03f6efb-kube-api-access-bh2p6\") pod \"nova-api-0\" (UID: \"0a4db194-fc59-4410-9782-698dd03f6efb\") " pod="openstack/nova-api-0" Feb 27 08:11:45 crc kubenswrapper[4612]: I0227 08:11:45.879895 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 27 08:11:45 crc kubenswrapper[4612]: I0227 08:11:45.887459 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/348494e3-ef53-4083-866a-da64ad196ed0-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"348494e3-ef53-4083-866a-da64ad196ed0\") " pod="openstack/nova-metadata-0" Feb 27 08:11:45 crc kubenswrapper[4612]: I0227 08:11:45.887491 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-twk4f\" (UniqueName: \"kubernetes.io/projected/348494e3-ef53-4083-866a-da64ad196ed0-kube-api-access-twk4f\") pod \"nova-metadata-0\" (UID: \"348494e3-ef53-4083-866a-da64ad196ed0\") " pod="openstack/nova-metadata-0" Feb 27 08:11:45 crc kubenswrapper[4612]: I0227 08:11:45.887516 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3c4d5bb6-8ca7-41d7-8e30-af438a0e267a-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"3c4d5bb6-8ca7-41d7-8e30-af438a0e267a\") " pod="openstack/nova-scheduler-0" Feb 27 08:11:45 crc kubenswrapper[4612]: I0227 08:11:45.887562 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/348494e3-ef53-4083-866a-da64ad196ed0-config-data\") pod \"nova-metadata-0\" (UID: \"348494e3-ef53-4083-866a-da64ad196ed0\") " pod="openstack/nova-metadata-0" Feb 27 08:11:45 crc kubenswrapper[4612]: I0227 08:11:45.887591 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nn7g6\" (UniqueName: \"kubernetes.io/projected/3c4d5bb6-8ca7-41d7-8e30-af438a0e267a-kube-api-access-nn7g6\") pod \"nova-scheduler-0\" (UID: \"3c4d5bb6-8ca7-41d7-8e30-af438a0e267a\") " pod="openstack/nova-scheduler-0" Feb 27 08:11:45 crc kubenswrapper[4612]: I0227 08:11:45.887626 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3c4d5bb6-8ca7-41d7-8e30-af438a0e267a-config-data\") pod \"nova-scheduler-0\" (UID: \"3c4d5bb6-8ca7-41d7-8e30-af438a0e267a\") " pod="openstack/nova-scheduler-0" Feb 27 08:11:45 crc kubenswrapper[4612]: I0227 08:11:45.887676 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/348494e3-ef53-4083-866a-da64ad196ed0-logs\") pod \"nova-metadata-0\" (UID: \"348494e3-ef53-4083-866a-da64ad196ed0\") " pod="openstack/nova-metadata-0" Feb 27 08:11:45 crc kubenswrapper[4612]: I0227 08:11:45.888028 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/348494e3-ef53-4083-866a-da64ad196ed0-logs\") pod \"nova-metadata-0\" (UID: \"348494e3-ef53-4083-866a-da64ad196ed0\") " pod="openstack/nova-metadata-0" Feb 27 08:11:45 crc kubenswrapper[4612]: I0227 08:11:45.892925 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-bccf8f775-kgn5g"] Feb 27 08:11:45 crc kubenswrapper[4612]: I0227 08:11:45.894371 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bccf8f775-kgn5g" Feb 27 08:11:45 crc kubenswrapper[4612]: I0227 08:11:45.908156 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/348494e3-ef53-4083-866a-da64ad196ed0-config-data\") pod \"nova-metadata-0\" (UID: \"348494e3-ef53-4083-866a-da64ad196ed0\") " pod="openstack/nova-metadata-0" Feb 27 08:11:45 crc kubenswrapper[4612]: I0227 08:11:45.908928 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/348494e3-ef53-4083-866a-da64ad196ed0-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"348494e3-ef53-4083-866a-da64ad196ed0\") " pod="openstack/nova-metadata-0" Feb 27 08:11:45 crc kubenswrapper[4612]: I0227 08:11:45.912028 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Feb 27 08:11:45 crc kubenswrapper[4612]: I0227 08:11:45.913104 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Feb 27 08:11:45 crc kubenswrapper[4612]: I0227 08:11:45.916087 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Feb 27 08:11:45 crc kubenswrapper[4612]: I0227 08:11:45.933133 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-twk4f\" (UniqueName: \"kubernetes.io/projected/348494e3-ef53-4083-866a-da64ad196ed0-kube-api-access-twk4f\") pod \"nova-metadata-0\" (UID: \"348494e3-ef53-4083-866a-da64ad196ed0\") " pod="openstack/nova-metadata-0" Feb 27 08:11:45 crc kubenswrapper[4612]: I0227 08:11:45.936086 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-bccf8f775-kgn5g"] Feb 27 08:11:45 crc kubenswrapper[4612]: I0227 08:11:45.980986 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 27 08:11:45 crc kubenswrapper[4612]: I0227 08:11:45.995004 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nn7g6\" (UniqueName: \"kubernetes.io/projected/3c4d5bb6-8ca7-41d7-8e30-af438a0e267a-kube-api-access-nn7g6\") pod \"nova-scheduler-0\" (UID: \"3c4d5bb6-8ca7-41d7-8e30-af438a0e267a\") " pod="openstack/nova-scheduler-0" Feb 27 08:11:45 crc kubenswrapper[4612]: I0227 08:11:45.995072 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lj2n5\" (UniqueName: \"kubernetes.io/projected/35ff05eb-318f-455c-85a9-7fc07a444bf6-kube-api-access-lj2n5\") pod \"dnsmasq-dns-bccf8f775-kgn5g\" (UID: \"35ff05eb-318f-455c-85a9-7fc07a444bf6\") " pod="openstack/dnsmasq-dns-bccf8f775-kgn5g" Feb 27 08:11:45 crc kubenswrapper[4612]: I0227 08:11:45.995098 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/35ff05eb-318f-455c-85a9-7fc07a444bf6-config\") pod \"dnsmasq-dns-bccf8f775-kgn5g\" (UID: \"35ff05eb-318f-455c-85a9-7fc07a444bf6\") " pod="openstack/dnsmasq-dns-bccf8f775-kgn5g" Feb 27 08:11:45 crc kubenswrapper[4612]: I0227 08:11:45.995119 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/35ff05eb-318f-455c-85a9-7fc07a444bf6-ovsdbserver-nb\") pod \"dnsmasq-dns-bccf8f775-kgn5g\" (UID: \"35ff05eb-318f-455c-85a9-7fc07a444bf6\") " pod="openstack/dnsmasq-dns-bccf8f775-kgn5g" Feb 27 08:11:45 crc kubenswrapper[4612]: I0227 08:11:45.995143 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sfz69\" (UniqueName: \"kubernetes.io/projected/a96cfd8e-57bd-4780-a843-e30c9d1a0e02-kube-api-access-sfz69\") pod \"nova-cell1-novncproxy-0\" (UID: \"a96cfd8e-57bd-4780-a843-e30c9d1a0e02\") " pod="openstack/nova-cell1-novncproxy-0" Feb 27 08:11:45 crc kubenswrapper[4612]: I0227 08:11:45.995171 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3c4d5bb6-8ca7-41d7-8e30-af438a0e267a-config-data\") pod \"nova-scheduler-0\" (UID: \"3c4d5bb6-8ca7-41d7-8e30-af438a0e267a\") " pod="openstack/nova-scheduler-0" Feb 27 08:11:45 crc kubenswrapper[4612]: I0227 08:11:45.995228 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a96cfd8e-57bd-4780-a843-e30c9d1a0e02-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"a96cfd8e-57bd-4780-a843-e30c9d1a0e02\") " pod="openstack/nova-cell1-novncproxy-0" Feb 27 08:11:45 crc kubenswrapper[4612]: I0227 08:11:45.995267 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/35ff05eb-318f-455c-85a9-7fc07a444bf6-ovsdbserver-sb\") pod \"dnsmasq-dns-bccf8f775-kgn5g\" (UID: \"35ff05eb-318f-455c-85a9-7fc07a444bf6\") " pod="openstack/dnsmasq-dns-bccf8f775-kgn5g" Feb 27 08:11:45 crc kubenswrapper[4612]: I0227 08:11:45.995318 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/35ff05eb-318f-455c-85a9-7fc07a444bf6-dns-swift-storage-0\") pod \"dnsmasq-dns-bccf8f775-kgn5g\" (UID: \"35ff05eb-318f-455c-85a9-7fc07a444bf6\") " pod="openstack/dnsmasq-dns-bccf8f775-kgn5g" Feb 27 08:11:45 crc kubenswrapper[4612]: I0227 08:11:45.995359 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/35ff05eb-318f-455c-85a9-7fc07a444bf6-dns-svc\") pod \"dnsmasq-dns-bccf8f775-kgn5g\" (UID: \"35ff05eb-318f-455c-85a9-7fc07a444bf6\") " pod="openstack/dnsmasq-dns-bccf8f775-kgn5g" Feb 27 08:11:45 crc kubenswrapper[4612]: I0227 08:11:45.995378 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a96cfd8e-57bd-4780-a843-e30c9d1a0e02-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"a96cfd8e-57bd-4780-a843-e30c9d1a0e02\") " pod="openstack/nova-cell1-novncproxy-0" Feb 27 08:11:45 crc kubenswrapper[4612]: I0227 08:11:45.995434 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3c4d5bb6-8ca7-41d7-8e30-af438a0e267a-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"3c4d5bb6-8ca7-41d7-8e30-af438a0e267a\") " pod="openstack/nova-scheduler-0" Feb 27 08:11:45 crc kubenswrapper[4612]: I0227 08:11:45.995818 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Feb 27 08:11:46 crc kubenswrapper[4612]: I0227 08:11:46.029941 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3c4d5bb6-8ca7-41d7-8e30-af438a0e267a-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"3c4d5bb6-8ca7-41d7-8e30-af438a0e267a\") " pod="openstack/nova-scheduler-0" Feb 27 08:11:46 crc kubenswrapper[4612]: I0227 08:11:46.030043 4612 patch_prober.go:28] interesting pod/machine-config-daemon-924vb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 27 08:11:46 crc kubenswrapper[4612]: I0227 08:11:46.030074 4612 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 27 08:11:46 crc kubenswrapper[4612]: I0227 08:11:46.030118 4612 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-924vb" Feb 27 08:11:46 crc kubenswrapper[4612]: I0227 08:11:46.034395 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nn7g6\" (UniqueName: \"kubernetes.io/projected/3c4d5bb6-8ca7-41d7-8e30-af438a0e267a-kube-api-access-nn7g6\") pod \"nova-scheduler-0\" (UID: \"3c4d5bb6-8ca7-41d7-8e30-af438a0e267a\") " pod="openstack/nova-scheduler-0" Feb 27 08:11:46 crc kubenswrapper[4612]: I0227 08:11:46.066338 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3c4d5bb6-8ca7-41d7-8e30-af438a0e267a-config-data\") pod \"nova-scheduler-0\" (UID: \"3c4d5bb6-8ca7-41d7-8e30-af438a0e267a\") " pod="openstack/nova-scheduler-0" Feb 27 08:11:46 crc kubenswrapper[4612]: I0227 08:11:46.097720 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/35ff05eb-318f-455c-85a9-7fc07a444bf6-ovsdbserver-sb\") pod \"dnsmasq-dns-bccf8f775-kgn5g\" (UID: \"35ff05eb-318f-455c-85a9-7fc07a444bf6\") " pod="openstack/dnsmasq-dns-bccf8f775-kgn5g" Feb 27 08:11:46 crc kubenswrapper[4612]: I0227 08:11:46.098032 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/35ff05eb-318f-455c-85a9-7fc07a444bf6-dns-swift-storage-0\") pod \"dnsmasq-dns-bccf8f775-kgn5g\" (UID: \"35ff05eb-318f-455c-85a9-7fc07a444bf6\") " pod="openstack/dnsmasq-dns-bccf8f775-kgn5g" Feb 27 08:11:46 crc kubenswrapper[4612]: I0227 08:11:46.098143 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/35ff05eb-318f-455c-85a9-7fc07a444bf6-dns-svc\") pod \"dnsmasq-dns-bccf8f775-kgn5g\" (UID: \"35ff05eb-318f-455c-85a9-7fc07a444bf6\") " pod="openstack/dnsmasq-dns-bccf8f775-kgn5g" Feb 27 08:11:46 crc kubenswrapper[4612]: I0227 08:11:46.102592 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a96cfd8e-57bd-4780-a843-e30c9d1a0e02-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"a96cfd8e-57bd-4780-a843-e30c9d1a0e02\") " pod="openstack/nova-cell1-novncproxy-0" Feb 27 08:11:46 crc kubenswrapper[4612]: I0227 08:11:46.103237 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lj2n5\" (UniqueName: \"kubernetes.io/projected/35ff05eb-318f-455c-85a9-7fc07a444bf6-kube-api-access-lj2n5\") pod \"dnsmasq-dns-bccf8f775-kgn5g\" (UID: \"35ff05eb-318f-455c-85a9-7fc07a444bf6\") " pod="openstack/dnsmasq-dns-bccf8f775-kgn5g" Feb 27 08:11:46 crc kubenswrapper[4612]: I0227 08:11:46.103344 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/35ff05eb-318f-455c-85a9-7fc07a444bf6-config\") pod \"dnsmasq-dns-bccf8f775-kgn5g\" (UID: \"35ff05eb-318f-455c-85a9-7fc07a444bf6\") " pod="openstack/dnsmasq-dns-bccf8f775-kgn5g" Feb 27 08:11:46 crc kubenswrapper[4612]: I0227 08:11:46.103464 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/35ff05eb-318f-455c-85a9-7fc07a444bf6-ovsdbserver-nb\") pod \"dnsmasq-dns-bccf8f775-kgn5g\" (UID: \"35ff05eb-318f-455c-85a9-7fc07a444bf6\") " pod="openstack/dnsmasq-dns-bccf8f775-kgn5g" Feb 27 08:11:46 crc kubenswrapper[4612]: I0227 08:11:46.103628 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sfz69\" (UniqueName: \"kubernetes.io/projected/a96cfd8e-57bd-4780-a843-e30c9d1a0e02-kube-api-access-sfz69\") pod \"nova-cell1-novncproxy-0\" (UID: \"a96cfd8e-57bd-4780-a843-e30c9d1a0e02\") " pod="openstack/nova-cell1-novncproxy-0" Feb 27 08:11:46 crc kubenswrapper[4612]: I0227 08:11:46.103781 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a96cfd8e-57bd-4780-a843-e30c9d1a0e02-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"a96cfd8e-57bd-4780-a843-e30c9d1a0e02\") " pod="openstack/nova-cell1-novncproxy-0" Feb 27 08:11:46 crc kubenswrapper[4612]: I0227 08:11:46.106848 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/35ff05eb-318f-455c-85a9-7fc07a444bf6-config\") pod \"dnsmasq-dns-bccf8f775-kgn5g\" (UID: \"35ff05eb-318f-455c-85a9-7fc07a444bf6\") " pod="openstack/dnsmasq-dns-bccf8f775-kgn5g" Feb 27 08:11:46 crc kubenswrapper[4612]: I0227 08:11:46.107888 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/35ff05eb-318f-455c-85a9-7fc07a444bf6-dns-swift-storage-0\") pod \"dnsmasq-dns-bccf8f775-kgn5g\" (UID: \"35ff05eb-318f-455c-85a9-7fc07a444bf6\") " pod="openstack/dnsmasq-dns-bccf8f775-kgn5g" Feb 27 08:11:46 crc kubenswrapper[4612]: I0227 08:11:46.108541 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/35ff05eb-318f-455c-85a9-7fc07a444bf6-ovsdbserver-sb\") pod \"dnsmasq-dns-bccf8f775-kgn5g\" (UID: \"35ff05eb-318f-455c-85a9-7fc07a444bf6\") " pod="openstack/dnsmasq-dns-bccf8f775-kgn5g" Feb 27 08:11:46 crc kubenswrapper[4612]: I0227 08:11:46.109122 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/35ff05eb-318f-455c-85a9-7fc07a444bf6-ovsdbserver-nb\") pod \"dnsmasq-dns-bccf8f775-kgn5g\" (UID: \"35ff05eb-318f-455c-85a9-7fc07a444bf6\") " pod="openstack/dnsmasq-dns-bccf8f775-kgn5g" Feb 27 08:11:46 crc kubenswrapper[4612]: I0227 08:11:46.109630 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/35ff05eb-318f-455c-85a9-7fc07a444bf6-dns-svc\") pod \"dnsmasq-dns-bccf8f775-kgn5g\" (UID: \"35ff05eb-318f-455c-85a9-7fc07a444bf6\") " pod="openstack/dnsmasq-dns-bccf8f775-kgn5g" Feb 27 08:11:46 crc kubenswrapper[4612]: I0227 08:11:46.138141 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lj2n5\" (UniqueName: \"kubernetes.io/projected/35ff05eb-318f-455c-85a9-7fc07a444bf6-kube-api-access-lj2n5\") pod \"dnsmasq-dns-bccf8f775-kgn5g\" (UID: \"35ff05eb-318f-455c-85a9-7fc07a444bf6\") " pod="openstack/dnsmasq-dns-bccf8f775-kgn5g" Feb 27 08:11:46 crc kubenswrapper[4612]: I0227 08:11:46.139281 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a96cfd8e-57bd-4780-a843-e30c9d1a0e02-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"a96cfd8e-57bd-4780-a843-e30c9d1a0e02\") " pod="openstack/nova-cell1-novncproxy-0" Feb 27 08:11:46 crc kubenswrapper[4612]: I0227 08:11:46.140874 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sfz69\" (UniqueName: \"kubernetes.io/projected/a96cfd8e-57bd-4780-a843-e30c9d1a0e02-kube-api-access-sfz69\") pod \"nova-cell1-novncproxy-0\" (UID: \"a96cfd8e-57bd-4780-a843-e30c9d1a0e02\") " pod="openstack/nova-cell1-novncproxy-0" Feb 27 08:11:46 crc kubenswrapper[4612]: I0227 08:11:46.143910 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a96cfd8e-57bd-4780-a843-e30c9d1a0e02-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"a96cfd8e-57bd-4780-a843-e30c9d1a0e02\") " pod="openstack/nova-cell1-novncproxy-0" Feb 27 08:11:46 crc kubenswrapper[4612]: I0227 08:11:46.186532 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 27 08:11:46 crc kubenswrapper[4612]: I0227 08:11:46.234889 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bccf8f775-kgn5g" Feb 27 08:11:46 crc kubenswrapper[4612]: I0227 08:11:46.256538 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Feb 27 08:11:46 crc kubenswrapper[4612]: I0227 08:11:46.410925 4612 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"650e97edc5e74c75f461feb02d9fe0c4cbdce4ff0887e7ec2d2f50e6d2e7c100"} pod="openshift-machine-config-operator/machine-config-daemon-924vb" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 27 08:11:46 crc kubenswrapper[4612]: I0227 08:11:46.411004 4612 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" containerName="machine-config-daemon" containerID="cri-o://650e97edc5e74c75f461feb02d9fe0c4cbdce4ff0887e7ec2d2f50e6d2e7c100" gracePeriod=600 Feb 27 08:11:46 crc kubenswrapper[4612]: I0227 08:11:46.694605 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Feb 27 08:11:46 crc kubenswrapper[4612]: I0227 08:11:46.783070 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-x5mhp"] Feb 27 08:11:46 crc kubenswrapper[4612]: I0227 08:11:46.907607 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Feb 27 08:11:47 crc kubenswrapper[4612]: I0227 08:11:47.117495 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-bccf8f775-kgn5g"] Feb 27 08:11:47 crc kubenswrapper[4612]: W0227 08:11:47.136067 4612 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3c4d5bb6_8ca7_41d7_8e30_af438a0e267a.slice/crio-3a425120e135e1aa509b9b0ea5561b0641992abe57e59d56dae08a9105f3e8f6 WatchSource:0}: Error finding container 3a425120e135e1aa509b9b0ea5561b0641992abe57e59d56dae08a9105f3e8f6: Status 404 returned error can't find the container with id 3a425120e135e1aa509b9b0ea5561b0641992abe57e59d56dae08a9105f3e8f6 Feb 27 08:11:47 crc kubenswrapper[4612]: I0227 08:11:47.145683 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Feb 27 08:11:47 crc kubenswrapper[4612]: I0227 08:11:47.318738 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Feb 27 08:11:47 crc kubenswrapper[4612]: I0227 08:11:47.446441 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-qckvl"] Feb 27 08:11:47 crc kubenswrapper[4612]: I0227 08:11:47.447597 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-qckvl" Feb 27 08:11:47 crc kubenswrapper[4612]: I0227 08:11:47.460088 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Feb 27 08:11:47 crc kubenswrapper[4612]: I0227 08:11:47.460411 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Feb 27 08:11:47 crc kubenswrapper[4612]: I0227 08:11:47.460411 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-x5mhp" event={"ID":"5d42ee9b-5073-4f40-8a3d-632914f14bbc","Type":"ContainerStarted","Data":"0b4e437518f07bffd24b4f73fca63432cbe5816775bc61ba305365ed6b924fe6"} Feb 27 08:11:47 crc kubenswrapper[4612]: I0227 08:11:47.461220 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-x5mhp" event={"ID":"5d42ee9b-5073-4f40-8a3d-632914f14bbc","Type":"ContainerStarted","Data":"761ac441cf15f1e500a95e92cba8c2bcb83552f5a290c39a1474ca83e24063f5"} Feb 27 08:11:47 crc kubenswrapper[4612]: I0227 08:11:47.464473 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"348494e3-ef53-4083-866a-da64ad196ed0","Type":"ContainerStarted","Data":"69bedad05dd45a75187c7bcb584a4b597d9f61b2b1f06fcc3b33d21e5c60339c"} Feb 27 08:11:47 crc kubenswrapper[4612]: I0227 08:11:47.465621 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"a96cfd8e-57bd-4780-a843-e30c9d1a0e02","Type":"ContainerStarted","Data":"06d0b8f94832cbd72d3834a653f08d2469aceb87280381ac4cbc39f838817f1c"} Feb 27 08:11:47 crc kubenswrapper[4612]: I0227 08:11:47.466392 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"0a4db194-fc59-4410-9782-698dd03f6efb","Type":"ContainerStarted","Data":"619505c020ad44021687ffaf53350d5e157222fc8787daa76d252abd2fae4cec"} Feb 27 08:11:47 crc kubenswrapper[4612]: I0227 08:11:47.467304 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-qckvl"] Feb 27 08:11:47 crc kubenswrapper[4612]: I0227 08:11:47.467470 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"3c4d5bb6-8ca7-41d7-8e30-af438a0e267a","Type":"ContainerStarted","Data":"3a425120e135e1aa509b9b0ea5561b0641992abe57e59d56dae08a9105f3e8f6"} Feb 27 08:11:47 crc kubenswrapper[4612]: I0227 08:11:47.494275 4612 generic.go:334] "Generic (PLEG): container finished" podID="24599373-7adc-4a1b-8bb4-797bf726f43d" containerID="650e97edc5e74c75f461feb02d9fe0c4cbdce4ff0887e7ec2d2f50e6d2e7c100" exitCode=0 Feb 27 08:11:47 crc kubenswrapper[4612]: I0227 08:11:47.494392 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-924vb" event={"ID":"24599373-7adc-4a1b-8bb4-797bf726f43d","Type":"ContainerDied","Data":"650e97edc5e74c75f461feb02d9fe0c4cbdce4ff0887e7ec2d2f50e6d2e7c100"} Feb 27 08:11:47 crc kubenswrapper[4612]: I0227 08:11:47.494424 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-924vb" event={"ID":"24599373-7adc-4a1b-8bb4-797bf726f43d","Type":"ContainerStarted","Data":"7c62ebf970b4c78c5d31a0ce6596e9b398afbc8a6935511ce1b20bcd4f436b08"} Feb 27 08:11:47 crc kubenswrapper[4612]: I0227 08:11:47.494441 4612 scope.go:117] "RemoveContainer" containerID="8e7aef6fa8ec4b3d8f700070d994aef18691e45f133524f54d3ffb63aa703a66" Feb 27 08:11:47 crc kubenswrapper[4612]: I0227 08:11:47.509607 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bccf8f775-kgn5g" event={"ID":"35ff05eb-318f-455c-85a9-7fc07a444bf6","Type":"ContainerStarted","Data":"a8c6e9a5808b6c72f39bf4d07cf08d027cee7be07395b714f4b359bffbb5486b"} Feb 27 08:11:47 crc kubenswrapper[4612]: I0227 08:11:47.540485 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-x5mhp" podStartSLOduration=2.540464071 podStartE2EDuration="2.540464071s" podCreationTimestamp="2026-02-27 08:11:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 08:11:47.48782874 +0000 UTC m=+1365.341758738" watchObservedRunningTime="2026-02-27 08:11:47.540464071 +0000 UTC m=+1365.394394069" Feb 27 08:11:47 crc kubenswrapper[4612]: I0227 08:11:47.545355 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7bca0ea-5f1b-4de7-83d1-54c76f32a813-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-qckvl\" (UID: \"a7bca0ea-5f1b-4de7-83d1-54c76f32a813\") " pod="openstack/nova-cell1-conductor-db-sync-qckvl" Feb 27 08:11:47 crc kubenswrapper[4612]: I0227 08:11:47.545559 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a7bca0ea-5f1b-4de7-83d1-54c76f32a813-scripts\") pod \"nova-cell1-conductor-db-sync-qckvl\" (UID: \"a7bca0ea-5f1b-4de7-83d1-54c76f32a813\") " pod="openstack/nova-cell1-conductor-db-sync-qckvl" Feb 27 08:11:47 crc kubenswrapper[4612]: I0227 08:11:47.545750 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r62rp\" (UniqueName: \"kubernetes.io/projected/a7bca0ea-5f1b-4de7-83d1-54c76f32a813-kube-api-access-r62rp\") pod \"nova-cell1-conductor-db-sync-qckvl\" (UID: \"a7bca0ea-5f1b-4de7-83d1-54c76f32a813\") " pod="openstack/nova-cell1-conductor-db-sync-qckvl" Feb 27 08:11:47 crc kubenswrapper[4612]: I0227 08:11:47.545902 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a7bca0ea-5f1b-4de7-83d1-54c76f32a813-config-data\") pod \"nova-cell1-conductor-db-sync-qckvl\" (UID: \"a7bca0ea-5f1b-4de7-83d1-54c76f32a813\") " pod="openstack/nova-cell1-conductor-db-sync-qckvl" Feb 27 08:11:47 crc kubenswrapper[4612]: I0227 08:11:47.648568 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7bca0ea-5f1b-4de7-83d1-54c76f32a813-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-qckvl\" (UID: \"a7bca0ea-5f1b-4de7-83d1-54c76f32a813\") " pod="openstack/nova-cell1-conductor-db-sync-qckvl" Feb 27 08:11:47 crc kubenswrapper[4612]: I0227 08:11:47.648656 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a7bca0ea-5f1b-4de7-83d1-54c76f32a813-scripts\") pod \"nova-cell1-conductor-db-sync-qckvl\" (UID: \"a7bca0ea-5f1b-4de7-83d1-54c76f32a813\") " pod="openstack/nova-cell1-conductor-db-sync-qckvl" Feb 27 08:11:47 crc kubenswrapper[4612]: I0227 08:11:47.648712 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r62rp\" (UniqueName: \"kubernetes.io/projected/a7bca0ea-5f1b-4de7-83d1-54c76f32a813-kube-api-access-r62rp\") pod \"nova-cell1-conductor-db-sync-qckvl\" (UID: \"a7bca0ea-5f1b-4de7-83d1-54c76f32a813\") " pod="openstack/nova-cell1-conductor-db-sync-qckvl" Feb 27 08:11:47 crc kubenswrapper[4612]: I0227 08:11:47.648777 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a7bca0ea-5f1b-4de7-83d1-54c76f32a813-config-data\") pod \"nova-cell1-conductor-db-sync-qckvl\" (UID: \"a7bca0ea-5f1b-4de7-83d1-54c76f32a813\") " pod="openstack/nova-cell1-conductor-db-sync-qckvl" Feb 27 08:11:47 crc kubenswrapper[4612]: I0227 08:11:47.668516 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7bca0ea-5f1b-4de7-83d1-54c76f32a813-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-qckvl\" (UID: \"a7bca0ea-5f1b-4de7-83d1-54c76f32a813\") " pod="openstack/nova-cell1-conductor-db-sync-qckvl" Feb 27 08:11:47 crc kubenswrapper[4612]: I0227 08:11:47.672610 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a7bca0ea-5f1b-4de7-83d1-54c76f32a813-config-data\") pod \"nova-cell1-conductor-db-sync-qckvl\" (UID: \"a7bca0ea-5f1b-4de7-83d1-54c76f32a813\") " pod="openstack/nova-cell1-conductor-db-sync-qckvl" Feb 27 08:11:47 crc kubenswrapper[4612]: I0227 08:11:47.673270 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r62rp\" (UniqueName: \"kubernetes.io/projected/a7bca0ea-5f1b-4de7-83d1-54c76f32a813-kube-api-access-r62rp\") pod \"nova-cell1-conductor-db-sync-qckvl\" (UID: \"a7bca0ea-5f1b-4de7-83d1-54c76f32a813\") " pod="openstack/nova-cell1-conductor-db-sync-qckvl" Feb 27 08:11:47 crc kubenswrapper[4612]: I0227 08:11:47.681983 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a7bca0ea-5f1b-4de7-83d1-54c76f32a813-scripts\") pod \"nova-cell1-conductor-db-sync-qckvl\" (UID: \"a7bca0ea-5f1b-4de7-83d1-54c76f32a813\") " pod="openstack/nova-cell1-conductor-db-sync-qckvl" Feb 27 08:11:47 crc kubenswrapper[4612]: I0227 08:11:47.925374 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-qckvl" Feb 27 08:11:48 crc kubenswrapper[4612]: I0227 08:11:48.491321 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-qckvl"] Feb 27 08:11:48 crc kubenswrapper[4612]: I0227 08:11:48.581159 4612 generic.go:334] "Generic (PLEG): container finished" podID="35ff05eb-318f-455c-85a9-7fc07a444bf6" containerID="91cae3a3762855836c712786ebf97534e2857c9913837c47a6c8f9edc9488675" exitCode=0 Feb 27 08:11:48 crc kubenswrapper[4612]: I0227 08:11:48.581833 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bccf8f775-kgn5g" event={"ID":"35ff05eb-318f-455c-85a9-7fc07a444bf6","Type":"ContainerDied","Data":"91cae3a3762855836c712786ebf97534e2857c9913837c47a6c8f9edc9488675"} Feb 27 08:11:48 crc kubenswrapper[4612]: I0227 08:11:48.593117 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-qckvl" event={"ID":"a7bca0ea-5f1b-4de7-83d1-54c76f32a813","Type":"ContainerStarted","Data":"8a623cb0e75a9d914cac0e9df1b0b5777df34f99124872844c0028295fddb2ad"} Feb 27 08:11:49 crc kubenswrapper[4612]: I0227 08:11:49.106968 4612 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-6b985547d4-mkfpv" podUID="76009243-16b3-4652-abfe-a6cdba363724" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.152:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.152:8443: connect: connection refused" Feb 27 08:11:49 crc kubenswrapper[4612]: I0227 08:11:49.107289 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-6b985547d4-mkfpv" Feb 27 08:11:49 crc kubenswrapper[4612]: I0227 08:11:49.354384 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Feb 27 08:11:49 crc kubenswrapper[4612]: I0227 08:11:49.371660 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Feb 27 08:11:49 crc kubenswrapper[4612]: I0227 08:11:49.619378 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bccf8f775-kgn5g" event={"ID":"35ff05eb-318f-455c-85a9-7fc07a444bf6","Type":"ContainerStarted","Data":"49cf9107e3aaa1ff37311e366694fb3372a1bd51a3c8e0756d414d85f9f79070"} Feb 27 08:11:49 crc kubenswrapper[4612]: I0227 08:11:49.619473 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-bccf8f775-kgn5g" Feb 27 08:11:49 crc kubenswrapper[4612]: I0227 08:11:49.620655 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-qckvl" event={"ID":"a7bca0ea-5f1b-4de7-83d1-54c76f32a813","Type":"ContainerStarted","Data":"55f1d5da6b2040184cc24875298b865e1ce09f44138529f1feec92112d3f4362"} Feb 27 08:11:49 crc kubenswrapper[4612]: I0227 08:11:49.642880 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-bccf8f775-kgn5g" podStartSLOduration=4.642859608 podStartE2EDuration="4.642859608s" podCreationTimestamp="2026-02-27 08:11:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 08:11:49.640196571 +0000 UTC m=+1367.494126569" watchObservedRunningTime="2026-02-27 08:11:49.642859608 +0000 UTC m=+1367.496789606" Feb 27 08:11:49 crc kubenswrapper[4612]: I0227 08:11:49.671127 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-qckvl" podStartSLOduration=2.671111409 podStartE2EDuration="2.671111409s" podCreationTimestamp="2026-02-27 08:11:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 08:11:49.663357606 +0000 UTC m=+1367.517287594" watchObservedRunningTime="2026-02-27 08:11:49.671111409 +0000 UTC m=+1367.525041407" Feb 27 08:11:51 crc kubenswrapper[4612]: I0227 08:11:51.641189 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"a96cfd8e-57bd-4780-a843-e30c9d1a0e02","Type":"ContainerStarted","Data":"a3f5f545e871de451788fd0c23694fd3101c0a3404666ccabcf72114b3416ed1"} Feb 27 08:11:51 crc kubenswrapper[4612]: I0227 08:11:51.641529 4612 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="a96cfd8e-57bd-4780-a843-e30c9d1a0e02" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://a3f5f545e871de451788fd0c23694fd3101c0a3404666ccabcf72114b3416ed1" gracePeriod=30 Feb 27 08:11:51 crc kubenswrapper[4612]: I0227 08:11:51.644833 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"0a4db194-fc59-4410-9782-698dd03f6efb","Type":"ContainerStarted","Data":"b569723412818ade307aee6c05130a45f30421f2ba063b311d0a32ebd456fcbf"} Feb 27 08:11:51 crc kubenswrapper[4612]: I0227 08:11:51.644919 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"0a4db194-fc59-4410-9782-698dd03f6efb","Type":"ContainerStarted","Data":"0b5812f027deb193e6de875cbc8fb3996e3a7515ce2d952698dfb9311d73d4d1"} Feb 27 08:11:51 crc kubenswrapper[4612]: I0227 08:11:51.653520 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"3c4d5bb6-8ca7-41d7-8e30-af438a0e267a","Type":"ContainerStarted","Data":"b20dc16d14af8f07dddc21f5ce3968bf2ad89f62cb1e10b67ad0a10affdda173"} Feb 27 08:11:51 crc kubenswrapper[4612]: I0227 08:11:51.655376 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"348494e3-ef53-4083-866a-da64ad196ed0","Type":"ContainerStarted","Data":"5176eb0ae2037c2f3ca8b5d7112b1f8d9b20b487b737eed5943cd399ed3c9366"} Feb 27 08:11:51 crc kubenswrapper[4612]: I0227 08:11:51.655804 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"348494e3-ef53-4083-866a-da64ad196ed0","Type":"ContainerStarted","Data":"b14a4c1a1ff6e469587da8b4ccbd09bb39f616ab37922abdfd673e32008cd1b9"} Feb 27 08:11:51 crc kubenswrapper[4612]: I0227 08:11:51.655758 4612 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="348494e3-ef53-4083-866a-da64ad196ed0" containerName="nova-metadata-metadata" containerID="cri-o://5176eb0ae2037c2f3ca8b5d7112b1f8d9b20b487b737eed5943cd399ed3c9366" gracePeriod=30 Feb 27 08:11:51 crc kubenswrapper[4612]: I0227 08:11:51.655538 4612 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="348494e3-ef53-4083-866a-da64ad196ed0" containerName="nova-metadata-log" containerID="cri-o://b14a4c1a1ff6e469587da8b4ccbd09bb39f616ab37922abdfd673e32008cd1b9" gracePeriod=30 Feb 27 08:11:51 crc kubenswrapper[4612]: I0227 08:11:51.663023 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=3.182524991 podStartE2EDuration="6.662989143s" podCreationTimestamp="2026-02-27 08:11:45 +0000 UTC" firstStartedPulling="2026-02-27 08:11:47.336246867 +0000 UTC m=+1365.190176865" lastFinishedPulling="2026-02-27 08:11:50.816710999 +0000 UTC m=+1368.670641017" observedRunningTime="2026-02-27 08:11:51.658256287 +0000 UTC m=+1369.512186295" watchObservedRunningTime="2026-02-27 08:11:51.662989143 +0000 UTC m=+1369.516919161" Feb 27 08:11:51 crc kubenswrapper[4612]: I0227 08:11:51.688941 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.699513419 podStartE2EDuration="6.688923318s" podCreationTimestamp="2026-02-27 08:11:45 +0000 UTC" firstStartedPulling="2026-02-27 08:11:46.828754172 +0000 UTC m=+1364.682684170" lastFinishedPulling="2026-02-27 08:11:50.818164071 +0000 UTC m=+1368.672094069" observedRunningTime="2026-02-27 08:11:51.683642636 +0000 UTC m=+1369.537572634" watchObservedRunningTime="2026-02-27 08:11:51.688923318 +0000 UTC m=+1369.542853316" Feb 27 08:11:51 crc kubenswrapper[4612]: I0227 08:11:51.700300 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=3.012920739 podStartE2EDuration="6.700286293s" podCreationTimestamp="2026-02-27 08:11:45 +0000 UTC" firstStartedPulling="2026-02-27 08:11:47.141056981 +0000 UTC m=+1364.994986979" lastFinishedPulling="2026-02-27 08:11:50.828422535 +0000 UTC m=+1368.682352533" observedRunningTime="2026-02-27 08:11:51.696240597 +0000 UTC m=+1369.550170595" watchObservedRunningTime="2026-02-27 08:11:51.700286293 +0000 UTC m=+1369.554216291" Feb 27 08:11:52 crc kubenswrapper[4612]: I0227 08:11:52.278842 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 27 08:11:52 crc kubenswrapper[4612]: I0227 08:11:52.369720 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/348494e3-ef53-4083-866a-da64ad196ed0-logs\") pod \"348494e3-ef53-4083-866a-da64ad196ed0\" (UID: \"348494e3-ef53-4083-866a-da64ad196ed0\") " Feb 27 08:11:52 crc kubenswrapper[4612]: I0227 08:11:52.369768 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-twk4f\" (UniqueName: \"kubernetes.io/projected/348494e3-ef53-4083-866a-da64ad196ed0-kube-api-access-twk4f\") pod \"348494e3-ef53-4083-866a-da64ad196ed0\" (UID: \"348494e3-ef53-4083-866a-da64ad196ed0\") " Feb 27 08:11:52 crc kubenswrapper[4612]: I0227 08:11:52.369816 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/348494e3-ef53-4083-866a-da64ad196ed0-combined-ca-bundle\") pod \"348494e3-ef53-4083-866a-da64ad196ed0\" (UID: \"348494e3-ef53-4083-866a-da64ad196ed0\") " Feb 27 08:11:52 crc kubenswrapper[4612]: I0227 08:11:52.370032 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/348494e3-ef53-4083-866a-da64ad196ed0-config-data\") pod \"348494e3-ef53-4083-866a-da64ad196ed0\" (UID: \"348494e3-ef53-4083-866a-da64ad196ed0\") " Feb 27 08:11:52 crc kubenswrapper[4612]: I0227 08:11:52.370288 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/348494e3-ef53-4083-866a-da64ad196ed0-logs" (OuterVolumeSpecName: "logs") pod "348494e3-ef53-4083-866a-da64ad196ed0" (UID: "348494e3-ef53-4083-866a-da64ad196ed0"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 08:11:52 crc kubenswrapper[4612]: I0227 08:11:52.371012 4612 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/348494e3-ef53-4083-866a-da64ad196ed0-logs\") on node \"crc\" DevicePath \"\"" Feb 27 08:11:52 crc kubenswrapper[4612]: I0227 08:11:52.392651 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/348494e3-ef53-4083-866a-da64ad196ed0-kube-api-access-twk4f" (OuterVolumeSpecName: "kube-api-access-twk4f") pod "348494e3-ef53-4083-866a-da64ad196ed0" (UID: "348494e3-ef53-4083-866a-da64ad196ed0"). InnerVolumeSpecName "kube-api-access-twk4f". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 08:11:52 crc kubenswrapper[4612]: I0227 08:11:52.400147 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/348494e3-ef53-4083-866a-da64ad196ed0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "348494e3-ef53-4083-866a-da64ad196ed0" (UID: "348494e3-ef53-4083-866a-da64ad196ed0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:11:52 crc kubenswrapper[4612]: I0227 08:11:52.405424 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/348494e3-ef53-4083-866a-da64ad196ed0-config-data" (OuterVolumeSpecName: "config-data") pod "348494e3-ef53-4083-866a-da64ad196ed0" (UID: "348494e3-ef53-4083-866a-da64ad196ed0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:11:52 crc kubenswrapper[4612]: I0227 08:11:52.472834 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-twk4f\" (UniqueName: \"kubernetes.io/projected/348494e3-ef53-4083-866a-da64ad196ed0-kube-api-access-twk4f\") on node \"crc\" DevicePath \"\"" Feb 27 08:11:52 crc kubenswrapper[4612]: I0227 08:11:52.472868 4612 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/348494e3-ef53-4083-866a-da64ad196ed0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 08:11:52 crc kubenswrapper[4612]: I0227 08:11:52.472881 4612 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/348494e3-ef53-4083-866a-da64ad196ed0-config-data\") on node \"crc\" DevicePath \"\"" Feb 27 08:11:52 crc kubenswrapper[4612]: I0227 08:11:52.665858 4612 generic.go:334] "Generic (PLEG): container finished" podID="348494e3-ef53-4083-866a-da64ad196ed0" containerID="5176eb0ae2037c2f3ca8b5d7112b1f8d9b20b487b737eed5943cd399ed3c9366" exitCode=0 Feb 27 08:11:52 crc kubenswrapper[4612]: I0227 08:11:52.665894 4612 generic.go:334] "Generic (PLEG): container finished" podID="348494e3-ef53-4083-866a-da64ad196ed0" containerID="b14a4c1a1ff6e469587da8b4ccbd09bb39f616ab37922abdfd673e32008cd1b9" exitCode=143 Feb 27 08:11:52 crc kubenswrapper[4612]: I0227 08:11:52.666868 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 27 08:11:52 crc kubenswrapper[4612]: I0227 08:11:52.667092 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"348494e3-ef53-4083-866a-da64ad196ed0","Type":"ContainerDied","Data":"5176eb0ae2037c2f3ca8b5d7112b1f8d9b20b487b737eed5943cd399ed3c9366"} Feb 27 08:11:52 crc kubenswrapper[4612]: I0227 08:11:52.667160 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"348494e3-ef53-4083-866a-da64ad196ed0","Type":"ContainerDied","Data":"b14a4c1a1ff6e469587da8b4ccbd09bb39f616ab37922abdfd673e32008cd1b9"} Feb 27 08:11:52 crc kubenswrapper[4612]: I0227 08:11:52.667189 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"348494e3-ef53-4083-866a-da64ad196ed0","Type":"ContainerDied","Data":"69bedad05dd45a75187c7bcb584a4b597d9f61b2b1f06fcc3b33d21e5c60339c"} Feb 27 08:11:52 crc kubenswrapper[4612]: I0227 08:11:52.667213 4612 scope.go:117] "RemoveContainer" containerID="5176eb0ae2037c2f3ca8b5d7112b1f8d9b20b487b737eed5943cd399ed3c9366" Feb 27 08:11:52 crc kubenswrapper[4612]: I0227 08:11:52.700205 4612 scope.go:117] "RemoveContainer" containerID="b14a4c1a1ff6e469587da8b4ccbd09bb39f616ab37922abdfd673e32008cd1b9" Feb 27 08:11:52 crc kubenswrapper[4612]: I0227 08:11:52.709378 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Feb 27 08:11:52 crc kubenswrapper[4612]: I0227 08:11:52.723222 4612 scope.go:117] "RemoveContainer" containerID="5176eb0ae2037c2f3ca8b5d7112b1f8d9b20b487b737eed5943cd399ed3c9366" Feb 27 08:11:52 crc kubenswrapper[4612]: E0227 08:11:52.727665 4612 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5176eb0ae2037c2f3ca8b5d7112b1f8d9b20b487b737eed5943cd399ed3c9366\": container with ID starting with 5176eb0ae2037c2f3ca8b5d7112b1f8d9b20b487b737eed5943cd399ed3c9366 not found: ID does not exist" containerID="5176eb0ae2037c2f3ca8b5d7112b1f8d9b20b487b737eed5943cd399ed3c9366" Feb 27 08:11:52 crc kubenswrapper[4612]: I0227 08:11:52.727787 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5176eb0ae2037c2f3ca8b5d7112b1f8d9b20b487b737eed5943cd399ed3c9366"} err="failed to get container status \"5176eb0ae2037c2f3ca8b5d7112b1f8d9b20b487b737eed5943cd399ed3c9366\": rpc error: code = NotFound desc = could not find container \"5176eb0ae2037c2f3ca8b5d7112b1f8d9b20b487b737eed5943cd399ed3c9366\": container with ID starting with 5176eb0ae2037c2f3ca8b5d7112b1f8d9b20b487b737eed5943cd399ed3c9366 not found: ID does not exist" Feb 27 08:11:52 crc kubenswrapper[4612]: I0227 08:11:52.727821 4612 scope.go:117] "RemoveContainer" containerID="b14a4c1a1ff6e469587da8b4ccbd09bb39f616ab37922abdfd673e32008cd1b9" Feb 27 08:11:52 crc kubenswrapper[4612]: E0227 08:11:52.728289 4612 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b14a4c1a1ff6e469587da8b4ccbd09bb39f616ab37922abdfd673e32008cd1b9\": container with ID starting with b14a4c1a1ff6e469587da8b4ccbd09bb39f616ab37922abdfd673e32008cd1b9 not found: ID does not exist" containerID="b14a4c1a1ff6e469587da8b4ccbd09bb39f616ab37922abdfd673e32008cd1b9" Feb 27 08:11:52 crc kubenswrapper[4612]: I0227 08:11:52.728335 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b14a4c1a1ff6e469587da8b4ccbd09bb39f616ab37922abdfd673e32008cd1b9"} err="failed to get container status \"b14a4c1a1ff6e469587da8b4ccbd09bb39f616ab37922abdfd673e32008cd1b9\": rpc error: code = NotFound desc = could not find container \"b14a4c1a1ff6e469587da8b4ccbd09bb39f616ab37922abdfd673e32008cd1b9\": container with ID starting with b14a4c1a1ff6e469587da8b4ccbd09bb39f616ab37922abdfd673e32008cd1b9 not found: ID does not exist" Feb 27 08:11:52 crc kubenswrapper[4612]: I0227 08:11:52.728363 4612 scope.go:117] "RemoveContainer" containerID="5176eb0ae2037c2f3ca8b5d7112b1f8d9b20b487b737eed5943cd399ed3c9366" Feb 27 08:11:52 crc kubenswrapper[4612]: I0227 08:11:52.729986 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5176eb0ae2037c2f3ca8b5d7112b1f8d9b20b487b737eed5943cd399ed3c9366"} err="failed to get container status \"5176eb0ae2037c2f3ca8b5d7112b1f8d9b20b487b737eed5943cd399ed3c9366\": rpc error: code = NotFound desc = could not find container \"5176eb0ae2037c2f3ca8b5d7112b1f8d9b20b487b737eed5943cd399ed3c9366\": container with ID starting with 5176eb0ae2037c2f3ca8b5d7112b1f8d9b20b487b737eed5943cd399ed3c9366 not found: ID does not exist" Feb 27 08:11:52 crc kubenswrapper[4612]: I0227 08:11:52.730015 4612 scope.go:117] "RemoveContainer" containerID="b14a4c1a1ff6e469587da8b4ccbd09bb39f616ab37922abdfd673e32008cd1b9" Feb 27 08:11:52 crc kubenswrapper[4612]: I0227 08:11:52.730582 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b14a4c1a1ff6e469587da8b4ccbd09bb39f616ab37922abdfd673e32008cd1b9"} err="failed to get container status \"b14a4c1a1ff6e469587da8b4ccbd09bb39f616ab37922abdfd673e32008cd1b9\": rpc error: code = NotFound desc = could not find container \"b14a4c1a1ff6e469587da8b4ccbd09bb39f616ab37922abdfd673e32008cd1b9\": container with ID starting with b14a4c1a1ff6e469587da8b4ccbd09bb39f616ab37922abdfd673e32008cd1b9 not found: ID does not exist" Feb 27 08:11:52 crc kubenswrapper[4612]: I0227 08:11:52.737278 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Feb 27 08:11:52 crc kubenswrapper[4612]: I0227 08:11:52.758618 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Feb 27 08:11:52 crc kubenswrapper[4612]: E0227 08:11:52.759093 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="348494e3-ef53-4083-866a-da64ad196ed0" containerName="nova-metadata-log" Feb 27 08:11:52 crc kubenswrapper[4612]: I0227 08:11:52.759116 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="348494e3-ef53-4083-866a-da64ad196ed0" containerName="nova-metadata-log" Feb 27 08:11:52 crc kubenswrapper[4612]: E0227 08:11:52.759145 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="348494e3-ef53-4083-866a-da64ad196ed0" containerName="nova-metadata-metadata" Feb 27 08:11:52 crc kubenswrapper[4612]: I0227 08:11:52.759154 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="348494e3-ef53-4083-866a-da64ad196ed0" containerName="nova-metadata-metadata" Feb 27 08:11:52 crc kubenswrapper[4612]: I0227 08:11:52.759392 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="348494e3-ef53-4083-866a-da64ad196ed0" containerName="nova-metadata-log" Feb 27 08:11:52 crc kubenswrapper[4612]: I0227 08:11:52.759426 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="348494e3-ef53-4083-866a-da64ad196ed0" containerName="nova-metadata-metadata" Feb 27 08:11:52 crc kubenswrapper[4612]: I0227 08:11:52.760924 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 27 08:11:52 crc kubenswrapper[4612]: I0227 08:11:52.766459 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Feb 27 08:11:52 crc kubenswrapper[4612]: I0227 08:11:52.788028 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Feb 27 08:11:52 crc kubenswrapper[4612]: I0227 08:11:52.821127 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Feb 27 08:11:52 crc kubenswrapper[4612]: I0227 08:11:52.880287 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="348494e3-ef53-4083-866a-da64ad196ed0" path="/var/lib/kubelet/pods/348494e3-ef53-4083-866a-da64ad196ed0/volumes" Feb 27 08:11:52 crc kubenswrapper[4612]: I0227 08:11:52.884166 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/46b77ad9-d0a9-40de-a765-47437a44d6a6-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"46b77ad9-d0a9-40de-a765-47437a44d6a6\") " pod="openstack/nova-metadata-0" Feb 27 08:11:52 crc kubenswrapper[4612]: I0227 08:11:52.884219 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/46b77ad9-d0a9-40de-a765-47437a44d6a6-config-data\") pod \"nova-metadata-0\" (UID: \"46b77ad9-d0a9-40de-a765-47437a44d6a6\") " pod="openstack/nova-metadata-0" Feb 27 08:11:52 crc kubenswrapper[4612]: I0227 08:11:52.884242 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46b77ad9-d0a9-40de-a765-47437a44d6a6-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"46b77ad9-d0a9-40de-a765-47437a44d6a6\") " pod="openstack/nova-metadata-0" Feb 27 08:11:52 crc kubenswrapper[4612]: I0227 08:11:52.884265 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/46b77ad9-d0a9-40de-a765-47437a44d6a6-logs\") pod \"nova-metadata-0\" (UID: \"46b77ad9-d0a9-40de-a765-47437a44d6a6\") " pod="openstack/nova-metadata-0" Feb 27 08:11:52 crc kubenswrapper[4612]: I0227 08:11:52.884428 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zd9sr\" (UniqueName: \"kubernetes.io/projected/46b77ad9-d0a9-40de-a765-47437a44d6a6-kube-api-access-zd9sr\") pod \"nova-metadata-0\" (UID: \"46b77ad9-d0a9-40de-a765-47437a44d6a6\") " pod="openstack/nova-metadata-0" Feb 27 08:11:52 crc kubenswrapper[4612]: I0227 08:11:52.987578 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/46b77ad9-d0a9-40de-a765-47437a44d6a6-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"46b77ad9-d0a9-40de-a765-47437a44d6a6\") " pod="openstack/nova-metadata-0" Feb 27 08:11:52 crc kubenswrapper[4612]: I0227 08:11:52.987679 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/46b77ad9-d0a9-40de-a765-47437a44d6a6-config-data\") pod \"nova-metadata-0\" (UID: \"46b77ad9-d0a9-40de-a765-47437a44d6a6\") " pod="openstack/nova-metadata-0" Feb 27 08:11:52 crc kubenswrapper[4612]: I0227 08:11:52.987716 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46b77ad9-d0a9-40de-a765-47437a44d6a6-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"46b77ad9-d0a9-40de-a765-47437a44d6a6\") " pod="openstack/nova-metadata-0" Feb 27 08:11:52 crc kubenswrapper[4612]: I0227 08:11:52.987738 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/46b77ad9-d0a9-40de-a765-47437a44d6a6-logs\") pod \"nova-metadata-0\" (UID: \"46b77ad9-d0a9-40de-a765-47437a44d6a6\") " pod="openstack/nova-metadata-0" Feb 27 08:11:52 crc kubenswrapper[4612]: I0227 08:11:52.987761 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zd9sr\" (UniqueName: \"kubernetes.io/projected/46b77ad9-d0a9-40de-a765-47437a44d6a6-kube-api-access-zd9sr\") pod \"nova-metadata-0\" (UID: \"46b77ad9-d0a9-40de-a765-47437a44d6a6\") " pod="openstack/nova-metadata-0" Feb 27 08:11:52 crc kubenswrapper[4612]: I0227 08:11:52.988450 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/46b77ad9-d0a9-40de-a765-47437a44d6a6-logs\") pod \"nova-metadata-0\" (UID: \"46b77ad9-d0a9-40de-a765-47437a44d6a6\") " pod="openstack/nova-metadata-0" Feb 27 08:11:52 crc kubenswrapper[4612]: I0227 08:11:52.991822 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/46b77ad9-d0a9-40de-a765-47437a44d6a6-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"46b77ad9-d0a9-40de-a765-47437a44d6a6\") " pod="openstack/nova-metadata-0" Feb 27 08:11:53 crc kubenswrapper[4612]: I0227 08:11:52.993893 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/46b77ad9-d0a9-40de-a765-47437a44d6a6-config-data\") pod \"nova-metadata-0\" (UID: \"46b77ad9-d0a9-40de-a765-47437a44d6a6\") " pod="openstack/nova-metadata-0" Feb 27 08:11:53 crc kubenswrapper[4612]: I0227 08:11:52.998498 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46b77ad9-d0a9-40de-a765-47437a44d6a6-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"46b77ad9-d0a9-40de-a765-47437a44d6a6\") " pod="openstack/nova-metadata-0" Feb 27 08:11:53 crc kubenswrapper[4612]: I0227 08:11:53.014994 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zd9sr\" (UniqueName: \"kubernetes.io/projected/46b77ad9-d0a9-40de-a765-47437a44d6a6-kube-api-access-zd9sr\") pod \"nova-metadata-0\" (UID: \"46b77ad9-d0a9-40de-a765-47437a44d6a6\") " pod="openstack/nova-metadata-0" Feb 27 08:11:53 crc kubenswrapper[4612]: I0227 08:11:53.111332 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 27 08:11:53 crc kubenswrapper[4612]: W0227 08:11:53.615399 4612 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod46b77ad9_d0a9_40de_a765_47437a44d6a6.slice/crio-0f8f94a59ccd9454f6b463221edaa8160d0ecf4ad50b9ed51a11fd00e5233a49 WatchSource:0}: Error finding container 0f8f94a59ccd9454f6b463221edaa8160d0ecf4ad50b9ed51a11fd00e5233a49: Status 404 returned error can't find the container with id 0f8f94a59ccd9454f6b463221edaa8160d0ecf4ad50b9ed51a11fd00e5233a49 Feb 27 08:11:53 crc kubenswrapper[4612]: I0227 08:11:53.616003 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Feb 27 08:11:53 crc kubenswrapper[4612]: I0227 08:11:53.679562 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"46b77ad9-d0a9-40de-a765-47437a44d6a6","Type":"ContainerStarted","Data":"0f8f94a59ccd9454f6b463221edaa8160d0ecf4ad50b9ed51a11fd00e5233a49"} Feb 27 08:11:54 crc kubenswrapper[4612]: I0227 08:11:54.718461 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"46b77ad9-d0a9-40de-a765-47437a44d6a6","Type":"ContainerStarted","Data":"1f0880b86bbc77a2d82a1ba76d4db6d467630a0d9fd239f55fcd2c881776f56e"} Feb 27 08:11:54 crc kubenswrapper[4612]: I0227 08:11:54.719158 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"46b77ad9-d0a9-40de-a765-47437a44d6a6","Type":"ContainerStarted","Data":"3b1ed2ade3e135a2f5916e902cb4d16e90bbf2549b3c6d6480cfa9fbe2f53bd5"} Feb 27 08:11:55 crc kubenswrapper[4612]: I0227 08:11:55.561685 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6b985547d4-mkfpv" Feb 27 08:11:55 crc kubenswrapper[4612]: I0227 08:11:55.586729 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=3.586711004 podStartE2EDuration="3.586711004s" podCreationTimestamp="2026-02-27 08:11:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 08:11:54.74528638 +0000 UTC m=+1372.599216388" watchObservedRunningTime="2026-02-27 08:11:55.586711004 +0000 UTC m=+1373.440641002" Feb 27 08:11:55 crc kubenswrapper[4612]: I0227 08:11:55.645266 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/76009243-16b3-4652-abfe-a6cdba363724-horizon-tls-certs\") pod \"76009243-16b3-4652-abfe-a6cdba363724\" (UID: \"76009243-16b3-4652-abfe-a6cdba363724\") " Feb 27 08:11:55 crc kubenswrapper[4612]: I0227 08:11:55.645316 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/76009243-16b3-4652-abfe-a6cdba363724-horizon-secret-key\") pod \"76009243-16b3-4652-abfe-a6cdba363724\" (UID: \"76009243-16b3-4652-abfe-a6cdba363724\") " Feb 27 08:11:55 crc kubenswrapper[4612]: I0227 08:11:55.645343 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9mg4q\" (UniqueName: \"kubernetes.io/projected/76009243-16b3-4652-abfe-a6cdba363724-kube-api-access-9mg4q\") pod \"76009243-16b3-4652-abfe-a6cdba363724\" (UID: \"76009243-16b3-4652-abfe-a6cdba363724\") " Feb 27 08:11:55 crc kubenswrapper[4612]: I0227 08:11:55.645408 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/76009243-16b3-4652-abfe-a6cdba363724-logs\") pod \"76009243-16b3-4652-abfe-a6cdba363724\" (UID: \"76009243-16b3-4652-abfe-a6cdba363724\") " Feb 27 08:11:55 crc kubenswrapper[4612]: I0227 08:11:55.645531 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/76009243-16b3-4652-abfe-a6cdba363724-config-data\") pod \"76009243-16b3-4652-abfe-a6cdba363724\" (UID: \"76009243-16b3-4652-abfe-a6cdba363724\") " Feb 27 08:11:55 crc kubenswrapper[4612]: I0227 08:11:55.645572 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/76009243-16b3-4652-abfe-a6cdba363724-scripts\") pod \"76009243-16b3-4652-abfe-a6cdba363724\" (UID: \"76009243-16b3-4652-abfe-a6cdba363724\") " Feb 27 08:11:55 crc kubenswrapper[4612]: I0227 08:11:55.645640 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76009243-16b3-4652-abfe-a6cdba363724-combined-ca-bundle\") pod \"76009243-16b3-4652-abfe-a6cdba363724\" (UID: \"76009243-16b3-4652-abfe-a6cdba363724\") " Feb 27 08:11:55 crc kubenswrapper[4612]: I0227 08:11:55.645983 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/76009243-16b3-4652-abfe-a6cdba363724-logs" (OuterVolumeSpecName: "logs") pod "76009243-16b3-4652-abfe-a6cdba363724" (UID: "76009243-16b3-4652-abfe-a6cdba363724"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 08:11:55 crc kubenswrapper[4612]: I0227 08:11:55.651486 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/76009243-16b3-4652-abfe-a6cdba363724-kube-api-access-9mg4q" (OuterVolumeSpecName: "kube-api-access-9mg4q") pod "76009243-16b3-4652-abfe-a6cdba363724" (UID: "76009243-16b3-4652-abfe-a6cdba363724"). InnerVolumeSpecName "kube-api-access-9mg4q". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 08:11:55 crc kubenswrapper[4612]: I0227 08:11:55.652482 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/76009243-16b3-4652-abfe-a6cdba363724-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "76009243-16b3-4652-abfe-a6cdba363724" (UID: "76009243-16b3-4652-abfe-a6cdba363724"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:11:55 crc kubenswrapper[4612]: I0227 08:11:55.673325 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/76009243-16b3-4652-abfe-a6cdba363724-config-data" (OuterVolumeSpecName: "config-data") pod "76009243-16b3-4652-abfe-a6cdba363724" (UID: "76009243-16b3-4652-abfe-a6cdba363724"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 08:11:55 crc kubenswrapper[4612]: I0227 08:11:55.677336 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/76009243-16b3-4652-abfe-a6cdba363724-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "76009243-16b3-4652-abfe-a6cdba363724" (UID: "76009243-16b3-4652-abfe-a6cdba363724"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:11:55 crc kubenswrapper[4612]: I0227 08:11:55.702708 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/76009243-16b3-4652-abfe-a6cdba363724-horizon-tls-certs" (OuterVolumeSpecName: "horizon-tls-certs") pod "76009243-16b3-4652-abfe-a6cdba363724" (UID: "76009243-16b3-4652-abfe-a6cdba363724"). InnerVolumeSpecName "horizon-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:11:55 crc kubenswrapper[4612]: I0227 08:11:55.703491 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/76009243-16b3-4652-abfe-a6cdba363724-scripts" (OuterVolumeSpecName: "scripts") pod "76009243-16b3-4652-abfe-a6cdba363724" (UID: "76009243-16b3-4652-abfe-a6cdba363724"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 08:11:55 crc kubenswrapper[4612]: I0227 08:11:55.729895 4612 generic.go:334] "Generic (PLEG): container finished" podID="76009243-16b3-4652-abfe-a6cdba363724" containerID="deeb662d0bcfb1868b750b03f6073e2ff029f55cf5b6b403283389a06b1b3b17" exitCode=137 Feb 27 08:11:55 crc kubenswrapper[4612]: I0227 08:11:55.729980 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6b985547d4-mkfpv" event={"ID":"76009243-16b3-4652-abfe-a6cdba363724","Type":"ContainerDied","Data":"deeb662d0bcfb1868b750b03f6073e2ff029f55cf5b6b403283389a06b1b3b17"} Feb 27 08:11:55 crc kubenswrapper[4612]: I0227 08:11:55.730651 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6b985547d4-mkfpv" event={"ID":"76009243-16b3-4652-abfe-a6cdba363724","Type":"ContainerDied","Data":"6222dbd09c56ca68cdf2e41ea355c459ebd6a049c501036b2337e5fe88da1409"} Feb 27 08:11:55 crc kubenswrapper[4612]: I0227 08:11:55.730711 4612 scope.go:117] "RemoveContainer" containerID="36510ad0ab2650616b7e36d7449187c82f7e47b4ef8d4d802e348d602213aa5a" Feb 27 08:11:55 crc kubenswrapper[4612]: I0227 08:11:55.730004 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6b985547d4-mkfpv" Feb 27 08:11:55 crc kubenswrapper[4612]: I0227 08:11:55.748197 4612 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/76009243-16b3-4652-abfe-a6cdba363724-config-data\") on node \"crc\" DevicePath \"\"" Feb 27 08:11:55 crc kubenswrapper[4612]: I0227 08:11:55.748235 4612 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/76009243-16b3-4652-abfe-a6cdba363724-scripts\") on node \"crc\" DevicePath \"\"" Feb 27 08:11:55 crc kubenswrapper[4612]: I0227 08:11:55.748247 4612 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76009243-16b3-4652-abfe-a6cdba363724-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 08:11:55 crc kubenswrapper[4612]: I0227 08:11:55.748258 4612 reconciler_common.go:293] "Volume detached for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/76009243-16b3-4652-abfe-a6cdba363724-horizon-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 27 08:11:55 crc kubenswrapper[4612]: I0227 08:11:55.748270 4612 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/76009243-16b3-4652-abfe-a6cdba363724-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Feb 27 08:11:55 crc kubenswrapper[4612]: I0227 08:11:55.748280 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9mg4q\" (UniqueName: \"kubernetes.io/projected/76009243-16b3-4652-abfe-a6cdba363724-kube-api-access-9mg4q\") on node \"crc\" DevicePath \"\"" Feb 27 08:11:55 crc kubenswrapper[4612]: I0227 08:11:55.748608 4612 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/76009243-16b3-4652-abfe-a6cdba363724-logs\") on node \"crc\" DevicePath \"\"" Feb 27 08:11:55 crc kubenswrapper[4612]: I0227 08:11:55.808734 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-6b985547d4-mkfpv"] Feb 27 08:11:55 crc kubenswrapper[4612]: I0227 08:11:55.817141 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-6b985547d4-mkfpv"] Feb 27 08:11:55 crc kubenswrapper[4612]: I0227 08:11:55.881210 4612 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Feb 27 08:11:55 crc kubenswrapper[4612]: I0227 08:11:55.881249 4612 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Feb 27 08:11:55 crc kubenswrapper[4612]: I0227 08:11:55.962497 4612 scope.go:117] "RemoveContainer" containerID="deeb662d0bcfb1868b750b03f6073e2ff029f55cf5b6b403283389a06b1b3b17" Feb 27 08:11:55 crc kubenswrapper[4612]: I0227 08:11:55.990619 4612 scope.go:117] "RemoveContainer" containerID="36510ad0ab2650616b7e36d7449187c82f7e47b4ef8d4d802e348d602213aa5a" Feb 27 08:11:55 crc kubenswrapper[4612]: E0227 08:11:55.991231 4612 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"36510ad0ab2650616b7e36d7449187c82f7e47b4ef8d4d802e348d602213aa5a\": container with ID starting with 36510ad0ab2650616b7e36d7449187c82f7e47b4ef8d4d802e348d602213aa5a not found: ID does not exist" containerID="36510ad0ab2650616b7e36d7449187c82f7e47b4ef8d4d802e348d602213aa5a" Feb 27 08:11:55 crc kubenswrapper[4612]: I0227 08:11:55.991286 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"36510ad0ab2650616b7e36d7449187c82f7e47b4ef8d4d802e348d602213aa5a"} err="failed to get container status \"36510ad0ab2650616b7e36d7449187c82f7e47b4ef8d4d802e348d602213aa5a\": rpc error: code = NotFound desc = could not find container \"36510ad0ab2650616b7e36d7449187c82f7e47b4ef8d4d802e348d602213aa5a\": container with ID starting with 36510ad0ab2650616b7e36d7449187c82f7e47b4ef8d4d802e348d602213aa5a not found: ID does not exist" Feb 27 08:11:55 crc kubenswrapper[4612]: I0227 08:11:55.991319 4612 scope.go:117] "RemoveContainer" containerID="deeb662d0bcfb1868b750b03f6073e2ff029f55cf5b6b403283389a06b1b3b17" Feb 27 08:11:55 crc kubenswrapper[4612]: E0227 08:11:55.991637 4612 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"deeb662d0bcfb1868b750b03f6073e2ff029f55cf5b6b403283389a06b1b3b17\": container with ID starting with deeb662d0bcfb1868b750b03f6073e2ff029f55cf5b6b403283389a06b1b3b17 not found: ID does not exist" containerID="deeb662d0bcfb1868b750b03f6073e2ff029f55cf5b6b403283389a06b1b3b17" Feb 27 08:11:55 crc kubenswrapper[4612]: I0227 08:11:55.991677 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"deeb662d0bcfb1868b750b03f6073e2ff029f55cf5b6b403283389a06b1b3b17"} err="failed to get container status \"deeb662d0bcfb1868b750b03f6073e2ff029f55cf5b6b403283389a06b1b3b17\": rpc error: code = NotFound desc = could not find container \"deeb662d0bcfb1868b750b03f6073e2ff029f55cf5b6b403283389a06b1b3b17\": container with ID starting with deeb662d0bcfb1868b750b03f6073e2ff029f55cf5b6b403283389a06b1b3b17 not found: ID does not exist" Feb 27 08:11:56 crc kubenswrapper[4612]: I0227 08:11:56.187225 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Feb 27 08:11:56 crc kubenswrapper[4612]: I0227 08:11:56.187306 4612 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Feb 27 08:11:56 crc kubenswrapper[4612]: I0227 08:11:56.239266 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-bccf8f775-kgn5g" Feb 27 08:11:56 crc kubenswrapper[4612]: I0227 08:11:56.241535 4612 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Feb 27 08:11:56 crc kubenswrapper[4612]: I0227 08:11:56.256726 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Feb 27 08:11:56 crc kubenswrapper[4612]: I0227 08:11:56.356343 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6578955fd5-ds7jf"] Feb 27 08:11:56 crc kubenswrapper[4612]: I0227 08:11:56.356591 4612 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6578955fd5-ds7jf" podUID="aedb2ee7-269c-49e6-9bf7-8ac34d6c1d48" containerName="dnsmasq-dns" containerID="cri-o://c4bbdaba9741e72544fb14edcdad80fe5dc614a3942c1d0557e3c416add35dd7" gracePeriod=10 Feb 27 08:11:56 crc kubenswrapper[4612]: I0227 08:11:56.744358 4612 generic.go:334] "Generic (PLEG): container finished" podID="aedb2ee7-269c-49e6-9bf7-8ac34d6c1d48" containerID="c4bbdaba9741e72544fb14edcdad80fe5dc614a3942c1d0557e3c416add35dd7" exitCode=0 Feb 27 08:11:56 crc kubenswrapper[4612]: I0227 08:11:56.744731 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6578955fd5-ds7jf" event={"ID":"aedb2ee7-269c-49e6-9bf7-8ac34d6c1d48","Type":"ContainerDied","Data":"c4bbdaba9741e72544fb14edcdad80fe5dc614a3942c1d0557e3c416add35dd7"} Feb 27 08:11:56 crc kubenswrapper[4612]: I0227 08:11:56.762935 4612 generic.go:334] "Generic (PLEG): container finished" podID="5d42ee9b-5073-4f40-8a3d-632914f14bbc" containerID="0b4e437518f07bffd24b4f73fca63432cbe5816775bc61ba305365ed6b924fe6" exitCode=0 Feb 27 08:11:56 crc kubenswrapper[4612]: I0227 08:11:56.763006 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-x5mhp" event={"ID":"5d42ee9b-5073-4f40-8a3d-632914f14bbc","Type":"ContainerDied","Data":"0b4e437518f07bffd24b4f73fca63432cbe5816775bc61ba305365ed6b924fe6"} Feb 27 08:11:56 crc kubenswrapper[4612]: I0227 08:11:56.808028 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Feb 27 08:11:56 crc kubenswrapper[4612]: I0227 08:11:56.865377 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="76009243-16b3-4652-abfe-a6cdba363724" path="/var/lib/kubelet/pods/76009243-16b3-4652-abfe-a6cdba363724/volumes" Feb 27 08:11:56 crc kubenswrapper[4612]: I0227 08:11:56.917120 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6578955fd5-ds7jf" Feb 27 08:11:56 crc kubenswrapper[4612]: I0227 08:11:56.964939 4612 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="0a4db194-fc59-4410-9782-698dd03f6efb" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.194:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Feb 27 08:11:56 crc kubenswrapper[4612]: I0227 08:11:56.965110 4612 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="0a4db194-fc59-4410-9782-698dd03f6efb" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.194:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Feb 27 08:11:56 crc kubenswrapper[4612]: I0227 08:11:56.985048 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/aedb2ee7-269c-49e6-9bf7-8ac34d6c1d48-ovsdbserver-sb\") pod \"aedb2ee7-269c-49e6-9bf7-8ac34d6c1d48\" (UID: \"aedb2ee7-269c-49e6-9bf7-8ac34d6c1d48\") " Feb 27 08:11:56 crc kubenswrapper[4612]: I0227 08:11:56.985144 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/aedb2ee7-269c-49e6-9bf7-8ac34d6c1d48-dns-svc\") pod \"aedb2ee7-269c-49e6-9bf7-8ac34d6c1d48\" (UID: \"aedb2ee7-269c-49e6-9bf7-8ac34d6c1d48\") " Feb 27 08:11:56 crc kubenswrapper[4612]: I0227 08:11:56.985207 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aedb2ee7-269c-49e6-9bf7-8ac34d6c1d48-config\") pod \"aedb2ee7-269c-49e6-9bf7-8ac34d6c1d48\" (UID: \"aedb2ee7-269c-49e6-9bf7-8ac34d6c1d48\") " Feb 27 08:11:56 crc kubenswrapper[4612]: I0227 08:11:56.985243 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/aedb2ee7-269c-49e6-9bf7-8ac34d6c1d48-dns-swift-storage-0\") pod \"aedb2ee7-269c-49e6-9bf7-8ac34d6c1d48\" (UID: \"aedb2ee7-269c-49e6-9bf7-8ac34d6c1d48\") " Feb 27 08:11:56 crc kubenswrapper[4612]: I0227 08:11:56.985290 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nvldg\" (UniqueName: \"kubernetes.io/projected/aedb2ee7-269c-49e6-9bf7-8ac34d6c1d48-kube-api-access-nvldg\") pod \"aedb2ee7-269c-49e6-9bf7-8ac34d6c1d48\" (UID: \"aedb2ee7-269c-49e6-9bf7-8ac34d6c1d48\") " Feb 27 08:11:56 crc kubenswrapper[4612]: I0227 08:11:56.985337 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/aedb2ee7-269c-49e6-9bf7-8ac34d6c1d48-ovsdbserver-nb\") pod \"aedb2ee7-269c-49e6-9bf7-8ac34d6c1d48\" (UID: \"aedb2ee7-269c-49e6-9bf7-8ac34d6c1d48\") " Feb 27 08:11:57 crc kubenswrapper[4612]: I0227 08:11:57.013791 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aedb2ee7-269c-49e6-9bf7-8ac34d6c1d48-kube-api-access-nvldg" (OuterVolumeSpecName: "kube-api-access-nvldg") pod "aedb2ee7-269c-49e6-9bf7-8ac34d6c1d48" (UID: "aedb2ee7-269c-49e6-9bf7-8ac34d6c1d48"). InnerVolumeSpecName "kube-api-access-nvldg". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 08:11:57 crc kubenswrapper[4612]: I0227 08:11:57.048845 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aedb2ee7-269c-49e6-9bf7-8ac34d6c1d48-config" (OuterVolumeSpecName: "config") pod "aedb2ee7-269c-49e6-9bf7-8ac34d6c1d48" (UID: "aedb2ee7-269c-49e6-9bf7-8ac34d6c1d48"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 08:11:57 crc kubenswrapper[4612]: I0227 08:11:57.107157 4612 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aedb2ee7-269c-49e6-9bf7-8ac34d6c1d48-config\") on node \"crc\" DevicePath \"\"" Feb 27 08:11:57 crc kubenswrapper[4612]: I0227 08:11:57.107272 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nvldg\" (UniqueName: \"kubernetes.io/projected/aedb2ee7-269c-49e6-9bf7-8ac34d6c1d48-kube-api-access-nvldg\") on node \"crc\" DevicePath \"\"" Feb 27 08:11:57 crc kubenswrapper[4612]: I0227 08:11:57.646175 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aedb2ee7-269c-49e6-9bf7-8ac34d6c1d48-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "aedb2ee7-269c-49e6-9bf7-8ac34d6c1d48" (UID: "aedb2ee7-269c-49e6-9bf7-8ac34d6c1d48"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 08:11:57 crc kubenswrapper[4612]: I0227 08:11:57.648371 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aedb2ee7-269c-49e6-9bf7-8ac34d6c1d48-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "aedb2ee7-269c-49e6-9bf7-8ac34d6c1d48" (UID: "aedb2ee7-269c-49e6-9bf7-8ac34d6c1d48"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 08:11:57 crc kubenswrapper[4612]: I0227 08:11:57.650964 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aedb2ee7-269c-49e6-9bf7-8ac34d6c1d48-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "aedb2ee7-269c-49e6-9bf7-8ac34d6c1d48" (UID: "aedb2ee7-269c-49e6-9bf7-8ac34d6c1d48"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 08:11:57 crc kubenswrapper[4612]: I0227 08:11:57.704026 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Feb 27 08:11:57 crc kubenswrapper[4612]: I0227 08:11:57.723895 4612 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/aedb2ee7-269c-49e6-9bf7-8ac34d6c1d48-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 27 08:11:57 crc kubenswrapper[4612]: I0227 08:11:57.724166 4612 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/aedb2ee7-269c-49e6-9bf7-8ac34d6c1d48-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 27 08:11:57 crc kubenswrapper[4612]: I0227 08:11:57.724177 4612 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/aedb2ee7-269c-49e6-9bf7-8ac34d6c1d48-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Feb 27 08:11:57 crc kubenswrapper[4612]: I0227 08:11:57.739106 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aedb2ee7-269c-49e6-9bf7-8ac34d6c1d48-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "aedb2ee7-269c-49e6-9bf7-8ac34d6c1d48" (UID: "aedb2ee7-269c-49e6-9bf7-8ac34d6c1d48"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 08:11:57 crc kubenswrapper[4612]: I0227 08:11:57.774172 4612 generic.go:334] "Generic (PLEG): container finished" podID="a7bca0ea-5f1b-4de7-83d1-54c76f32a813" containerID="55f1d5da6b2040184cc24875298b865e1ce09f44138529f1feec92112d3f4362" exitCode=0 Feb 27 08:11:57 crc kubenswrapper[4612]: I0227 08:11:57.774231 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-qckvl" event={"ID":"a7bca0ea-5f1b-4de7-83d1-54c76f32a813","Type":"ContainerDied","Data":"55f1d5da6b2040184cc24875298b865e1ce09f44138529f1feec92112d3f4362"} Feb 27 08:11:57 crc kubenswrapper[4612]: I0227 08:11:57.780969 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6578955fd5-ds7jf" Feb 27 08:11:57 crc kubenswrapper[4612]: I0227 08:11:57.781124 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6578955fd5-ds7jf" event={"ID":"aedb2ee7-269c-49e6-9bf7-8ac34d6c1d48","Type":"ContainerDied","Data":"ff55bc6b8616236eae12d1bc0aae268a90ceceb39912082d2ae50274b6bd1b60"} Feb 27 08:11:57 crc kubenswrapper[4612]: I0227 08:11:57.781200 4612 scope.go:117] "RemoveContainer" containerID="c4bbdaba9741e72544fb14edcdad80fe5dc614a3942c1d0557e3c416add35dd7" Feb 27 08:11:57 crc kubenswrapper[4612]: I0227 08:11:57.825571 4612 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/aedb2ee7-269c-49e6-9bf7-8ac34d6c1d48-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 27 08:11:57 crc kubenswrapper[4612]: I0227 08:11:57.859112 4612 scope.go:117] "RemoveContainer" containerID="26a9589d9d888d24be276369ce5e1ce8a8c8a9d4d5b6ff3df566a8f91e230e69" Feb 27 08:11:57 crc kubenswrapper[4612]: I0227 08:11:57.876156 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6578955fd5-ds7jf"] Feb 27 08:11:57 crc kubenswrapper[4612]: I0227 08:11:57.916996 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6578955fd5-ds7jf"] Feb 27 08:11:58 crc kubenswrapper[4612]: I0227 08:11:58.114616 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Feb 27 08:11:58 crc kubenswrapper[4612]: I0227 08:11:58.115925 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Feb 27 08:11:58 crc kubenswrapper[4612]: I0227 08:11:58.211840 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-x5mhp" Feb 27 08:11:58 crc kubenswrapper[4612]: I0227 08:11:58.342201 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d42ee9b-5073-4f40-8a3d-632914f14bbc-combined-ca-bundle\") pod \"5d42ee9b-5073-4f40-8a3d-632914f14bbc\" (UID: \"5d42ee9b-5073-4f40-8a3d-632914f14bbc\") " Feb 27 08:11:58 crc kubenswrapper[4612]: I0227 08:11:58.342267 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dx6ks\" (UniqueName: \"kubernetes.io/projected/5d42ee9b-5073-4f40-8a3d-632914f14bbc-kube-api-access-dx6ks\") pod \"5d42ee9b-5073-4f40-8a3d-632914f14bbc\" (UID: \"5d42ee9b-5073-4f40-8a3d-632914f14bbc\") " Feb 27 08:11:58 crc kubenswrapper[4612]: I0227 08:11:58.342311 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5d42ee9b-5073-4f40-8a3d-632914f14bbc-config-data\") pod \"5d42ee9b-5073-4f40-8a3d-632914f14bbc\" (UID: \"5d42ee9b-5073-4f40-8a3d-632914f14bbc\") " Feb 27 08:11:58 crc kubenswrapper[4612]: I0227 08:11:58.342423 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5d42ee9b-5073-4f40-8a3d-632914f14bbc-scripts\") pod \"5d42ee9b-5073-4f40-8a3d-632914f14bbc\" (UID: \"5d42ee9b-5073-4f40-8a3d-632914f14bbc\") " Feb 27 08:11:58 crc kubenswrapper[4612]: I0227 08:11:58.347076 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5d42ee9b-5073-4f40-8a3d-632914f14bbc-scripts" (OuterVolumeSpecName: "scripts") pod "5d42ee9b-5073-4f40-8a3d-632914f14bbc" (UID: "5d42ee9b-5073-4f40-8a3d-632914f14bbc"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:11:58 crc kubenswrapper[4612]: I0227 08:11:58.347461 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5d42ee9b-5073-4f40-8a3d-632914f14bbc-kube-api-access-dx6ks" (OuterVolumeSpecName: "kube-api-access-dx6ks") pod "5d42ee9b-5073-4f40-8a3d-632914f14bbc" (UID: "5d42ee9b-5073-4f40-8a3d-632914f14bbc"). InnerVolumeSpecName "kube-api-access-dx6ks". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 08:11:58 crc kubenswrapper[4612]: I0227 08:11:58.370179 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5d42ee9b-5073-4f40-8a3d-632914f14bbc-config-data" (OuterVolumeSpecName: "config-data") pod "5d42ee9b-5073-4f40-8a3d-632914f14bbc" (UID: "5d42ee9b-5073-4f40-8a3d-632914f14bbc"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:11:58 crc kubenswrapper[4612]: I0227 08:11:58.372145 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5d42ee9b-5073-4f40-8a3d-632914f14bbc-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5d42ee9b-5073-4f40-8a3d-632914f14bbc" (UID: "5d42ee9b-5073-4f40-8a3d-632914f14bbc"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:11:58 crc kubenswrapper[4612]: I0227 08:11:58.444110 4612 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5d42ee9b-5073-4f40-8a3d-632914f14bbc-scripts\") on node \"crc\" DevicePath \"\"" Feb 27 08:11:58 crc kubenswrapper[4612]: I0227 08:11:58.444146 4612 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d42ee9b-5073-4f40-8a3d-632914f14bbc-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 08:11:58 crc kubenswrapper[4612]: I0227 08:11:58.444158 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dx6ks\" (UniqueName: \"kubernetes.io/projected/5d42ee9b-5073-4f40-8a3d-632914f14bbc-kube-api-access-dx6ks\") on node \"crc\" DevicePath \"\"" Feb 27 08:11:58 crc kubenswrapper[4612]: I0227 08:11:58.444167 4612 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5d42ee9b-5073-4f40-8a3d-632914f14bbc-config-data\") on node \"crc\" DevicePath \"\"" Feb 27 08:11:58 crc kubenswrapper[4612]: I0227 08:11:58.817534 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-x5mhp" event={"ID":"5d42ee9b-5073-4f40-8a3d-632914f14bbc","Type":"ContainerDied","Data":"761ac441cf15f1e500a95e92cba8c2bcb83552f5a290c39a1474ca83e24063f5"} Feb 27 08:11:58 crc kubenswrapper[4612]: I0227 08:11:58.817593 4612 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="761ac441cf15f1e500a95e92cba8c2bcb83552f5a290c39a1474ca83e24063f5" Feb 27 08:11:58 crc kubenswrapper[4612]: I0227 08:11:58.817897 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-x5mhp" Feb 27 08:11:58 crc kubenswrapper[4612]: I0227 08:11:58.864285 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aedb2ee7-269c-49e6-9bf7-8ac34d6c1d48" path="/var/lib/kubelet/pods/aedb2ee7-269c-49e6-9bf7-8ac34d6c1d48/volumes" Feb 27 08:11:58 crc kubenswrapper[4612]: I0227 08:11:58.997218 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Feb 27 08:11:58 crc kubenswrapper[4612]: I0227 08:11:58.997909 4612 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="3c4d5bb6-8ca7-41d7-8e30-af438a0e267a" containerName="nova-scheduler-scheduler" containerID="cri-o://b20dc16d14af8f07dddc21f5ce3968bf2ad89f62cb1e10b67ad0a10affdda173" gracePeriod=30 Feb 27 08:11:59 crc kubenswrapper[4612]: I0227 08:11:59.017823 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Feb 27 08:11:59 crc kubenswrapper[4612]: I0227 08:11:59.018070 4612 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="0a4db194-fc59-4410-9782-698dd03f6efb" containerName="nova-api-log" containerID="cri-o://0b5812f027deb193e6de875cbc8fb3996e3a7515ce2d952698dfb9311d73d4d1" gracePeriod=30 Feb 27 08:11:59 crc kubenswrapper[4612]: I0227 08:11:59.018206 4612 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="0a4db194-fc59-4410-9782-698dd03f6efb" containerName="nova-api-api" containerID="cri-o://b569723412818ade307aee6c05130a45f30421f2ba063b311d0a32ebd456fcbf" gracePeriod=30 Feb 27 08:11:59 crc kubenswrapper[4612]: I0227 08:11:59.027871 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Feb 27 08:11:59 crc kubenswrapper[4612]: I0227 08:11:59.225710 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-qckvl" Feb 27 08:11:59 crc kubenswrapper[4612]: I0227 08:11:59.392892 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a7bca0ea-5f1b-4de7-83d1-54c76f32a813-config-data\") pod \"a7bca0ea-5f1b-4de7-83d1-54c76f32a813\" (UID: \"a7bca0ea-5f1b-4de7-83d1-54c76f32a813\") " Feb 27 08:11:59 crc kubenswrapper[4612]: I0227 08:11:59.392984 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a7bca0ea-5f1b-4de7-83d1-54c76f32a813-scripts\") pod \"a7bca0ea-5f1b-4de7-83d1-54c76f32a813\" (UID: \"a7bca0ea-5f1b-4de7-83d1-54c76f32a813\") " Feb 27 08:11:59 crc kubenswrapper[4612]: I0227 08:11:59.393066 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r62rp\" (UniqueName: \"kubernetes.io/projected/a7bca0ea-5f1b-4de7-83d1-54c76f32a813-kube-api-access-r62rp\") pod \"a7bca0ea-5f1b-4de7-83d1-54c76f32a813\" (UID: \"a7bca0ea-5f1b-4de7-83d1-54c76f32a813\") " Feb 27 08:11:59 crc kubenswrapper[4612]: I0227 08:11:59.393112 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7bca0ea-5f1b-4de7-83d1-54c76f32a813-combined-ca-bundle\") pod \"a7bca0ea-5f1b-4de7-83d1-54c76f32a813\" (UID: \"a7bca0ea-5f1b-4de7-83d1-54c76f32a813\") " Feb 27 08:11:59 crc kubenswrapper[4612]: I0227 08:11:59.398823 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a7bca0ea-5f1b-4de7-83d1-54c76f32a813-scripts" (OuterVolumeSpecName: "scripts") pod "a7bca0ea-5f1b-4de7-83d1-54c76f32a813" (UID: "a7bca0ea-5f1b-4de7-83d1-54c76f32a813"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:11:59 crc kubenswrapper[4612]: I0227 08:11:59.412834 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a7bca0ea-5f1b-4de7-83d1-54c76f32a813-kube-api-access-r62rp" (OuterVolumeSpecName: "kube-api-access-r62rp") pod "a7bca0ea-5f1b-4de7-83d1-54c76f32a813" (UID: "a7bca0ea-5f1b-4de7-83d1-54c76f32a813"). InnerVolumeSpecName "kube-api-access-r62rp". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 08:11:59 crc kubenswrapper[4612]: I0227 08:11:59.431387 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a7bca0ea-5f1b-4de7-83d1-54c76f32a813-config-data" (OuterVolumeSpecName: "config-data") pod "a7bca0ea-5f1b-4de7-83d1-54c76f32a813" (UID: "a7bca0ea-5f1b-4de7-83d1-54c76f32a813"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:11:59 crc kubenswrapper[4612]: I0227 08:11:59.433876 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a7bca0ea-5f1b-4de7-83d1-54c76f32a813-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a7bca0ea-5f1b-4de7-83d1-54c76f32a813" (UID: "a7bca0ea-5f1b-4de7-83d1-54c76f32a813"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:11:59 crc kubenswrapper[4612]: I0227 08:11:59.495095 4612 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a7bca0ea-5f1b-4de7-83d1-54c76f32a813-config-data\") on node \"crc\" DevicePath \"\"" Feb 27 08:11:59 crc kubenswrapper[4612]: I0227 08:11:59.495120 4612 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a7bca0ea-5f1b-4de7-83d1-54c76f32a813-scripts\") on node \"crc\" DevicePath \"\"" Feb 27 08:11:59 crc kubenswrapper[4612]: I0227 08:11:59.495130 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r62rp\" (UniqueName: \"kubernetes.io/projected/a7bca0ea-5f1b-4de7-83d1-54c76f32a813-kube-api-access-r62rp\") on node \"crc\" DevicePath \"\"" Feb 27 08:11:59 crc kubenswrapper[4612]: I0227 08:11:59.495140 4612 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7bca0ea-5f1b-4de7-83d1-54c76f32a813-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 08:11:59 crc kubenswrapper[4612]: I0227 08:11:59.831021 4612 generic.go:334] "Generic (PLEG): container finished" podID="0a4db194-fc59-4410-9782-698dd03f6efb" containerID="0b5812f027deb193e6de875cbc8fb3996e3a7515ce2d952698dfb9311d73d4d1" exitCode=143 Feb 27 08:11:59 crc kubenswrapper[4612]: I0227 08:11:59.831296 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"0a4db194-fc59-4410-9782-698dd03f6efb","Type":"ContainerDied","Data":"0b5812f027deb193e6de875cbc8fb3996e3a7515ce2d952698dfb9311d73d4d1"} Feb 27 08:11:59 crc kubenswrapper[4612]: I0227 08:11:59.832685 4612 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="46b77ad9-d0a9-40de-a765-47437a44d6a6" containerName="nova-metadata-log" containerID="cri-o://3b1ed2ade3e135a2f5916e902cb4d16e90bbf2549b3c6d6480cfa9fbe2f53bd5" gracePeriod=30 Feb 27 08:11:59 crc kubenswrapper[4612]: I0227 08:11:59.833006 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-qckvl" Feb 27 08:11:59 crc kubenswrapper[4612]: I0227 08:11:59.837348 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-qckvl" event={"ID":"a7bca0ea-5f1b-4de7-83d1-54c76f32a813","Type":"ContainerDied","Data":"8a623cb0e75a9d914cac0e9df1b0b5777df34f99124872844c0028295fddb2ad"} Feb 27 08:11:59 crc kubenswrapper[4612]: I0227 08:11:59.837378 4612 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8a623cb0e75a9d914cac0e9df1b0b5777df34f99124872844c0028295fddb2ad" Feb 27 08:11:59 crc kubenswrapper[4612]: I0227 08:11:59.837461 4612 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="46b77ad9-d0a9-40de-a765-47437a44d6a6" containerName="nova-metadata-metadata" containerID="cri-o://1f0880b86bbc77a2d82a1ba76d4db6d467630a0d9fd239f55fcd2c881776f56e" gracePeriod=30 Feb 27 08:11:59 crc kubenswrapper[4612]: I0227 08:11:59.891473 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Feb 27 08:11:59 crc kubenswrapper[4612]: E0227 08:11:59.892014 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aedb2ee7-269c-49e6-9bf7-8ac34d6c1d48" containerName="dnsmasq-dns" Feb 27 08:11:59 crc kubenswrapper[4612]: I0227 08:11:59.892035 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="aedb2ee7-269c-49e6-9bf7-8ac34d6c1d48" containerName="dnsmasq-dns" Feb 27 08:11:59 crc kubenswrapper[4612]: E0227 08:11:59.892051 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a7bca0ea-5f1b-4de7-83d1-54c76f32a813" containerName="nova-cell1-conductor-db-sync" Feb 27 08:11:59 crc kubenswrapper[4612]: I0227 08:11:59.892062 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="a7bca0ea-5f1b-4de7-83d1-54c76f32a813" containerName="nova-cell1-conductor-db-sync" Feb 27 08:11:59 crc kubenswrapper[4612]: E0227 08:11:59.892079 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="76009243-16b3-4652-abfe-a6cdba363724" containerName="horizon" Feb 27 08:11:59 crc kubenswrapper[4612]: I0227 08:11:59.892086 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="76009243-16b3-4652-abfe-a6cdba363724" containerName="horizon" Feb 27 08:11:59 crc kubenswrapper[4612]: E0227 08:11:59.892102 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5d42ee9b-5073-4f40-8a3d-632914f14bbc" containerName="nova-manage" Feb 27 08:11:59 crc kubenswrapper[4612]: I0227 08:11:59.892109 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="5d42ee9b-5073-4f40-8a3d-632914f14bbc" containerName="nova-manage" Feb 27 08:11:59 crc kubenswrapper[4612]: E0227 08:11:59.892126 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="76009243-16b3-4652-abfe-a6cdba363724" containerName="horizon" Feb 27 08:11:59 crc kubenswrapper[4612]: I0227 08:11:59.892132 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="76009243-16b3-4652-abfe-a6cdba363724" containerName="horizon" Feb 27 08:11:59 crc kubenswrapper[4612]: E0227 08:11:59.892151 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aedb2ee7-269c-49e6-9bf7-8ac34d6c1d48" containerName="init" Feb 27 08:11:59 crc kubenswrapper[4612]: I0227 08:11:59.892157 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="aedb2ee7-269c-49e6-9bf7-8ac34d6c1d48" containerName="init" Feb 27 08:11:59 crc kubenswrapper[4612]: E0227 08:11:59.892168 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="76009243-16b3-4652-abfe-a6cdba363724" containerName="horizon-log" Feb 27 08:11:59 crc kubenswrapper[4612]: I0227 08:11:59.892174 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="76009243-16b3-4652-abfe-a6cdba363724" containerName="horizon-log" Feb 27 08:11:59 crc kubenswrapper[4612]: I0227 08:11:59.892337 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="76009243-16b3-4652-abfe-a6cdba363724" containerName="horizon-log" Feb 27 08:11:59 crc kubenswrapper[4612]: I0227 08:11:59.892350 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="76009243-16b3-4652-abfe-a6cdba363724" containerName="horizon" Feb 27 08:11:59 crc kubenswrapper[4612]: I0227 08:11:59.892359 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="5d42ee9b-5073-4f40-8a3d-632914f14bbc" containerName="nova-manage" Feb 27 08:11:59 crc kubenswrapper[4612]: I0227 08:11:59.892373 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="a7bca0ea-5f1b-4de7-83d1-54c76f32a813" containerName="nova-cell1-conductor-db-sync" Feb 27 08:11:59 crc kubenswrapper[4612]: I0227 08:11:59.892387 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="aedb2ee7-269c-49e6-9bf7-8ac34d6c1d48" containerName="dnsmasq-dns" Feb 27 08:11:59 crc kubenswrapper[4612]: I0227 08:11:59.893003 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Feb 27 08:11:59 crc kubenswrapper[4612]: I0227 08:11:59.894509 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Feb 27 08:11:59 crc kubenswrapper[4612]: I0227 08:11:59.901495 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Feb 27 08:12:00 crc kubenswrapper[4612]: I0227 08:12:00.003403 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pflhl\" (UniqueName: \"kubernetes.io/projected/fb906a42-d4ee-4e3e-b4c2-8924890a9e26-kube-api-access-pflhl\") pod \"nova-cell1-conductor-0\" (UID: \"fb906a42-d4ee-4e3e-b4c2-8924890a9e26\") " pod="openstack/nova-cell1-conductor-0" Feb 27 08:12:00 crc kubenswrapper[4612]: I0227 08:12:00.003564 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb906a42-d4ee-4e3e-b4c2-8924890a9e26-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"fb906a42-d4ee-4e3e-b4c2-8924890a9e26\") " pod="openstack/nova-cell1-conductor-0" Feb 27 08:12:00 crc kubenswrapper[4612]: I0227 08:12:00.003593 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fb906a42-d4ee-4e3e-b4c2-8924890a9e26-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"fb906a42-d4ee-4e3e-b4c2-8924890a9e26\") " pod="openstack/nova-cell1-conductor-0" Feb 27 08:12:00 crc kubenswrapper[4612]: I0227 08:12:00.105671 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb906a42-d4ee-4e3e-b4c2-8924890a9e26-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"fb906a42-d4ee-4e3e-b4c2-8924890a9e26\") " pod="openstack/nova-cell1-conductor-0" Feb 27 08:12:00 crc kubenswrapper[4612]: I0227 08:12:00.105737 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fb906a42-d4ee-4e3e-b4c2-8924890a9e26-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"fb906a42-d4ee-4e3e-b4c2-8924890a9e26\") " pod="openstack/nova-cell1-conductor-0" Feb 27 08:12:00 crc kubenswrapper[4612]: I0227 08:12:00.105839 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pflhl\" (UniqueName: \"kubernetes.io/projected/fb906a42-d4ee-4e3e-b4c2-8924890a9e26-kube-api-access-pflhl\") pod \"nova-cell1-conductor-0\" (UID: \"fb906a42-d4ee-4e3e-b4c2-8924890a9e26\") " pod="openstack/nova-cell1-conductor-0" Feb 27 08:12:00 crc kubenswrapper[4612]: I0227 08:12:00.113334 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fb906a42-d4ee-4e3e-b4c2-8924890a9e26-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"fb906a42-d4ee-4e3e-b4c2-8924890a9e26\") " pod="openstack/nova-cell1-conductor-0" Feb 27 08:12:00 crc kubenswrapper[4612]: I0227 08:12:00.114237 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb906a42-d4ee-4e3e-b4c2-8924890a9e26-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"fb906a42-d4ee-4e3e-b4c2-8924890a9e26\") " pod="openstack/nova-cell1-conductor-0" Feb 27 08:12:00 crc kubenswrapper[4612]: I0227 08:12:00.124093 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pflhl\" (UniqueName: \"kubernetes.io/projected/fb906a42-d4ee-4e3e-b4c2-8924890a9e26-kube-api-access-pflhl\") pod \"nova-cell1-conductor-0\" (UID: \"fb906a42-d4ee-4e3e-b4c2-8924890a9e26\") " pod="openstack/nova-cell1-conductor-0" Feb 27 08:12:00 crc kubenswrapper[4612]: I0227 08:12:00.138470 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29536332-8qjw9"] Feb 27 08:12:00 crc kubenswrapper[4612]: I0227 08:12:00.138995 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="76009243-16b3-4652-abfe-a6cdba363724" containerName="horizon" Feb 27 08:12:00 crc kubenswrapper[4612]: I0227 08:12:00.139533 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536332-8qjw9" Feb 27 08:12:00 crc kubenswrapper[4612]: I0227 08:12:00.141823 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-5zwfx" Feb 27 08:12:00 crc kubenswrapper[4612]: I0227 08:12:00.144279 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 27 08:12:00 crc kubenswrapper[4612]: I0227 08:12:00.146850 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 27 08:12:00 crc kubenswrapper[4612]: I0227 08:12:00.149607 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536332-8qjw9"] Feb 27 08:12:00 crc kubenswrapper[4612]: I0227 08:12:00.207230 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Feb 27 08:12:00 crc kubenswrapper[4612]: I0227 08:12:00.215379 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7lhsv\" (UniqueName: \"kubernetes.io/projected/cf3edace-2d20-49a4-95fc-ff3e04fafb77-kube-api-access-7lhsv\") pod \"auto-csr-approver-29536332-8qjw9\" (UID: \"cf3edace-2d20-49a4-95fc-ff3e04fafb77\") " pod="openshift-infra/auto-csr-approver-29536332-8qjw9" Feb 27 08:12:00 crc kubenswrapper[4612]: I0227 08:12:00.316924 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7lhsv\" (UniqueName: \"kubernetes.io/projected/cf3edace-2d20-49a4-95fc-ff3e04fafb77-kube-api-access-7lhsv\") pod \"auto-csr-approver-29536332-8qjw9\" (UID: \"cf3edace-2d20-49a4-95fc-ff3e04fafb77\") " pod="openshift-infra/auto-csr-approver-29536332-8qjw9" Feb 27 08:12:00 crc kubenswrapper[4612]: I0227 08:12:00.333869 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7lhsv\" (UniqueName: \"kubernetes.io/projected/cf3edace-2d20-49a4-95fc-ff3e04fafb77-kube-api-access-7lhsv\") pod \"auto-csr-approver-29536332-8qjw9\" (UID: \"cf3edace-2d20-49a4-95fc-ff3e04fafb77\") " pod="openshift-infra/auto-csr-approver-29536332-8qjw9" Feb 27 08:12:00 crc kubenswrapper[4612]: I0227 08:12:00.364635 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 27 08:12:00 crc kubenswrapper[4612]: I0227 08:12:00.515316 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536332-8qjw9" Feb 27 08:12:00 crc kubenswrapper[4612]: I0227 08:12:00.526984 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/46b77ad9-d0a9-40de-a765-47437a44d6a6-logs\") pod \"46b77ad9-d0a9-40de-a765-47437a44d6a6\" (UID: \"46b77ad9-d0a9-40de-a765-47437a44d6a6\") " Feb 27 08:12:00 crc kubenswrapper[4612]: I0227 08:12:00.527080 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zd9sr\" (UniqueName: \"kubernetes.io/projected/46b77ad9-d0a9-40de-a765-47437a44d6a6-kube-api-access-zd9sr\") pod \"46b77ad9-d0a9-40de-a765-47437a44d6a6\" (UID: \"46b77ad9-d0a9-40de-a765-47437a44d6a6\") " Feb 27 08:12:00 crc kubenswrapper[4612]: I0227 08:12:00.527160 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/46b77ad9-d0a9-40de-a765-47437a44d6a6-nova-metadata-tls-certs\") pod \"46b77ad9-d0a9-40de-a765-47437a44d6a6\" (UID: \"46b77ad9-d0a9-40de-a765-47437a44d6a6\") " Feb 27 08:12:00 crc kubenswrapper[4612]: I0227 08:12:00.527187 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46b77ad9-d0a9-40de-a765-47437a44d6a6-combined-ca-bundle\") pod \"46b77ad9-d0a9-40de-a765-47437a44d6a6\" (UID: \"46b77ad9-d0a9-40de-a765-47437a44d6a6\") " Feb 27 08:12:00 crc kubenswrapper[4612]: I0227 08:12:00.527207 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/46b77ad9-d0a9-40de-a765-47437a44d6a6-config-data\") pod \"46b77ad9-d0a9-40de-a765-47437a44d6a6\" (UID: \"46b77ad9-d0a9-40de-a765-47437a44d6a6\") " Feb 27 08:12:00 crc kubenswrapper[4612]: I0227 08:12:00.527490 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/46b77ad9-d0a9-40de-a765-47437a44d6a6-logs" (OuterVolumeSpecName: "logs") pod "46b77ad9-d0a9-40de-a765-47437a44d6a6" (UID: "46b77ad9-d0a9-40de-a765-47437a44d6a6"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 08:12:00 crc kubenswrapper[4612]: I0227 08:12:00.527632 4612 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/46b77ad9-d0a9-40de-a765-47437a44d6a6-logs\") on node \"crc\" DevicePath \"\"" Feb 27 08:12:00 crc kubenswrapper[4612]: I0227 08:12:00.535001 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/46b77ad9-d0a9-40de-a765-47437a44d6a6-kube-api-access-zd9sr" (OuterVolumeSpecName: "kube-api-access-zd9sr") pod "46b77ad9-d0a9-40de-a765-47437a44d6a6" (UID: "46b77ad9-d0a9-40de-a765-47437a44d6a6"). InnerVolumeSpecName "kube-api-access-zd9sr". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 08:12:00 crc kubenswrapper[4612]: I0227 08:12:00.569830 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/46b77ad9-d0a9-40de-a765-47437a44d6a6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "46b77ad9-d0a9-40de-a765-47437a44d6a6" (UID: "46b77ad9-d0a9-40de-a765-47437a44d6a6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:12:00 crc kubenswrapper[4612]: I0227 08:12:00.584845 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/46b77ad9-d0a9-40de-a765-47437a44d6a6-config-data" (OuterVolumeSpecName: "config-data") pod "46b77ad9-d0a9-40de-a765-47437a44d6a6" (UID: "46b77ad9-d0a9-40de-a765-47437a44d6a6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:12:00 crc kubenswrapper[4612]: I0227 08:12:00.601886 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/46b77ad9-d0a9-40de-a765-47437a44d6a6-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "46b77ad9-d0a9-40de-a765-47437a44d6a6" (UID: "46b77ad9-d0a9-40de-a765-47437a44d6a6"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:12:00 crc kubenswrapper[4612]: I0227 08:12:00.628954 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zd9sr\" (UniqueName: \"kubernetes.io/projected/46b77ad9-d0a9-40de-a765-47437a44d6a6-kube-api-access-zd9sr\") on node \"crc\" DevicePath \"\"" Feb 27 08:12:00 crc kubenswrapper[4612]: I0227 08:12:00.628991 4612 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/46b77ad9-d0a9-40de-a765-47437a44d6a6-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 27 08:12:00 crc kubenswrapper[4612]: I0227 08:12:00.629003 4612 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46b77ad9-d0a9-40de-a765-47437a44d6a6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 08:12:00 crc kubenswrapper[4612]: I0227 08:12:00.629011 4612 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/46b77ad9-d0a9-40de-a765-47437a44d6a6-config-data\") on node \"crc\" DevicePath \"\"" Feb 27 08:12:00 crc kubenswrapper[4612]: I0227 08:12:00.729107 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Feb 27 08:12:00 crc kubenswrapper[4612]: I0227 08:12:00.865374 4612 generic.go:334] "Generic (PLEG): container finished" podID="46b77ad9-d0a9-40de-a765-47437a44d6a6" containerID="1f0880b86bbc77a2d82a1ba76d4db6d467630a0d9fd239f55fcd2c881776f56e" exitCode=0 Feb 27 08:12:00 crc kubenswrapper[4612]: I0227 08:12:00.865403 4612 generic.go:334] "Generic (PLEG): container finished" podID="46b77ad9-d0a9-40de-a765-47437a44d6a6" containerID="3b1ed2ade3e135a2f5916e902cb4d16e90bbf2549b3c6d6480cfa9fbe2f53bd5" exitCode=143 Feb 27 08:12:00 crc kubenswrapper[4612]: I0227 08:12:00.865474 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 27 08:12:00 crc kubenswrapper[4612]: I0227 08:12:00.870554 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"fb906a42-d4ee-4e3e-b4c2-8924890a9e26","Type":"ContainerStarted","Data":"afa71bf5a70264890c7324766a6edd030b6af9bc407e9d15265966979b9d41b9"} Feb 27 08:12:00 crc kubenswrapper[4612]: I0227 08:12:00.870589 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"46b77ad9-d0a9-40de-a765-47437a44d6a6","Type":"ContainerDied","Data":"1f0880b86bbc77a2d82a1ba76d4db6d467630a0d9fd239f55fcd2c881776f56e"} Feb 27 08:12:00 crc kubenswrapper[4612]: I0227 08:12:00.870611 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"46b77ad9-d0a9-40de-a765-47437a44d6a6","Type":"ContainerDied","Data":"3b1ed2ade3e135a2f5916e902cb4d16e90bbf2549b3c6d6480cfa9fbe2f53bd5"} Feb 27 08:12:00 crc kubenswrapper[4612]: I0227 08:12:00.870622 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"46b77ad9-d0a9-40de-a765-47437a44d6a6","Type":"ContainerDied","Data":"0f8f94a59ccd9454f6b463221edaa8160d0ecf4ad50b9ed51a11fd00e5233a49"} Feb 27 08:12:00 crc kubenswrapper[4612]: I0227 08:12:00.870637 4612 scope.go:117] "RemoveContainer" containerID="1f0880b86bbc77a2d82a1ba76d4db6d467630a0d9fd239f55fcd2c881776f56e" Feb 27 08:12:00 crc kubenswrapper[4612]: I0227 08:12:00.910989 4612 scope.go:117] "RemoveContainer" containerID="3b1ed2ade3e135a2f5916e902cb4d16e90bbf2549b3c6d6480cfa9fbe2f53bd5" Feb 27 08:12:00 crc kubenswrapper[4612]: I0227 08:12:00.923957 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Feb 27 08:12:00 crc kubenswrapper[4612]: I0227 08:12:00.936610 4612 scope.go:117] "RemoveContainer" containerID="1f0880b86bbc77a2d82a1ba76d4db6d467630a0d9fd239f55fcd2c881776f56e" Feb 27 08:12:00 crc kubenswrapper[4612]: I0227 08:12:00.936902 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Feb 27 08:12:00 crc kubenswrapper[4612]: E0227 08:12:00.937046 4612 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1f0880b86bbc77a2d82a1ba76d4db6d467630a0d9fd239f55fcd2c881776f56e\": container with ID starting with 1f0880b86bbc77a2d82a1ba76d4db6d467630a0d9fd239f55fcd2c881776f56e not found: ID does not exist" containerID="1f0880b86bbc77a2d82a1ba76d4db6d467630a0d9fd239f55fcd2c881776f56e" Feb 27 08:12:00 crc kubenswrapper[4612]: I0227 08:12:00.937083 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1f0880b86bbc77a2d82a1ba76d4db6d467630a0d9fd239f55fcd2c881776f56e"} err="failed to get container status \"1f0880b86bbc77a2d82a1ba76d4db6d467630a0d9fd239f55fcd2c881776f56e\": rpc error: code = NotFound desc = could not find container \"1f0880b86bbc77a2d82a1ba76d4db6d467630a0d9fd239f55fcd2c881776f56e\": container with ID starting with 1f0880b86bbc77a2d82a1ba76d4db6d467630a0d9fd239f55fcd2c881776f56e not found: ID does not exist" Feb 27 08:12:00 crc kubenswrapper[4612]: I0227 08:12:00.937104 4612 scope.go:117] "RemoveContainer" containerID="3b1ed2ade3e135a2f5916e902cb4d16e90bbf2549b3c6d6480cfa9fbe2f53bd5" Feb 27 08:12:00 crc kubenswrapper[4612]: E0227 08:12:00.937516 4612 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3b1ed2ade3e135a2f5916e902cb4d16e90bbf2549b3c6d6480cfa9fbe2f53bd5\": container with ID starting with 3b1ed2ade3e135a2f5916e902cb4d16e90bbf2549b3c6d6480cfa9fbe2f53bd5 not found: ID does not exist" containerID="3b1ed2ade3e135a2f5916e902cb4d16e90bbf2549b3c6d6480cfa9fbe2f53bd5" Feb 27 08:12:00 crc kubenswrapper[4612]: I0227 08:12:00.937561 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3b1ed2ade3e135a2f5916e902cb4d16e90bbf2549b3c6d6480cfa9fbe2f53bd5"} err="failed to get container status \"3b1ed2ade3e135a2f5916e902cb4d16e90bbf2549b3c6d6480cfa9fbe2f53bd5\": rpc error: code = NotFound desc = could not find container \"3b1ed2ade3e135a2f5916e902cb4d16e90bbf2549b3c6d6480cfa9fbe2f53bd5\": container with ID starting with 3b1ed2ade3e135a2f5916e902cb4d16e90bbf2549b3c6d6480cfa9fbe2f53bd5 not found: ID does not exist" Feb 27 08:12:00 crc kubenswrapper[4612]: I0227 08:12:00.937589 4612 scope.go:117] "RemoveContainer" containerID="1f0880b86bbc77a2d82a1ba76d4db6d467630a0d9fd239f55fcd2c881776f56e" Feb 27 08:12:00 crc kubenswrapper[4612]: I0227 08:12:00.938919 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1f0880b86bbc77a2d82a1ba76d4db6d467630a0d9fd239f55fcd2c881776f56e"} err="failed to get container status \"1f0880b86bbc77a2d82a1ba76d4db6d467630a0d9fd239f55fcd2c881776f56e\": rpc error: code = NotFound desc = could not find container \"1f0880b86bbc77a2d82a1ba76d4db6d467630a0d9fd239f55fcd2c881776f56e\": container with ID starting with 1f0880b86bbc77a2d82a1ba76d4db6d467630a0d9fd239f55fcd2c881776f56e not found: ID does not exist" Feb 27 08:12:00 crc kubenswrapper[4612]: I0227 08:12:00.938959 4612 scope.go:117] "RemoveContainer" containerID="3b1ed2ade3e135a2f5916e902cb4d16e90bbf2549b3c6d6480cfa9fbe2f53bd5" Feb 27 08:12:00 crc kubenswrapper[4612]: I0227 08:12:00.939383 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3b1ed2ade3e135a2f5916e902cb4d16e90bbf2549b3c6d6480cfa9fbe2f53bd5"} err="failed to get container status \"3b1ed2ade3e135a2f5916e902cb4d16e90bbf2549b3c6d6480cfa9fbe2f53bd5\": rpc error: code = NotFound desc = could not find container \"3b1ed2ade3e135a2f5916e902cb4d16e90bbf2549b3c6d6480cfa9fbe2f53bd5\": container with ID starting with 3b1ed2ade3e135a2f5916e902cb4d16e90bbf2549b3c6d6480cfa9fbe2f53bd5 not found: ID does not exist" Feb 27 08:12:00 crc kubenswrapper[4612]: I0227 08:12:00.947553 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Feb 27 08:12:00 crc kubenswrapper[4612]: E0227 08:12:00.947945 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="46b77ad9-d0a9-40de-a765-47437a44d6a6" containerName="nova-metadata-metadata" Feb 27 08:12:00 crc kubenswrapper[4612]: I0227 08:12:00.947961 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="46b77ad9-d0a9-40de-a765-47437a44d6a6" containerName="nova-metadata-metadata" Feb 27 08:12:00 crc kubenswrapper[4612]: E0227 08:12:00.947989 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="46b77ad9-d0a9-40de-a765-47437a44d6a6" containerName="nova-metadata-log" Feb 27 08:12:00 crc kubenswrapper[4612]: I0227 08:12:00.947995 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="46b77ad9-d0a9-40de-a765-47437a44d6a6" containerName="nova-metadata-log" Feb 27 08:12:00 crc kubenswrapper[4612]: I0227 08:12:00.948146 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="46b77ad9-d0a9-40de-a765-47437a44d6a6" containerName="nova-metadata-log" Feb 27 08:12:00 crc kubenswrapper[4612]: I0227 08:12:00.948171 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="46b77ad9-d0a9-40de-a765-47437a44d6a6" containerName="nova-metadata-metadata" Feb 27 08:12:00 crc kubenswrapper[4612]: I0227 08:12:00.949104 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 27 08:12:00 crc kubenswrapper[4612]: I0227 08:12:00.955168 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Feb 27 08:12:00 crc kubenswrapper[4612]: I0227 08:12:00.955388 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Feb 27 08:12:00 crc kubenswrapper[4612]: I0227 08:12:00.955427 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Feb 27 08:12:00 crc kubenswrapper[4612]: I0227 08:12:00.993331 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536332-8qjw9"] Feb 27 08:12:01 crc kubenswrapper[4612]: I0227 08:12:01.039603 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/59d65625-1a7d-4f75-95a3-82ce5976b27a-config-data\") pod \"nova-metadata-0\" (UID: \"59d65625-1a7d-4f75-95a3-82ce5976b27a\") " pod="openstack/nova-metadata-0" Feb 27 08:12:01 crc kubenswrapper[4612]: I0227 08:12:01.039853 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/59d65625-1a7d-4f75-95a3-82ce5976b27a-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"59d65625-1a7d-4f75-95a3-82ce5976b27a\") " pod="openstack/nova-metadata-0" Feb 27 08:12:01 crc kubenswrapper[4612]: I0227 08:12:01.039901 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9qnp2\" (UniqueName: \"kubernetes.io/projected/59d65625-1a7d-4f75-95a3-82ce5976b27a-kube-api-access-9qnp2\") pod \"nova-metadata-0\" (UID: \"59d65625-1a7d-4f75-95a3-82ce5976b27a\") " pod="openstack/nova-metadata-0" Feb 27 08:12:01 crc kubenswrapper[4612]: I0227 08:12:01.040052 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/59d65625-1a7d-4f75-95a3-82ce5976b27a-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"59d65625-1a7d-4f75-95a3-82ce5976b27a\") " pod="openstack/nova-metadata-0" Feb 27 08:12:01 crc kubenswrapper[4612]: I0227 08:12:01.040132 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/59d65625-1a7d-4f75-95a3-82ce5976b27a-logs\") pod \"nova-metadata-0\" (UID: \"59d65625-1a7d-4f75-95a3-82ce5976b27a\") " pod="openstack/nova-metadata-0" Feb 27 08:12:01 crc kubenswrapper[4612]: I0227 08:12:01.141853 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/59d65625-1a7d-4f75-95a3-82ce5976b27a-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"59d65625-1a7d-4f75-95a3-82ce5976b27a\") " pod="openstack/nova-metadata-0" Feb 27 08:12:01 crc kubenswrapper[4612]: I0227 08:12:01.141897 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9qnp2\" (UniqueName: \"kubernetes.io/projected/59d65625-1a7d-4f75-95a3-82ce5976b27a-kube-api-access-9qnp2\") pod \"nova-metadata-0\" (UID: \"59d65625-1a7d-4f75-95a3-82ce5976b27a\") " pod="openstack/nova-metadata-0" Feb 27 08:12:01 crc kubenswrapper[4612]: I0227 08:12:01.141969 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/59d65625-1a7d-4f75-95a3-82ce5976b27a-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"59d65625-1a7d-4f75-95a3-82ce5976b27a\") " pod="openstack/nova-metadata-0" Feb 27 08:12:01 crc kubenswrapper[4612]: I0227 08:12:01.142007 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/59d65625-1a7d-4f75-95a3-82ce5976b27a-logs\") pod \"nova-metadata-0\" (UID: \"59d65625-1a7d-4f75-95a3-82ce5976b27a\") " pod="openstack/nova-metadata-0" Feb 27 08:12:01 crc kubenswrapper[4612]: I0227 08:12:01.142044 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/59d65625-1a7d-4f75-95a3-82ce5976b27a-config-data\") pod \"nova-metadata-0\" (UID: \"59d65625-1a7d-4f75-95a3-82ce5976b27a\") " pod="openstack/nova-metadata-0" Feb 27 08:12:01 crc kubenswrapper[4612]: I0227 08:12:01.143175 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/59d65625-1a7d-4f75-95a3-82ce5976b27a-logs\") pod \"nova-metadata-0\" (UID: \"59d65625-1a7d-4f75-95a3-82ce5976b27a\") " pod="openstack/nova-metadata-0" Feb 27 08:12:01 crc kubenswrapper[4612]: I0227 08:12:01.145926 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/59d65625-1a7d-4f75-95a3-82ce5976b27a-config-data\") pod \"nova-metadata-0\" (UID: \"59d65625-1a7d-4f75-95a3-82ce5976b27a\") " pod="openstack/nova-metadata-0" Feb 27 08:12:01 crc kubenswrapper[4612]: I0227 08:12:01.149306 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/59d65625-1a7d-4f75-95a3-82ce5976b27a-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"59d65625-1a7d-4f75-95a3-82ce5976b27a\") " pod="openstack/nova-metadata-0" Feb 27 08:12:01 crc kubenswrapper[4612]: I0227 08:12:01.151711 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/59d65625-1a7d-4f75-95a3-82ce5976b27a-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"59d65625-1a7d-4f75-95a3-82ce5976b27a\") " pod="openstack/nova-metadata-0" Feb 27 08:12:01 crc kubenswrapper[4612]: I0227 08:12:01.168537 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9qnp2\" (UniqueName: \"kubernetes.io/projected/59d65625-1a7d-4f75-95a3-82ce5976b27a-kube-api-access-9qnp2\") pod \"nova-metadata-0\" (UID: \"59d65625-1a7d-4f75-95a3-82ce5976b27a\") " pod="openstack/nova-metadata-0" Feb 27 08:12:01 crc kubenswrapper[4612]: E0227 08:12:01.192023 4612 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="b20dc16d14af8f07dddc21f5ce3968bf2ad89f62cb1e10b67ad0a10affdda173" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Feb 27 08:12:01 crc kubenswrapper[4612]: E0227 08:12:01.196090 4612 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="b20dc16d14af8f07dddc21f5ce3968bf2ad89f62cb1e10b67ad0a10affdda173" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Feb 27 08:12:01 crc kubenswrapper[4612]: E0227 08:12:01.197574 4612 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="b20dc16d14af8f07dddc21f5ce3968bf2ad89f62cb1e10b67ad0a10affdda173" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Feb 27 08:12:01 crc kubenswrapper[4612]: E0227 08:12:01.197678 4612 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="3c4d5bb6-8ca7-41d7-8e30-af438a0e267a" containerName="nova-scheduler-scheduler" Feb 27 08:12:01 crc kubenswrapper[4612]: I0227 08:12:01.363398 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 27 08:12:01 crc kubenswrapper[4612]: I0227 08:12:01.879627 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Feb 27 08:12:01 crc kubenswrapper[4612]: W0227 08:12:01.881893 4612 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod59d65625_1a7d_4f75_95a3_82ce5976b27a.slice/crio-1ce7ea348234bd6cfcc81b61b16af5d3eabed4cab4d7609771b3f81ecc8e2606 WatchSource:0}: Error finding container 1ce7ea348234bd6cfcc81b61b16af5d3eabed4cab4d7609771b3f81ecc8e2606: Status 404 returned error can't find the container with id 1ce7ea348234bd6cfcc81b61b16af5d3eabed4cab4d7609771b3f81ecc8e2606 Feb 27 08:12:01 crc kubenswrapper[4612]: I0227 08:12:01.882579 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"fb906a42-d4ee-4e3e-b4c2-8924890a9e26","Type":"ContainerStarted","Data":"740a05883afc2786dc67a7efeaad85c2e66b9426e3b22fb6f7b75c764acf7bfb"} Feb 27 08:12:01 crc kubenswrapper[4612]: I0227 08:12:01.882839 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Feb 27 08:12:01 crc kubenswrapper[4612]: I0227 08:12:01.884651 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536332-8qjw9" event={"ID":"cf3edace-2d20-49a4-95fc-ff3e04fafb77","Type":"ContainerStarted","Data":"7694416574becdef04582a5a33703fb5df5985fa60bca19dde62aecd47f93315"} Feb 27 08:12:01 crc kubenswrapper[4612]: I0227 08:12:01.908532 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.908512675 podStartE2EDuration="2.908512675s" podCreationTimestamp="2026-02-27 08:11:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 08:12:01.905148418 +0000 UTC m=+1379.759078456" watchObservedRunningTime="2026-02-27 08:12:01.908512675 +0000 UTC m=+1379.762442663" Feb 27 08:12:02 crc kubenswrapper[4612]: I0227 08:12:02.863606 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="46b77ad9-d0a9-40de-a765-47437a44d6a6" path="/var/lib/kubelet/pods/46b77ad9-d0a9-40de-a765-47437a44d6a6/volumes" Feb 27 08:12:02 crc kubenswrapper[4612]: I0227 08:12:02.918344 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536332-8qjw9" event={"ID":"cf3edace-2d20-49a4-95fc-ff3e04fafb77","Type":"ContainerStarted","Data":"22a8e7a514287d732f249a859cb167853f917dc5c1a0c1423aea971bfb518228"} Feb 27 08:12:02 crc kubenswrapper[4612]: I0227 08:12:02.922682 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"59d65625-1a7d-4f75-95a3-82ce5976b27a","Type":"ContainerStarted","Data":"e04e3c34eac9644f55d0e1c95060a48f1ec164a86ffdcf54bf34f17667296e37"} Feb 27 08:12:02 crc kubenswrapper[4612]: I0227 08:12:02.922858 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"59d65625-1a7d-4f75-95a3-82ce5976b27a","Type":"ContainerStarted","Data":"2f2643fefc567894c7e71a70aa2c5809517c1801c609eac1707c40bcf9c76d76"} Feb 27 08:12:02 crc kubenswrapper[4612]: I0227 08:12:02.922936 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"59d65625-1a7d-4f75-95a3-82ce5976b27a","Type":"ContainerStarted","Data":"1ce7ea348234bd6cfcc81b61b16af5d3eabed4cab4d7609771b3f81ecc8e2606"} Feb 27 08:12:02 crc kubenswrapper[4612]: I0227 08:12:02.961130 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29536332-8qjw9" podStartSLOduration=1.474959933 podStartE2EDuration="2.961107572s" podCreationTimestamp="2026-02-27 08:12:00 +0000 UTC" firstStartedPulling="2026-02-27 08:12:00.998234853 +0000 UTC m=+1378.852164851" lastFinishedPulling="2026-02-27 08:12:02.484382502 +0000 UTC m=+1380.338312490" observedRunningTime="2026-02-27 08:12:02.930451322 +0000 UTC m=+1380.784381320" watchObservedRunningTime="2026-02-27 08:12:02.961107572 +0000 UTC m=+1380.815037570" Feb 27 08:12:02 crc kubenswrapper[4612]: I0227 08:12:02.963031 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.963025007 podStartE2EDuration="2.963025007s" podCreationTimestamp="2026-02-27 08:12:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 08:12:02.949748856 +0000 UTC m=+1380.803678854" watchObservedRunningTime="2026-02-27 08:12:02.963025007 +0000 UTC m=+1380.816955005" Feb 27 08:12:03 crc kubenswrapper[4612]: I0227 08:12:03.390145 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 27 08:12:03 crc kubenswrapper[4612]: I0227 08:12:03.486610 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3c4d5bb6-8ca7-41d7-8e30-af438a0e267a-config-data\") pod \"3c4d5bb6-8ca7-41d7-8e30-af438a0e267a\" (UID: \"3c4d5bb6-8ca7-41d7-8e30-af438a0e267a\") " Feb 27 08:12:03 crc kubenswrapper[4612]: I0227 08:12:03.486726 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nn7g6\" (UniqueName: \"kubernetes.io/projected/3c4d5bb6-8ca7-41d7-8e30-af438a0e267a-kube-api-access-nn7g6\") pod \"3c4d5bb6-8ca7-41d7-8e30-af438a0e267a\" (UID: \"3c4d5bb6-8ca7-41d7-8e30-af438a0e267a\") " Feb 27 08:12:03 crc kubenswrapper[4612]: I0227 08:12:03.486897 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3c4d5bb6-8ca7-41d7-8e30-af438a0e267a-combined-ca-bundle\") pod \"3c4d5bb6-8ca7-41d7-8e30-af438a0e267a\" (UID: \"3c4d5bb6-8ca7-41d7-8e30-af438a0e267a\") " Feb 27 08:12:03 crc kubenswrapper[4612]: I0227 08:12:03.513937 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3c4d5bb6-8ca7-41d7-8e30-af438a0e267a-kube-api-access-nn7g6" (OuterVolumeSpecName: "kube-api-access-nn7g6") pod "3c4d5bb6-8ca7-41d7-8e30-af438a0e267a" (UID: "3c4d5bb6-8ca7-41d7-8e30-af438a0e267a"). InnerVolumeSpecName "kube-api-access-nn7g6". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 08:12:03 crc kubenswrapper[4612]: I0227 08:12:03.523857 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3c4d5bb6-8ca7-41d7-8e30-af438a0e267a-config-data" (OuterVolumeSpecName: "config-data") pod "3c4d5bb6-8ca7-41d7-8e30-af438a0e267a" (UID: "3c4d5bb6-8ca7-41d7-8e30-af438a0e267a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:12:03 crc kubenswrapper[4612]: I0227 08:12:03.530826 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3c4d5bb6-8ca7-41d7-8e30-af438a0e267a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3c4d5bb6-8ca7-41d7-8e30-af438a0e267a" (UID: "3c4d5bb6-8ca7-41d7-8e30-af438a0e267a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:12:03 crc kubenswrapper[4612]: I0227 08:12:03.588636 4612 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3c4d5bb6-8ca7-41d7-8e30-af438a0e267a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 08:12:03 crc kubenswrapper[4612]: I0227 08:12:03.588674 4612 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3c4d5bb6-8ca7-41d7-8e30-af438a0e267a-config-data\") on node \"crc\" DevicePath \"\"" Feb 27 08:12:03 crc kubenswrapper[4612]: I0227 08:12:03.588684 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nn7g6\" (UniqueName: \"kubernetes.io/projected/3c4d5bb6-8ca7-41d7-8e30-af438a0e267a-kube-api-access-nn7g6\") on node \"crc\" DevicePath \"\"" Feb 27 08:12:03 crc kubenswrapper[4612]: I0227 08:12:03.931577 4612 generic.go:334] "Generic (PLEG): container finished" podID="cf3edace-2d20-49a4-95fc-ff3e04fafb77" containerID="22a8e7a514287d732f249a859cb167853f917dc5c1a0c1423aea971bfb518228" exitCode=0 Feb 27 08:12:03 crc kubenswrapper[4612]: I0227 08:12:03.931648 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536332-8qjw9" event={"ID":"cf3edace-2d20-49a4-95fc-ff3e04fafb77","Type":"ContainerDied","Data":"22a8e7a514287d732f249a859cb167853f917dc5c1a0c1423aea971bfb518228"} Feb 27 08:12:03 crc kubenswrapper[4612]: I0227 08:12:03.933914 4612 generic.go:334] "Generic (PLEG): container finished" podID="3c4d5bb6-8ca7-41d7-8e30-af438a0e267a" containerID="b20dc16d14af8f07dddc21f5ce3968bf2ad89f62cb1e10b67ad0a10affdda173" exitCode=0 Feb 27 08:12:03 crc kubenswrapper[4612]: I0227 08:12:03.934311 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 27 08:12:03 crc kubenswrapper[4612]: I0227 08:12:03.934953 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"3c4d5bb6-8ca7-41d7-8e30-af438a0e267a","Type":"ContainerDied","Data":"b20dc16d14af8f07dddc21f5ce3968bf2ad89f62cb1e10b67ad0a10affdda173"} Feb 27 08:12:03 crc kubenswrapper[4612]: I0227 08:12:03.934989 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"3c4d5bb6-8ca7-41d7-8e30-af438a0e267a","Type":"ContainerDied","Data":"3a425120e135e1aa509b9b0ea5561b0641992abe57e59d56dae08a9105f3e8f6"} Feb 27 08:12:03 crc kubenswrapper[4612]: I0227 08:12:03.935005 4612 scope.go:117] "RemoveContainer" containerID="b20dc16d14af8f07dddc21f5ce3968bf2ad89f62cb1e10b67ad0a10affdda173" Feb 27 08:12:03 crc kubenswrapper[4612]: I0227 08:12:03.961899 4612 scope.go:117] "RemoveContainer" containerID="b20dc16d14af8f07dddc21f5ce3968bf2ad89f62cb1e10b67ad0a10affdda173" Feb 27 08:12:03 crc kubenswrapper[4612]: E0227 08:12:03.962288 4612 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b20dc16d14af8f07dddc21f5ce3968bf2ad89f62cb1e10b67ad0a10affdda173\": container with ID starting with b20dc16d14af8f07dddc21f5ce3968bf2ad89f62cb1e10b67ad0a10affdda173 not found: ID does not exist" containerID="b20dc16d14af8f07dddc21f5ce3968bf2ad89f62cb1e10b67ad0a10affdda173" Feb 27 08:12:03 crc kubenswrapper[4612]: I0227 08:12:03.962347 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b20dc16d14af8f07dddc21f5ce3968bf2ad89f62cb1e10b67ad0a10affdda173"} err="failed to get container status \"b20dc16d14af8f07dddc21f5ce3968bf2ad89f62cb1e10b67ad0a10affdda173\": rpc error: code = NotFound desc = could not find container \"b20dc16d14af8f07dddc21f5ce3968bf2ad89f62cb1e10b67ad0a10affdda173\": container with ID starting with b20dc16d14af8f07dddc21f5ce3968bf2ad89f62cb1e10b67ad0a10affdda173 not found: ID does not exist" Feb 27 08:12:03 crc kubenswrapper[4612]: I0227 08:12:03.981865 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Feb 27 08:12:03 crc kubenswrapper[4612]: I0227 08:12:03.989321 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Feb 27 08:12:04 crc kubenswrapper[4612]: I0227 08:12:04.027604 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Feb 27 08:12:04 crc kubenswrapper[4612]: E0227 08:12:04.028103 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c4d5bb6-8ca7-41d7-8e30-af438a0e267a" containerName="nova-scheduler-scheduler" Feb 27 08:12:04 crc kubenswrapper[4612]: I0227 08:12:04.028120 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c4d5bb6-8ca7-41d7-8e30-af438a0e267a" containerName="nova-scheduler-scheduler" Feb 27 08:12:04 crc kubenswrapper[4612]: I0227 08:12:04.028295 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="3c4d5bb6-8ca7-41d7-8e30-af438a0e267a" containerName="nova-scheduler-scheduler" Feb 27 08:12:04 crc kubenswrapper[4612]: I0227 08:12:04.028976 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 27 08:12:04 crc kubenswrapper[4612]: I0227 08:12:04.031982 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Feb 27 08:12:04 crc kubenswrapper[4612]: I0227 08:12:04.034712 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Feb 27 08:12:04 crc kubenswrapper[4612]: I0227 08:12:04.098503 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rtkd2\" (UniqueName: \"kubernetes.io/projected/ff8b7423-9c8a-4c8a-8d7f-11e70ded2244-kube-api-access-rtkd2\") pod \"nova-scheduler-0\" (UID: \"ff8b7423-9c8a-4c8a-8d7f-11e70ded2244\") " pod="openstack/nova-scheduler-0" Feb 27 08:12:04 crc kubenswrapper[4612]: I0227 08:12:04.098668 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff8b7423-9c8a-4c8a-8d7f-11e70ded2244-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"ff8b7423-9c8a-4c8a-8d7f-11e70ded2244\") " pod="openstack/nova-scheduler-0" Feb 27 08:12:04 crc kubenswrapper[4612]: I0227 08:12:04.098913 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ff8b7423-9c8a-4c8a-8d7f-11e70ded2244-config-data\") pod \"nova-scheduler-0\" (UID: \"ff8b7423-9c8a-4c8a-8d7f-11e70ded2244\") " pod="openstack/nova-scheduler-0" Feb 27 08:12:04 crc kubenswrapper[4612]: I0227 08:12:04.200660 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ff8b7423-9c8a-4c8a-8d7f-11e70ded2244-config-data\") pod \"nova-scheduler-0\" (UID: \"ff8b7423-9c8a-4c8a-8d7f-11e70ded2244\") " pod="openstack/nova-scheduler-0" Feb 27 08:12:04 crc kubenswrapper[4612]: I0227 08:12:04.200752 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rtkd2\" (UniqueName: \"kubernetes.io/projected/ff8b7423-9c8a-4c8a-8d7f-11e70ded2244-kube-api-access-rtkd2\") pod \"nova-scheduler-0\" (UID: \"ff8b7423-9c8a-4c8a-8d7f-11e70ded2244\") " pod="openstack/nova-scheduler-0" Feb 27 08:12:04 crc kubenswrapper[4612]: I0227 08:12:04.200854 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff8b7423-9c8a-4c8a-8d7f-11e70ded2244-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"ff8b7423-9c8a-4c8a-8d7f-11e70ded2244\") " pod="openstack/nova-scheduler-0" Feb 27 08:12:04 crc kubenswrapper[4612]: I0227 08:12:04.208207 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ff8b7423-9c8a-4c8a-8d7f-11e70ded2244-config-data\") pod \"nova-scheduler-0\" (UID: \"ff8b7423-9c8a-4c8a-8d7f-11e70ded2244\") " pod="openstack/nova-scheduler-0" Feb 27 08:12:04 crc kubenswrapper[4612]: I0227 08:12:04.208298 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff8b7423-9c8a-4c8a-8d7f-11e70ded2244-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"ff8b7423-9c8a-4c8a-8d7f-11e70ded2244\") " pod="openstack/nova-scheduler-0" Feb 27 08:12:04 crc kubenswrapper[4612]: I0227 08:12:04.218102 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rtkd2\" (UniqueName: \"kubernetes.io/projected/ff8b7423-9c8a-4c8a-8d7f-11e70ded2244-kube-api-access-rtkd2\") pod \"nova-scheduler-0\" (UID: \"ff8b7423-9c8a-4c8a-8d7f-11e70ded2244\") " pod="openstack/nova-scheduler-0" Feb 27 08:12:04 crc kubenswrapper[4612]: I0227 08:12:04.362384 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 27 08:12:04 crc kubenswrapper[4612]: I0227 08:12:04.847514 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Feb 27 08:12:04 crc kubenswrapper[4612]: I0227 08:12:04.850725 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 27 08:12:04 crc kubenswrapper[4612]: W0227 08:12:04.855521 4612 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podff8b7423_9c8a_4c8a_8d7f_11e70ded2244.slice/crio-83e9f6a60dbda9b9f4133346657757915edd38f1e46079776f3938e91d88f745 WatchSource:0}: Error finding container 83e9f6a60dbda9b9f4133346657757915edd38f1e46079776f3938e91d88f745: Status 404 returned error can't find the container with id 83e9f6a60dbda9b9f4133346657757915edd38f1e46079776f3938e91d88f745 Feb 27 08:12:04 crc kubenswrapper[4612]: I0227 08:12:04.867169 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3c4d5bb6-8ca7-41d7-8e30-af438a0e267a" path="/var/lib/kubelet/pods/3c4d5bb6-8ca7-41d7-8e30-af438a0e267a/volumes" Feb 27 08:12:04 crc kubenswrapper[4612]: I0227 08:12:04.917038 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a4db194-fc59-4410-9782-698dd03f6efb-combined-ca-bundle\") pod \"0a4db194-fc59-4410-9782-698dd03f6efb\" (UID: \"0a4db194-fc59-4410-9782-698dd03f6efb\") " Feb 27 08:12:04 crc kubenswrapper[4612]: I0227 08:12:04.917092 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bh2p6\" (UniqueName: \"kubernetes.io/projected/0a4db194-fc59-4410-9782-698dd03f6efb-kube-api-access-bh2p6\") pod \"0a4db194-fc59-4410-9782-698dd03f6efb\" (UID: \"0a4db194-fc59-4410-9782-698dd03f6efb\") " Feb 27 08:12:04 crc kubenswrapper[4612]: I0227 08:12:04.917140 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a4db194-fc59-4410-9782-698dd03f6efb-config-data\") pod \"0a4db194-fc59-4410-9782-698dd03f6efb\" (UID: \"0a4db194-fc59-4410-9782-698dd03f6efb\") " Feb 27 08:12:04 crc kubenswrapper[4612]: I0227 08:12:04.917226 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0a4db194-fc59-4410-9782-698dd03f6efb-logs\") pod \"0a4db194-fc59-4410-9782-698dd03f6efb\" (UID: \"0a4db194-fc59-4410-9782-698dd03f6efb\") " Feb 27 08:12:04 crc kubenswrapper[4612]: I0227 08:12:04.926287 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0a4db194-fc59-4410-9782-698dd03f6efb-logs" (OuterVolumeSpecName: "logs") pod "0a4db194-fc59-4410-9782-698dd03f6efb" (UID: "0a4db194-fc59-4410-9782-698dd03f6efb"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 08:12:04 crc kubenswrapper[4612]: I0227 08:12:04.937108 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0a4db194-fc59-4410-9782-698dd03f6efb-kube-api-access-bh2p6" (OuterVolumeSpecName: "kube-api-access-bh2p6") pod "0a4db194-fc59-4410-9782-698dd03f6efb" (UID: "0a4db194-fc59-4410-9782-698dd03f6efb"). InnerVolumeSpecName "kube-api-access-bh2p6". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 08:12:04 crc kubenswrapper[4612]: I0227 08:12:04.946587 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0a4db194-fc59-4410-9782-698dd03f6efb-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0a4db194-fc59-4410-9782-698dd03f6efb" (UID: "0a4db194-fc59-4410-9782-698dd03f6efb"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:12:04 crc kubenswrapper[4612]: I0227 08:12:04.953295 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"ff8b7423-9c8a-4c8a-8d7f-11e70ded2244","Type":"ContainerStarted","Data":"83e9f6a60dbda9b9f4133346657757915edd38f1e46079776f3938e91d88f745"} Feb 27 08:12:04 crc kubenswrapper[4612]: I0227 08:12:04.955144 4612 generic.go:334] "Generic (PLEG): container finished" podID="0a4db194-fc59-4410-9782-698dd03f6efb" containerID="b569723412818ade307aee6c05130a45f30421f2ba063b311d0a32ebd456fcbf" exitCode=0 Feb 27 08:12:04 crc kubenswrapper[4612]: I0227 08:12:04.955185 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"0a4db194-fc59-4410-9782-698dd03f6efb","Type":"ContainerDied","Data":"b569723412818ade307aee6c05130a45f30421f2ba063b311d0a32ebd456fcbf"} Feb 27 08:12:04 crc kubenswrapper[4612]: I0227 08:12:04.955249 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"0a4db194-fc59-4410-9782-698dd03f6efb","Type":"ContainerDied","Data":"619505c020ad44021687ffaf53350d5e157222fc8787daa76d252abd2fae4cec"} Feb 27 08:12:04 crc kubenswrapper[4612]: I0227 08:12:04.955269 4612 scope.go:117] "RemoveContainer" containerID="b569723412818ade307aee6c05130a45f30421f2ba063b311d0a32ebd456fcbf" Feb 27 08:12:04 crc kubenswrapper[4612]: I0227 08:12:04.955396 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 27 08:12:04 crc kubenswrapper[4612]: I0227 08:12:04.960279 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0a4db194-fc59-4410-9782-698dd03f6efb-config-data" (OuterVolumeSpecName: "config-data") pod "0a4db194-fc59-4410-9782-698dd03f6efb" (UID: "0a4db194-fc59-4410-9782-698dd03f6efb"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:12:04 crc kubenswrapper[4612]: I0227 08:12:04.980288 4612 scope.go:117] "RemoveContainer" containerID="0b5812f027deb193e6de875cbc8fb3996e3a7515ce2d952698dfb9311d73d4d1" Feb 27 08:12:05 crc kubenswrapper[4612]: I0227 08:12:05.002547 4612 scope.go:117] "RemoveContainer" containerID="b569723412818ade307aee6c05130a45f30421f2ba063b311d0a32ebd456fcbf" Feb 27 08:12:05 crc kubenswrapper[4612]: E0227 08:12:05.003574 4612 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b569723412818ade307aee6c05130a45f30421f2ba063b311d0a32ebd456fcbf\": container with ID starting with b569723412818ade307aee6c05130a45f30421f2ba063b311d0a32ebd456fcbf not found: ID does not exist" containerID="b569723412818ade307aee6c05130a45f30421f2ba063b311d0a32ebd456fcbf" Feb 27 08:12:05 crc kubenswrapper[4612]: I0227 08:12:05.003637 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b569723412818ade307aee6c05130a45f30421f2ba063b311d0a32ebd456fcbf"} err="failed to get container status \"b569723412818ade307aee6c05130a45f30421f2ba063b311d0a32ebd456fcbf\": rpc error: code = NotFound desc = could not find container \"b569723412818ade307aee6c05130a45f30421f2ba063b311d0a32ebd456fcbf\": container with ID starting with b569723412818ade307aee6c05130a45f30421f2ba063b311d0a32ebd456fcbf not found: ID does not exist" Feb 27 08:12:05 crc kubenswrapper[4612]: I0227 08:12:05.003677 4612 scope.go:117] "RemoveContainer" containerID="0b5812f027deb193e6de875cbc8fb3996e3a7515ce2d952698dfb9311d73d4d1" Feb 27 08:12:05 crc kubenswrapper[4612]: E0227 08:12:05.004235 4612 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0b5812f027deb193e6de875cbc8fb3996e3a7515ce2d952698dfb9311d73d4d1\": container with ID starting with 0b5812f027deb193e6de875cbc8fb3996e3a7515ce2d952698dfb9311d73d4d1 not found: ID does not exist" containerID="0b5812f027deb193e6de875cbc8fb3996e3a7515ce2d952698dfb9311d73d4d1" Feb 27 08:12:05 crc kubenswrapper[4612]: I0227 08:12:05.004276 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0b5812f027deb193e6de875cbc8fb3996e3a7515ce2d952698dfb9311d73d4d1"} err="failed to get container status \"0b5812f027deb193e6de875cbc8fb3996e3a7515ce2d952698dfb9311d73d4d1\": rpc error: code = NotFound desc = could not find container \"0b5812f027deb193e6de875cbc8fb3996e3a7515ce2d952698dfb9311d73d4d1\": container with ID starting with 0b5812f027deb193e6de875cbc8fb3996e3a7515ce2d952698dfb9311d73d4d1 not found: ID does not exist" Feb 27 08:12:05 crc kubenswrapper[4612]: I0227 08:12:05.019231 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bh2p6\" (UniqueName: \"kubernetes.io/projected/0a4db194-fc59-4410-9782-698dd03f6efb-kube-api-access-bh2p6\") on node \"crc\" DevicePath \"\"" Feb 27 08:12:05 crc kubenswrapper[4612]: I0227 08:12:05.019256 4612 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a4db194-fc59-4410-9782-698dd03f6efb-config-data\") on node \"crc\" DevicePath \"\"" Feb 27 08:12:05 crc kubenswrapper[4612]: I0227 08:12:05.019266 4612 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0a4db194-fc59-4410-9782-698dd03f6efb-logs\") on node \"crc\" DevicePath \"\"" Feb 27 08:12:05 crc kubenswrapper[4612]: I0227 08:12:05.019274 4612 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a4db194-fc59-4410-9782-698dd03f6efb-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 08:12:05 crc kubenswrapper[4612]: I0227 08:12:05.233007 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Feb 27 08:12:05 crc kubenswrapper[4612]: I0227 08:12:05.327545 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536332-8qjw9" Feb 27 08:12:05 crc kubenswrapper[4612]: I0227 08:12:05.340634 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Feb 27 08:12:05 crc kubenswrapper[4612]: I0227 08:12:05.356927 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Feb 27 08:12:05 crc kubenswrapper[4612]: I0227 08:12:05.373323 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Feb 27 08:12:05 crc kubenswrapper[4612]: E0227 08:12:05.373855 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a4db194-fc59-4410-9782-698dd03f6efb" containerName="nova-api-log" Feb 27 08:12:05 crc kubenswrapper[4612]: I0227 08:12:05.373934 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a4db194-fc59-4410-9782-698dd03f6efb" containerName="nova-api-log" Feb 27 08:12:05 crc kubenswrapper[4612]: E0227 08:12:05.374013 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cf3edace-2d20-49a4-95fc-ff3e04fafb77" containerName="oc" Feb 27 08:12:05 crc kubenswrapper[4612]: I0227 08:12:05.374067 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="cf3edace-2d20-49a4-95fc-ff3e04fafb77" containerName="oc" Feb 27 08:12:05 crc kubenswrapper[4612]: E0227 08:12:05.374125 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a4db194-fc59-4410-9782-698dd03f6efb" containerName="nova-api-api" Feb 27 08:12:05 crc kubenswrapper[4612]: I0227 08:12:05.374173 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a4db194-fc59-4410-9782-698dd03f6efb" containerName="nova-api-api" Feb 27 08:12:05 crc kubenswrapper[4612]: I0227 08:12:05.374396 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="0a4db194-fc59-4410-9782-698dd03f6efb" containerName="nova-api-api" Feb 27 08:12:05 crc kubenswrapper[4612]: I0227 08:12:05.374456 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="0a4db194-fc59-4410-9782-698dd03f6efb" containerName="nova-api-log" Feb 27 08:12:05 crc kubenswrapper[4612]: I0227 08:12:05.374512 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="cf3edace-2d20-49a4-95fc-ff3e04fafb77" containerName="oc" Feb 27 08:12:05 crc kubenswrapper[4612]: I0227 08:12:05.376353 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 27 08:12:05 crc kubenswrapper[4612]: I0227 08:12:05.398336 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Feb 27 08:12:05 crc kubenswrapper[4612]: I0227 08:12:05.406083 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Feb 27 08:12:05 crc kubenswrapper[4612]: I0227 08:12:05.425530 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7lhsv\" (UniqueName: \"kubernetes.io/projected/cf3edace-2d20-49a4-95fc-ff3e04fafb77-kube-api-access-7lhsv\") pod \"cf3edace-2d20-49a4-95fc-ff3e04fafb77\" (UID: \"cf3edace-2d20-49a4-95fc-ff3e04fafb77\") " Feb 27 08:12:05 crc kubenswrapper[4612]: I0227 08:12:05.431199 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cf3edace-2d20-49a4-95fc-ff3e04fafb77-kube-api-access-7lhsv" (OuterVolumeSpecName: "kube-api-access-7lhsv") pod "cf3edace-2d20-49a4-95fc-ff3e04fafb77" (UID: "cf3edace-2d20-49a4-95fc-ff3e04fafb77"). InnerVolumeSpecName "kube-api-access-7lhsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 08:12:05 crc kubenswrapper[4612]: I0227 08:12:05.528541 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d6aaa653-6c5a-4809-957b-6bdefa994b16-config-data\") pod \"nova-api-0\" (UID: \"d6aaa653-6c5a-4809-957b-6bdefa994b16\") " pod="openstack/nova-api-0" Feb 27 08:12:05 crc kubenswrapper[4612]: I0227 08:12:05.528640 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d6aaa653-6c5a-4809-957b-6bdefa994b16-logs\") pod \"nova-api-0\" (UID: \"d6aaa653-6c5a-4809-957b-6bdefa994b16\") " pod="openstack/nova-api-0" Feb 27 08:12:05 crc kubenswrapper[4612]: I0227 08:12:05.528707 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-glztw\" (UniqueName: \"kubernetes.io/projected/d6aaa653-6c5a-4809-957b-6bdefa994b16-kube-api-access-glztw\") pod \"nova-api-0\" (UID: \"d6aaa653-6c5a-4809-957b-6bdefa994b16\") " pod="openstack/nova-api-0" Feb 27 08:12:05 crc kubenswrapper[4612]: I0227 08:12:05.528792 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6aaa653-6c5a-4809-957b-6bdefa994b16-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"d6aaa653-6c5a-4809-957b-6bdefa994b16\") " pod="openstack/nova-api-0" Feb 27 08:12:05 crc kubenswrapper[4612]: I0227 08:12:05.528853 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7lhsv\" (UniqueName: \"kubernetes.io/projected/cf3edace-2d20-49a4-95fc-ff3e04fafb77-kube-api-access-7lhsv\") on node \"crc\" DevicePath \"\"" Feb 27 08:12:05 crc kubenswrapper[4612]: I0227 08:12:05.630305 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-glztw\" (UniqueName: \"kubernetes.io/projected/d6aaa653-6c5a-4809-957b-6bdefa994b16-kube-api-access-glztw\") pod \"nova-api-0\" (UID: \"d6aaa653-6c5a-4809-957b-6bdefa994b16\") " pod="openstack/nova-api-0" Feb 27 08:12:05 crc kubenswrapper[4612]: I0227 08:12:05.630399 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6aaa653-6c5a-4809-957b-6bdefa994b16-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"d6aaa653-6c5a-4809-957b-6bdefa994b16\") " pod="openstack/nova-api-0" Feb 27 08:12:05 crc kubenswrapper[4612]: I0227 08:12:05.630457 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d6aaa653-6c5a-4809-957b-6bdefa994b16-config-data\") pod \"nova-api-0\" (UID: \"d6aaa653-6c5a-4809-957b-6bdefa994b16\") " pod="openstack/nova-api-0" Feb 27 08:12:05 crc kubenswrapper[4612]: I0227 08:12:05.630537 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d6aaa653-6c5a-4809-957b-6bdefa994b16-logs\") pod \"nova-api-0\" (UID: \"d6aaa653-6c5a-4809-957b-6bdefa994b16\") " pod="openstack/nova-api-0" Feb 27 08:12:05 crc kubenswrapper[4612]: I0227 08:12:05.631077 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d6aaa653-6c5a-4809-957b-6bdefa994b16-logs\") pod \"nova-api-0\" (UID: \"d6aaa653-6c5a-4809-957b-6bdefa994b16\") " pod="openstack/nova-api-0" Feb 27 08:12:05 crc kubenswrapper[4612]: I0227 08:12:05.635097 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d6aaa653-6c5a-4809-957b-6bdefa994b16-config-data\") pod \"nova-api-0\" (UID: \"d6aaa653-6c5a-4809-957b-6bdefa994b16\") " pod="openstack/nova-api-0" Feb 27 08:12:05 crc kubenswrapper[4612]: I0227 08:12:05.635429 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6aaa653-6c5a-4809-957b-6bdefa994b16-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"d6aaa653-6c5a-4809-957b-6bdefa994b16\") " pod="openstack/nova-api-0" Feb 27 08:12:05 crc kubenswrapper[4612]: I0227 08:12:05.644821 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-glztw\" (UniqueName: \"kubernetes.io/projected/d6aaa653-6c5a-4809-957b-6bdefa994b16-kube-api-access-glztw\") pod \"nova-api-0\" (UID: \"d6aaa653-6c5a-4809-957b-6bdefa994b16\") " pod="openstack/nova-api-0" Feb 27 08:12:05 crc kubenswrapper[4612]: I0227 08:12:05.769783 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 27 08:12:06 crc kubenswrapper[4612]: I0227 08:12:06.008864 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"ff8b7423-9c8a-4c8a-8d7f-11e70ded2244","Type":"ContainerStarted","Data":"205e0e7c8976d96647c6a5c7b44e2fbf349cb1346ac71036d6ee20ad722ab20b"} Feb 27 08:12:06 crc kubenswrapper[4612]: I0227 08:12:06.024880 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536332-8qjw9" event={"ID":"cf3edace-2d20-49a4-95fc-ff3e04fafb77","Type":"ContainerDied","Data":"7694416574becdef04582a5a33703fb5df5985fa60bca19dde62aecd47f93315"} Feb 27 08:12:06 crc kubenswrapper[4612]: I0227 08:12:06.024920 4612 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7694416574becdef04582a5a33703fb5df5985fa60bca19dde62aecd47f93315" Feb 27 08:12:06 crc kubenswrapper[4612]: I0227 08:12:06.024986 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536332-8qjw9" Feb 27 08:12:06 crc kubenswrapper[4612]: I0227 08:12:06.040744 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29536326-pjlzf"] Feb 27 08:12:06 crc kubenswrapper[4612]: I0227 08:12:06.051203 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29536326-pjlzf"] Feb 27 08:12:06 crc kubenswrapper[4612]: I0227 08:12:06.052329 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=3.052312096 podStartE2EDuration="3.052312096s" podCreationTimestamp="2026-02-27 08:12:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 08:12:06.031075186 +0000 UTC m=+1383.885005184" watchObservedRunningTime="2026-02-27 08:12:06.052312096 +0000 UTC m=+1383.906242094" Feb 27 08:12:06 crc kubenswrapper[4612]: I0227 08:12:06.361031 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Feb 27 08:12:06 crc kubenswrapper[4612]: I0227 08:12:06.363895 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Feb 27 08:12:06 crc kubenswrapper[4612]: I0227 08:12:06.363932 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Feb 27 08:12:06 crc kubenswrapper[4612]: W0227 08:12:06.372471 4612 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd6aaa653_6c5a_4809_957b_6bdefa994b16.slice/crio-3da8467bc27b960ab3f5c5e2ca8f4dbd708a2c9ae862bd6229372a0e8c31c5ba WatchSource:0}: Error finding container 3da8467bc27b960ab3f5c5e2ca8f4dbd708a2c9ae862bd6229372a0e8c31c5ba: Status 404 returned error can't find the container with id 3da8467bc27b960ab3f5c5e2ca8f4dbd708a2c9ae862bd6229372a0e8c31c5ba Feb 27 08:12:06 crc kubenswrapper[4612]: I0227 08:12:06.862814 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0a4db194-fc59-4410-9782-698dd03f6efb" path="/var/lib/kubelet/pods/0a4db194-fc59-4410-9782-698dd03f6efb/volumes" Feb 27 08:12:06 crc kubenswrapper[4612]: I0227 08:12:06.865638 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3aec7965-4aeb-432f-8b2a-f17d58427093" path="/var/lib/kubelet/pods/3aec7965-4aeb-432f-8b2a-f17d58427093/volumes" Feb 27 08:12:07 crc kubenswrapper[4612]: I0227 08:12:07.041845 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d6aaa653-6c5a-4809-957b-6bdefa994b16","Type":"ContainerStarted","Data":"29b7410f9f42497aeb02a86bb006d6b74ad02aaf6af9b4aa7ad66a42e0b5faf8"} Feb 27 08:12:07 crc kubenswrapper[4612]: I0227 08:12:07.041899 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d6aaa653-6c5a-4809-957b-6bdefa994b16","Type":"ContainerStarted","Data":"d78fb4b2a5a2ceeff13f9cc9454d569fedad601d2c42a08163d50af833bb6288"} Feb 27 08:12:07 crc kubenswrapper[4612]: I0227 08:12:07.041914 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d6aaa653-6c5a-4809-957b-6bdefa994b16","Type":"ContainerStarted","Data":"3da8467bc27b960ab3f5c5e2ca8f4dbd708a2c9ae862bd6229372a0e8c31c5ba"} Feb 27 08:12:07 crc kubenswrapper[4612]: I0227 08:12:07.077438 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.077417825 podStartE2EDuration="2.077417825s" podCreationTimestamp="2026-02-27 08:12:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 08:12:07.061999002 +0000 UTC m=+1384.915929010" watchObservedRunningTime="2026-02-27 08:12:07.077417825 +0000 UTC m=+1384.931347823" Feb 27 08:12:08 crc kubenswrapper[4612]: E0227 08:12:08.351683 4612 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3c4d5bb6_8ca7_41d7_8e30_af438a0e267a.slice/crio-b20dc16d14af8f07dddc21f5ce3968bf2ad89f62cb1e10b67ad0a10affdda173.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3c4d5bb6_8ca7_41d7_8e30_af438a0e267a.slice/crio-conmon-b20dc16d14af8f07dddc21f5ce3968bf2ad89f62cb1e10b67ad0a10affdda173.scope\": RecentStats: unable to find data in memory cache]" Feb 27 08:12:09 crc kubenswrapper[4612]: I0227 08:12:09.362577 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Feb 27 08:12:11 crc kubenswrapper[4612]: I0227 08:12:11.364946 4612 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Feb 27 08:12:11 crc kubenswrapper[4612]: I0227 08:12:11.365330 4612 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Feb 27 08:12:12 crc kubenswrapper[4612]: I0227 08:12:12.381904 4612 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="59d65625-1a7d-4f75-95a3-82ce5976b27a" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.203:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Feb 27 08:12:12 crc kubenswrapper[4612]: I0227 08:12:12.382295 4612 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="59d65625-1a7d-4f75-95a3-82ce5976b27a" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.203:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Feb 27 08:12:14 crc kubenswrapper[4612]: I0227 08:12:14.363488 4612 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Feb 27 08:12:14 crc kubenswrapper[4612]: I0227 08:12:14.396192 4612 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Feb 27 08:12:15 crc kubenswrapper[4612]: I0227 08:12:15.169303 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Feb 27 08:12:15 crc kubenswrapper[4612]: I0227 08:12:15.770580 4612 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Feb 27 08:12:15 crc kubenswrapper[4612]: I0227 08:12:15.771003 4612 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Feb 27 08:12:16 crc kubenswrapper[4612]: I0227 08:12:16.852943 4612 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="d6aaa653-6c5a-4809-957b-6bdefa994b16" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.205:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Feb 27 08:12:16 crc kubenswrapper[4612]: I0227 08:12:16.853400 4612 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="d6aaa653-6c5a-4809-957b-6bdefa994b16" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.205:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Feb 27 08:12:18 crc kubenswrapper[4612]: E0227 08:12:18.627660 4612 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3c4d5bb6_8ca7_41d7_8e30_af438a0e267a.slice/crio-b20dc16d14af8f07dddc21f5ce3968bf2ad89f62cb1e10b67ad0a10affdda173.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3c4d5bb6_8ca7_41d7_8e30_af438a0e267a.slice/crio-conmon-b20dc16d14af8f07dddc21f5ce3968bf2ad89f62cb1e10b67ad0a10affdda173.scope\": RecentStats: unable to find data in memory cache]" Feb 27 08:12:21 crc kubenswrapper[4612]: I0227 08:12:21.370755 4612 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Feb 27 08:12:21 crc kubenswrapper[4612]: I0227 08:12:21.379769 4612 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Feb 27 08:12:21 crc kubenswrapper[4612]: I0227 08:12:21.383490 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Feb 27 08:12:22 crc kubenswrapper[4612]: I0227 08:12:22.132977 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Feb 27 08:12:22 crc kubenswrapper[4612]: I0227 08:12:22.225276 4612 generic.go:334] "Generic (PLEG): container finished" podID="a96cfd8e-57bd-4780-a843-e30c9d1a0e02" containerID="a3f5f545e871de451788fd0c23694fd3101c0a3404666ccabcf72114b3416ed1" exitCode=137 Feb 27 08:12:22 crc kubenswrapper[4612]: I0227 08:12:22.226625 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Feb 27 08:12:22 crc kubenswrapper[4612]: I0227 08:12:22.227124 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"a96cfd8e-57bd-4780-a843-e30c9d1a0e02","Type":"ContainerDied","Data":"a3f5f545e871de451788fd0c23694fd3101c0a3404666ccabcf72114b3416ed1"} Feb 27 08:12:22 crc kubenswrapper[4612]: I0227 08:12:22.227164 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"a96cfd8e-57bd-4780-a843-e30c9d1a0e02","Type":"ContainerDied","Data":"06d0b8f94832cbd72d3834a653f08d2469aceb87280381ac4cbc39f838817f1c"} Feb 27 08:12:22 crc kubenswrapper[4612]: I0227 08:12:22.227181 4612 scope.go:117] "RemoveContainer" containerID="a3f5f545e871de451788fd0c23694fd3101c0a3404666ccabcf72114b3416ed1" Feb 27 08:12:22 crc kubenswrapper[4612]: I0227 08:12:22.237034 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Feb 27 08:12:22 crc kubenswrapper[4612]: I0227 08:12:22.281199 4612 scope.go:117] "RemoveContainer" containerID="a3f5f545e871de451788fd0c23694fd3101c0a3404666ccabcf72114b3416ed1" Feb 27 08:12:22 crc kubenswrapper[4612]: E0227 08:12:22.281685 4612 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a3f5f545e871de451788fd0c23694fd3101c0a3404666ccabcf72114b3416ed1\": container with ID starting with a3f5f545e871de451788fd0c23694fd3101c0a3404666ccabcf72114b3416ed1 not found: ID does not exist" containerID="a3f5f545e871de451788fd0c23694fd3101c0a3404666ccabcf72114b3416ed1" Feb 27 08:12:22 crc kubenswrapper[4612]: I0227 08:12:22.281734 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a3f5f545e871de451788fd0c23694fd3101c0a3404666ccabcf72114b3416ed1"} err="failed to get container status \"a3f5f545e871de451788fd0c23694fd3101c0a3404666ccabcf72114b3416ed1\": rpc error: code = NotFound desc = could not find container \"a3f5f545e871de451788fd0c23694fd3101c0a3404666ccabcf72114b3416ed1\": container with ID starting with a3f5f545e871de451788fd0c23694fd3101c0a3404666ccabcf72114b3416ed1 not found: ID does not exist" Feb 27 08:12:22 crc kubenswrapper[4612]: I0227 08:12:22.286757 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a96cfd8e-57bd-4780-a843-e30c9d1a0e02-combined-ca-bundle\") pod \"a96cfd8e-57bd-4780-a843-e30c9d1a0e02\" (UID: \"a96cfd8e-57bd-4780-a843-e30c9d1a0e02\") " Feb 27 08:12:22 crc kubenswrapper[4612]: I0227 08:12:22.286854 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a96cfd8e-57bd-4780-a843-e30c9d1a0e02-config-data\") pod \"a96cfd8e-57bd-4780-a843-e30c9d1a0e02\" (UID: \"a96cfd8e-57bd-4780-a843-e30c9d1a0e02\") " Feb 27 08:12:22 crc kubenswrapper[4612]: I0227 08:12:22.286978 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sfz69\" (UniqueName: \"kubernetes.io/projected/a96cfd8e-57bd-4780-a843-e30c9d1a0e02-kube-api-access-sfz69\") pod \"a96cfd8e-57bd-4780-a843-e30c9d1a0e02\" (UID: \"a96cfd8e-57bd-4780-a843-e30c9d1a0e02\") " Feb 27 08:12:22 crc kubenswrapper[4612]: I0227 08:12:22.310221 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a96cfd8e-57bd-4780-a843-e30c9d1a0e02-kube-api-access-sfz69" (OuterVolumeSpecName: "kube-api-access-sfz69") pod "a96cfd8e-57bd-4780-a843-e30c9d1a0e02" (UID: "a96cfd8e-57bd-4780-a843-e30c9d1a0e02"). InnerVolumeSpecName "kube-api-access-sfz69". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 08:12:22 crc kubenswrapper[4612]: I0227 08:12:22.323982 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a96cfd8e-57bd-4780-a843-e30c9d1a0e02-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a96cfd8e-57bd-4780-a843-e30c9d1a0e02" (UID: "a96cfd8e-57bd-4780-a843-e30c9d1a0e02"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:12:22 crc kubenswrapper[4612]: I0227 08:12:22.336649 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a96cfd8e-57bd-4780-a843-e30c9d1a0e02-config-data" (OuterVolumeSpecName: "config-data") pod "a96cfd8e-57bd-4780-a843-e30c9d1a0e02" (UID: "a96cfd8e-57bd-4780-a843-e30c9d1a0e02"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:12:22 crc kubenswrapper[4612]: I0227 08:12:22.394375 4612 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a96cfd8e-57bd-4780-a843-e30c9d1a0e02-config-data\") on node \"crc\" DevicePath \"\"" Feb 27 08:12:22 crc kubenswrapper[4612]: I0227 08:12:22.394404 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sfz69\" (UniqueName: \"kubernetes.io/projected/a96cfd8e-57bd-4780-a843-e30c9d1a0e02-kube-api-access-sfz69\") on node \"crc\" DevicePath \"\"" Feb 27 08:12:22 crc kubenswrapper[4612]: I0227 08:12:22.394416 4612 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a96cfd8e-57bd-4780-a843-e30c9d1a0e02-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 08:12:22 crc kubenswrapper[4612]: I0227 08:12:22.555015 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Feb 27 08:12:22 crc kubenswrapper[4612]: I0227 08:12:22.567016 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Feb 27 08:12:22 crc kubenswrapper[4612]: I0227 08:12:22.581171 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Feb 27 08:12:22 crc kubenswrapper[4612]: E0227 08:12:22.581601 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a96cfd8e-57bd-4780-a843-e30c9d1a0e02" containerName="nova-cell1-novncproxy-novncproxy" Feb 27 08:12:22 crc kubenswrapper[4612]: I0227 08:12:22.581620 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="a96cfd8e-57bd-4780-a843-e30c9d1a0e02" containerName="nova-cell1-novncproxy-novncproxy" Feb 27 08:12:22 crc kubenswrapper[4612]: I0227 08:12:22.581826 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="a96cfd8e-57bd-4780-a843-e30c9d1a0e02" containerName="nova-cell1-novncproxy-novncproxy" Feb 27 08:12:22 crc kubenswrapper[4612]: I0227 08:12:22.582502 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Feb 27 08:12:22 crc kubenswrapper[4612]: I0227 08:12:22.586353 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Feb 27 08:12:22 crc kubenswrapper[4612]: I0227 08:12:22.586485 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Feb 27 08:12:22 crc kubenswrapper[4612]: I0227 08:12:22.586561 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Feb 27 08:12:22 crc kubenswrapper[4612]: I0227 08:12:22.608875 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Feb 27 08:12:22 crc kubenswrapper[4612]: I0227 08:12:22.702204 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/211046a5-967c-44f0-8320-ddcc223dda97-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"211046a5-967c-44f0-8320-ddcc223dda97\") " pod="openstack/nova-cell1-novncproxy-0" Feb 27 08:12:22 crc kubenswrapper[4612]: I0227 08:12:22.702275 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/211046a5-967c-44f0-8320-ddcc223dda97-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"211046a5-967c-44f0-8320-ddcc223dda97\") " pod="openstack/nova-cell1-novncproxy-0" Feb 27 08:12:22 crc kubenswrapper[4612]: I0227 08:12:22.702333 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-48n9r\" (UniqueName: \"kubernetes.io/projected/211046a5-967c-44f0-8320-ddcc223dda97-kube-api-access-48n9r\") pod \"nova-cell1-novncproxy-0\" (UID: \"211046a5-967c-44f0-8320-ddcc223dda97\") " pod="openstack/nova-cell1-novncproxy-0" Feb 27 08:12:22 crc kubenswrapper[4612]: I0227 08:12:22.702472 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/211046a5-967c-44f0-8320-ddcc223dda97-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"211046a5-967c-44f0-8320-ddcc223dda97\") " pod="openstack/nova-cell1-novncproxy-0" Feb 27 08:12:22 crc kubenswrapper[4612]: I0227 08:12:22.702492 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/211046a5-967c-44f0-8320-ddcc223dda97-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"211046a5-967c-44f0-8320-ddcc223dda97\") " pod="openstack/nova-cell1-novncproxy-0" Feb 27 08:12:22 crc kubenswrapper[4612]: I0227 08:12:22.804439 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/211046a5-967c-44f0-8320-ddcc223dda97-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"211046a5-967c-44f0-8320-ddcc223dda97\") " pod="openstack/nova-cell1-novncproxy-0" Feb 27 08:12:22 crc kubenswrapper[4612]: I0227 08:12:22.804903 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-48n9r\" (UniqueName: \"kubernetes.io/projected/211046a5-967c-44f0-8320-ddcc223dda97-kube-api-access-48n9r\") pod \"nova-cell1-novncproxy-0\" (UID: \"211046a5-967c-44f0-8320-ddcc223dda97\") " pod="openstack/nova-cell1-novncproxy-0" Feb 27 08:12:22 crc kubenswrapper[4612]: I0227 08:12:22.805064 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/211046a5-967c-44f0-8320-ddcc223dda97-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"211046a5-967c-44f0-8320-ddcc223dda97\") " pod="openstack/nova-cell1-novncproxy-0" Feb 27 08:12:22 crc kubenswrapper[4612]: I0227 08:12:22.805099 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/211046a5-967c-44f0-8320-ddcc223dda97-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"211046a5-967c-44f0-8320-ddcc223dda97\") " pod="openstack/nova-cell1-novncproxy-0" Feb 27 08:12:22 crc kubenswrapper[4612]: I0227 08:12:22.805142 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/211046a5-967c-44f0-8320-ddcc223dda97-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"211046a5-967c-44f0-8320-ddcc223dda97\") " pod="openstack/nova-cell1-novncproxy-0" Feb 27 08:12:22 crc kubenswrapper[4612]: I0227 08:12:22.808035 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/211046a5-967c-44f0-8320-ddcc223dda97-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"211046a5-967c-44f0-8320-ddcc223dda97\") " pod="openstack/nova-cell1-novncproxy-0" Feb 27 08:12:22 crc kubenswrapper[4612]: I0227 08:12:22.809000 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/211046a5-967c-44f0-8320-ddcc223dda97-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"211046a5-967c-44f0-8320-ddcc223dda97\") " pod="openstack/nova-cell1-novncproxy-0" Feb 27 08:12:22 crc kubenswrapper[4612]: I0227 08:12:22.810817 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/211046a5-967c-44f0-8320-ddcc223dda97-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"211046a5-967c-44f0-8320-ddcc223dda97\") " pod="openstack/nova-cell1-novncproxy-0" Feb 27 08:12:22 crc kubenswrapper[4612]: I0227 08:12:22.821040 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/211046a5-967c-44f0-8320-ddcc223dda97-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"211046a5-967c-44f0-8320-ddcc223dda97\") " pod="openstack/nova-cell1-novncproxy-0" Feb 27 08:12:22 crc kubenswrapper[4612]: I0227 08:12:22.825719 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-48n9r\" (UniqueName: \"kubernetes.io/projected/211046a5-967c-44f0-8320-ddcc223dda97-kube-api-access-48n9r\") pod \"nova-cell1-novncproxy-0\" (UID: \"211046a5-967c-44f0-8320-ddcc223dda97\") " pod="openstack/nova-cell1-novncproxy-0" Feb 27 08:12:22 crc kubenswrapper[4612]: I0227 08:12:22.866366 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a96cfd8e-57bd-4780-a843-e30c9d1a0e02" path="/var/lib/kubelet/pods/a96cfd8e-57bd-4780-a843-e30c9d1a0e02/volumes" Feb 27 08:12:22 crc kubenswrapper[4612]: I0227 08:12:22.903732 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Feb 27 08:12:23 crc kubenswrapper[4612]: I0227 08:12:23.364037 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Feb 27 08:12:23 crc kubenswrapper[4612]: W0227 08:12:23.371815 4612 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod211046a5_967c_44f0_8320_ddcc223dda97.slice/crio-4e3acfed1e4116b24699916071c9ecb62959b64a17b8499da9528cf648d3b30d WatchSource:0}: Error finding container 4e3acfed1e4116b24699916071c9ecb62959b64a17b8499da9528cf648d3b30d: Status 404 returned error can't find the container with id 4e3acfed1e4116b24699916071c9ecb62959b64a17b8499da9528cf648d3b30d Feb 27 08:12:24 crc kubenswrapper[4612]: I0227 08:12:24.247568 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"211046a5-967c-44f0-8320-ddcc223dda97","Type":"ContainerStarted","Data":"3955d8898d81f3b7c4e2fb46d7370a05fcd92912e3ece35eda8299532d199acd"} Feb 27 08:12:24 crc kubenswrapper[4612]: I0227 08:12:24.248027 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"211046a5-967c-44f0-8320-ddcc223dda97","Type":"ContainerStarted","Data":"4e3acfed1e4116b24699916071c9ecb62959b64a17b8499da9528cf648d3b30d"} Feb 27 08:12:24 crc kubenswrapper[4612]: I0227 08:12:24.271984 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.27196826 podStartE2EDuration="2.27196826s" podCreationTimestamp="2026-02-27 08:12:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 08:12:24.270157308 +0000 UTC m=+1402.124087306" watchObservedRunningTime="2026-02-27 08:12:24.27196826 +0000 UTC m=+1402.125898258" Feb 27 08:12:25 crc kubenswrapper[4612]: I0227 08:12:25.776558 4612 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Feb 27 08:12:25 crc kubenswrapper[4612]: I0227 08:12:25.777408 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Feb 27 08:12:25 crc kubenswrapper[4612]: I0227 08:12:25.779049 4612 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Feb 27 08:12:25 crc kubenswrapper[4612]: I0227 08:12:25.783183 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Feb 27 08:12:26 crc kubenswrapper[4612]: I0227 08:12:26.266145 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Feb 27 08:12:26 crc kubenswrapper[4612]: I0227 08:12:26.271365 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Feb 27 08:12:26 crc kubenswrapper[4612]: I0227 08:12:26.570285 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-cd5cbd7b9-dczkk"] Feb 27 08:12:26 crc kubenswrapper[4612]: I0227 08:12:26.571790 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cd5cbd7b9-dczkk" Feb 27 08:12:26 crc kubenswrapper[4612]: I0227 08:12:26.587546 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/63e633cb-a49f-4adb-89f2-dddff70954b6-dns-swift-storage-0\") pod \"dnsmasq-dns-cd5cbd7b9-dczkk\" (UID: \"63e633cb-a49f-4adb-89f2-dddff70954b6\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-dczkk" Feb 27 08:12:26 crc kubenswrapper[4612]: I0227 08:12:26.587660 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/63e633cb-a49f-4adb-89f2-dddff70954b6-ovsdbserver-nb\") pod \"dnsmasq-dns-cd5cbd7b9-dczkk\" (UID: \"63e633cb-a49f-4adb-89f2-dddff70954b6\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-dczkk" Feb 27 08:12:26 crc kubenswrapper[4612]: I0227 08:12:26.587758 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lk7xj\" (UniqueName: \"kubernetes.io/projected/63e633cb-a49f-4adb-89f2-dddff70954b6-kube-api-access-lk7xj\") pod \"dnsmasq-dns-cd5cbd7b9-dczkk\" (UID: \"63e633cb-a49f-4adb-89f2-dddff70954b6\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-dczkk" Feb 27 08:12:26 crc kubenswrapper[4612]: I0227 08:12:26.587829 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/63e633cb-a49f-4adb-89f2-dddff70954b6-dns-svc\") pod \"dnsmasq-dns-cd5cbd7b9-dczkk\" (UID: \"63e633cb-a49f-4adb-89f2-dddff70954b6\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-dczkk" Feb 27 08:12:26 crc kubenswrapper[4612]: I0227 08:12:26.588042 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/63e633cb-a49f-4adb-89f2-dddff70954b6-ovsdbserver-sb\") pod \"dnsmasq-dns-cd5cbd7b9-dczkk\" (UID: \"63e633cb-a49f-4adb-89f2-dddff70954b6\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-dczkk" Feb 27 08:12:26 crc kubenswrapper[4612]: I0227 08:12:26.588078 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/63e633cb-a49f-4adb-89f2-dddff70954b6-config\") pod \"dnsmasq-dns-cd5cbd7b9-dczkk\" (UID: \"63e633cb-a49f-4adb-89f2-dddff70954b6\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-dczkk" Feb 27 08:12:26 crc kubenswrapper[4612]: I0227 08:12:26.590882 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-cd5cbd7b9-dczkk"] Feb 27 08:12:26 crc kubenswrapper[4612]: I0227 08:12:26.689056 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/63e633cb-a49f-4adb-89f2-dddff70954b6-ovsdbserver-nb\") pod \"dnsmasq-dns-cd5cbd7b9-dczkk\" (UID: \"63e633cb-a49f-4adb-89f2-dddff70954b6\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-dczkk" Feb 27 08:12:26 crc kubenswrapper[4612]: I0227 08:12:26.689112 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lk7xj\" (UniqueName: \"kubernetes.io/projected/63e633cb-a49f-4adb-89f2-dddff70954b6-kube-api-access-lk7xj\") pod \"dnsmasq-dns-cd5cbd7b9-dczkk\" (UID: \"63e633cb-a49f-4adb-89f2-dddff70954b6\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-dczkk" Feb 27 08:12:26 crc kubenswrapper[4612]: I0227 08:12:26.689145 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/63e633cb-a49f-4adb-89f2-dddff70954b6-dns-svc\") pod \"dnsmasq-dns-cd5cbd7b9-dczkk\" (UID: \"63e633cb-a49f-4adb-89f2-dddff70954b6\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-dczkk" Feb 27 08:12:26 crc kubenswrapper[4612]: I0227 08:12:26.689229 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/63e633cb-a49f-4adb-89f2-dddff70954b6-ovsdbserver-sb\") pod \"dnsmasq-dns-cd5cbd7b9-dczkk\" (UID: \"63e633cb-a49f-4adb-89f2-dddff70954b6\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-dczkk" Feb 27 08:12:26 crc kubenswrapper[4612]: I0227 08:12:26.689249 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/63e633cb-a49f-4adb-89f2-dddff70954b6-config\") pod \"dnsmasq-dns-cd5cbd7b9-dczkk\" (UID: \"63e633cb-a49f-4adb-89f2-dddff70954b6\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-dczkk" Feb 27 08:12:26 crc kubenswrapper[4612]: I0227 08:12:26.689280 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/63e633cb-a49f-4adb-89f2-dddff70954b6-dns-swift-storage-0\") pod \"dnsmasq-dns-cd5cbd7b9-dczkk\" (UID: \"63e633cb-a49f-4adb-89f2-dddff70954b6\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-dczkk" Feb 27 08:12:26 crc kubenswrapper[4612]: I0227 08:12:26.690461 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/63e633cb-a49f-4adb-89f2-dddff70954b6-dns-svc\") pod \"dnsmasq-dns-cd5cbd7b9-dczkk\" (UID: \"63e633cb-a49f-4adb-89f2-dddff70954b6\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-dczkk" Feb 27 08:12:26 crc kubenswrapper[4612]: I0227 08:12:26.691674 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/63e633cb-a49f-4adb-89f2-dddff70954b6-ovsdbserver-nb\") pod \"dnsmasq-dns-cd5cbd7b9-dczkk\" (UID: \"63e633cb-a49f-4adb-89f2-dddff70954b6\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-dczkk" Feb 27 08:12:26 crc kubenswrapper[4612]: I0227 08:12:26.692210 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/63e633cb-a49f-4adb-89f2-dddff70954b6-ovsdbserver-sb\") pod \"dnsmasq-dns-cd5cbd7b9-dczkk\" (UID: \"63e633cb-a49f-4adb-89f2-dddff70954b6\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-dczkk" Feb 27 08:12:26 crc kubenswrapper[4612]: I0227 08:12:26.692741 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/63e633cb-a49f-4adb-89f2-dddff70954b6-config\") pod \"dnsmasq-dns-cd5cbd7b9-dczkk\" (UID: \"63e633cb-a49f-4adb-89f2-dddff70954b6\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-dczkk" Feb 27 08:12:26 crc kubenswrapper[4612]: I0227 08:12:26.692900 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/63e633cb-a49f-4adb-89f2-dddff70954b6-dns-swift-storage-0\") pod \"dnsmasq-dns-cd5cbd7b9-dczkk\" (UID: \"63e633cb-a49f-4adb-89f2-dddff70954b6\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-dczkk" Feb 27 08:12:26 crc kubenswrapper[4612]: I0227 08:12:26.717007 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lk7xj\" (UniqueName: \"kubernetes.io/projected/63e633cb-a49f-4adb-89f2-dddff70954b6-kube-api-access-lk7xj\") pod \"dnsmasq-dns-cd5cbd7b9-dczkk\" (UID: \"63e633cb-a49f-4adb-89f2-dddff70954b6\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-dczkk" Feb 27 08:12:26 crc kubenswrapper[4612]: I0227 08:12:26.888780 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cd5cbd7b9-dczkk" Feb 27 08:12:27 crc kubenswrapper[4612]: I0227 08:12:27.386412 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-cd5cbd7b9-dczkk"] Feb 27 08:12:27 crc kubenswrapper[4612]: W0227 08:12:27.395553 4612 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod63e633cb_a49f_4adb_89f2_dddff70954b6.slice/crio-f0f5b202d1415e736d16a149f2bb068b1a55fa5bf017d30829b3d46650fed7b7 WatchSource:0}: Error finding container f0f5b202d1415e736d16a149f2bb068b1a55fa5bf017d30829b3d46650fed7b7: Status 404 returned error can't find the container with id f0f5b202d1415e736d16a149f2bb068b1a55fa5bf017d30829b3d46650fed7b7 Feb 27 08:12:27 crc kubenswrapper[4612]: I0227 08:12:27.904133 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Feb 27 08:12:28 crc kubenswrapper[4612]: I0227 08:12:28.283996 4612 generic.go:334] "Generic (PLEG): container finished" podID="63e633cb-a49f-4adb-89f2-dddff70954b6" containerID="6f2784c8b6f934cebc87686e3e4cd3c59b6ab6954f42c1fb0f94b8e1f080e216" exitCode=0 Feb 27 08:12:28 crc kubenswrapper[4612]: I0227 08:12:28.284098 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cd5cbd7b9-dczkk" event={"ID":"63e633cb-a49f-4adb-89f2-dddff70954b6","Type":"ContainerDied","Data":"6f2784c8b6f934cebc87686e3e4cd3c59b6ab6954f42c1fb0f94b8e1f080e216"} Feb 27 08:12:28 crc kubenswrapper[4612]: I0227 08:12:28.284294 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cd5cbd7b9-dczkk" event={"ID":"63e633cb-a49f-4adb-89f2-dddff70954b6","Type":"ContainerStarted","Data":"f0f5b202d1415e736d16a149f2bb068b1a55fa5bf017d30829b3d46650fed7b7"} Feb 27 08:12:28 crc kubenswrapper[4612]: I0227 08:12:28.910073 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Feb 27 08:12:28 crc kubenswrapper[4612]: E0227 08:12:28.921410 4612 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3c4d5bb6_8ca7_41d7_8e30_af438a0e267a.slice/crio-b20dc16d14af8f07dddc21f5ce3968bf2ad89f62cb1e10b67ad0a10affdda173.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3c4d5bb6_8ca7_41d7_8e30_af438a0e267a.slice/crio-conmon-b20dc16d14af8f07dddc21f5ce3968bf2ad89f62cb1e10b67ad0a10affdda173.scope\": RecentStats: unable to find data in memory cache]" Feb 27 08:12:29 crc kubenswrapper[4612]: I0227 08:12:29.002268 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 27 08:12:29 crc kubenswrapper[4612]: I0227 08:12:29.002544 4612 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f6ca7ac0-4618-4a5b-af1a-7df5fa0e3701" containerName="ceilometer-central-agent" containerID="cri-o://dc0c3dd3d7b1c7ed0866ab1f050022d49c1a0f1c9020462c96b150e41116c369" gracePeriod=30 Feb 27 08:12:29 crc kubenswrapper[4612]: I0227 08:12:29.002658 4612 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f6ca7ac0-4618-4a5b-af1a-7df5fa0e3701" containerName="ceilometer-notification-agent" containerID="cri-o://01be7c2b40fa8ae2e3fc259c649bbe57b7ff91643eb9d4b0a6310dc040b990e8" gracePeriod=30 Feb 27 08:12:29 crc kubenswrapper[4612]: I0227 08:12:29.002679 4612 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f6ca7ac0-4618-4a5b-af1a-7df5fa0e3701" containerName="sg-core" containerID="cri-o://f70a60fbef6664a97cb3a6aea70caf097b359a8295be9fd0b3ce433a45ffd9db" gracePeriod=30 Feb 27 08:12:29 crc kubenswrapper[4612]: I0227 08:12:29.002870 4612 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f6ca7ac0-4618-4a5b-af1a-7df5fa0e3701" containerName="proxy-httpd" containerID="cri-o://4ff36f9dedc579a142a58de2fc0e4ceb22548dad53a45b4ad10082fe2343907e" gracePeriod=30 Feb 27 08:12:29 crc kubenswrapper[4612]: I0227 08:12:29.294856 4612 generic.go:334] "Generic (PLEG): container finished" podID="f6ca7ac0-4618-4a5b-af1a-7df5fa0e3701" containerID="4ff36f9dedc579a142a58de2fc0e4ceb22548dad53a45b4ad10082fe2343907e" exitCode=0 Feb 27 08:12:29 crc kubenswrapper[4612]: I0227 08:12:29.294896 4612 generic.go:334] "Generic (PLEG): container finished" podID="f6ca7ac0-4618-4a5b-af1a-7df5fa0e3701" containerID="f70a60fbef6664a97cb3a6aea70caf097b359a8295be9fd0b3ce433a45ffd9db" exitCode=2 Feb 27 08:12:29 crc kubenswrapper[4612]: I0227 08:12:29.294993 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f6ca7ac0-4618-4a5b-af1a-7df5fa0e3701","Type":"ContainerDied","Data":"4ff36f9dedc579a142a58de2fc0e4ceb22548dad53a45b4ad10082fe2343907e"} Feb 27 08:12:29 crc kubenswrapper[4612]: I0227 08:12:29.295025 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f6ca7ac0-4618-4a5b-af1a-7df5fa0e3701","Type":"ContainerDied","Data":"f70a60fbef6664a97cb3a6aea70caf097b359a8295be9fd0b3ce433a45ffd9db"} Feb 27 08:12:29 crc kubenswrapper[4612]: I0227 08:12:29.296630 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cd5cbd7b9-dczkk" event={"ID":"63e633cb-a49f-4adb-89f2-dddff70954b6","Type":"ContainerStarted","Data":"11796b4f0faab1b25a428fd96b25aa78aace1d64498d85d49d882191d4817abe"} Feb 27 08:12:29 crc kubenswrapper[4612]: I0227 08:12:29.296738 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-cd5cbd7b9-dczkk" Feb 27 08:12:29 crc kubenswrapper[4612]: I0227 08:12:29.297172 4612 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="d6aaa653-6c5a-4809-957b-6bdefa994b16" containerName="nova-api-log" containerID="cri-o://d78fb4b2a5a2ceeff13f9cc9454d569fedad601d2c42a08163d50af833bb6288" gracePeriod=30 Feb 27 08:12:29 crc kubenswrapper[4612]: I0227 08:12:29.297246 4612 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="d6aaa653-6c5a-4809-957b-6bdefa994b16" containerName="nova-api-api" containerID="cri-o://29b7410f9f42497aeb02a86bb006d6b74ad02aaf6af9b4aa7ad66a42e0b5faf8" gracePeriod=30 Feb 27 08:12:29 crc kubenswrapper[4612]: I0227 08:12:29.324509 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-cd5cbd7b9-dczkk" podStartSLOduration=3.324495609 podStartE2EDuration="3.324495609s" podCreationTimestamp="2026-02-27 08:12:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 08:12:29.317739525 +0000 UTC m=+1407.171669523" watchObservedRunningTime="2026-02-27 08:12:29.324495609 +0000 UTC m=+1407.178425607" Feb 27 08:12:30 crc kubenswrapper[4612]: I0227 08:12:30.309284 4612 generic.go:334] "Generic (PLEG): container finished" podID="f6ca7ac0-4618-4a5b-af1a-7df5fa0e3701" containerID="dc0c3dd3d7b1c7ed0866ab1f050022d49c1a0f1c9020462c96b150e41116c369" exitCode=0 Feb 27 08:12:30 crc kubenswrapper[4612]: I0227 08:12:30.309373 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f6ca7ac0-4618-4a5b-af1a-7df5fa0e3701","Type":"ContainerDied","Data":"dc0c3dd3d7b1c7ed0866ab1f050022d49c1a0f1c9020462c96b150e41116c369"} Feb 27 08:12:30 crc kubenswrapper[4612]: I0227 08:12:30.312319 4612 generic.go:334] "Generic (PLEG): container finished" podID="d6aaa653-6c5a-4809-957b-6bdefa994b16" containerID="d78fb4b2a5a2ceeff13f9cc9454d569fedad601d2c42a08163d50af833bb6288" exitCode=143 Feb 27 08:12:30 crc kubenswrapper[4612]: I0227 08:12:30.312473 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d6aaa653-6c5a-4809-957b-6bdefa994b16","Type":"ContainerDied","Data":"d78fb4b2a5a2ceeff13f9cc9454d569fedad601d2c42a08163d50af833bb6288"} Feb 27 08:12:32 crc kubenswrapper[4612]: I0227 08:12:32.904887 4612 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Feb 27 08:12:32 crc kubenswrapper[4612]: I0227 08:12:32.936146 4612 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Feb 27 08:12:33 crc kubenswrapper[4612]: I0227 08:12:33.115049 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 27 08:12:33 crc kubenswrapper[4612]: I0227 08:12:33.212390 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d6aaa653-6c5a-4809-957b-6bdefa994b16-logs\") pod \"d6aaa653-6c5a-4809-957b-6bdefa994b16\" (UID: \"d6aaa653-6c5a-4809-957b-6bdefa994b16\") " Feb 27 08:12:33 crc kubenswrapper[4612]: I0227 08:12:33.212454 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6aaa653-6c5a-4809-957b-6bdefa994b16-combined-ca-bundle\") pod \"d6aaa653-6c5a-4809-957b-6bdefa994b16\" (UID: \"d6aaa653-6c5a-4809-957b-6bdefa994b16\") " Feb 27 08:12:33 crc kubenswrapper[4612]: I0227 08:12:33.212539 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-glztw\" (UniqueName: \"kubernetes.io/projected/d6aaa653-6c5a-4809-957b-6bdefa994b16-kube-api-access-glztw\") pod \"d6aaa653-6c5a-4809-957b-6bdefa994b16\" (UID: \"d6aaa653-6c5a-4809-957b-6bdefa994b16\") " Feb 27 08:12:33 crc kubenswrapper[4612]: I0227 08:12:33.212716 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d6aaa653-6c5a-4809-957b-6bdefa994b16-config-data\") pod \"d6aaa653-6c5a-4809-957b-6bdefa994b16\" (UID: \"d6aaa653-6c5a-4809-957b-6bdefa994b16\") " Feb 27 08:12:33 crc kubenswrapper[4612]: I0227 08:12:33.216749 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d6aaa653-6c5a-4809-957b-6bdefa994b16-logs" (OuterVolumeSpecName: "logs") pod "d6aaa653-6c5a-4809-957b-6bdefa994b16" (UID: "d6aaa653-6c5a-4809-957b-6bdefa994b16"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 08:12:33 crc kubenswrapper[4612]: I0227 08:12:33.235944 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d6aaa653-6c5a-4809-957b-6bdefa994b16-kube-api-access-glztw" (OuterVolumeSpecName: "kube-api-access-glztw") pod "d6aaa653-6c5a-4809-957b-6bdefa994b16" (UID: "d6aaa653-6c5a-4809-957b-6bdefa994b16"). InnerVolumeSpecName "kube-api-access-glztw". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 08:12:33 crc kubenswrapper[4612]: I0227 08:12:33.257685 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d6aaa653-6c5a-4809-957b-6bdefa994b16-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d6aaa653-6c5a-4809-957b-6bdefa994b16" (UID: "d6aaa653-6c5a-4809-957b-6bdefa994b16"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:12:33 crc kubenswrapper[4612]: I0227 08:12:33.289003 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d6aaa653-6c5a-4809-957b-6bdefa994b16-config-data" (OuterVolumeSpecName: "config-data") pod "d6aaa653-6c5a-4809-957b-6bdefa994b16" (UID: "d6aaa653-6c5a-4809-957b-6bdefa994b16"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:12:33 crc kubenswrapper[4612]: I0227 08:12:33.311049 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 27 08:12:33 crc kubenswrapper[4612]: I0227 08:12:33.314399 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f6ca7ac0-4618-4a5b-af1a-7df5fa0e3701-run-httpd\") pod \"f6ca7ac0-4618-4a5b-af1a-7df5fa0e3701\" (UID: \"f6ca7ac0-4618-4a5b-af1a-7df5fa0e3701\") " Feb 27 08:12:33 crc kubenswrapper[4612]: I0227 08:12:33.314479 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6ca7ac0-4618-4a5b-af1a-7df5fa0e3701-combined-ca-bundle\") pod \"f6ca7ac0-4618-4a5b-af1a-7df5fa0e3701\" (UID: \"f6ca7ac0-4618-4a5b-af1a-7df5fa0e3701\") " Feb 27 08:12:33 crc kubenswrapper[4612]: I0227 08:12:33.314522 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f6ca7ac0-4618-4a5b-af1a-7df5fa0e3701-scripts\") pod \"f6ca7ac0-4618-4a5b-af1a-7df5fa0e3701\" (UID: \"f6ca7ac0-4618-4a5b-af1a-7df5fa0e3701\") " Feb 27 08:12:33 crc kubenswrapper[4612]: I0227 08:12:33.314566 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f6ca7ac0-4618-4a5b-af1a-7df5fa0e3701-sg-core-conf-yaml\") pod \"f6ca7ac0-4618-4a5b-af1a-7df5fa0e3701\" (UID: \"f6ca7ac0-4618-4a5b-af1a-7df5fa0e3701\") " Feb 27 08:12:33 crc kubenswrapper[4612]: I0227 08:12:33.314604 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z728g\" (UniqueName: \"kubernetes.io/projected/f6ca7ac0-4618-4a5b-af1a-7df5fa0e3701-kube-api-access-z728g\") pod \"f6ca7ac0-4618-4a5b-af1a-7df5fa0e3701\" (UID: \"f6ca7ac0-4618-4a5b-af1a-7df5fa0e3701\") " Feb 27 08:12:33 crc kubenswrapper[4612]: I0227 08:12:33.314730 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/f6ca7ac0-4618-4a5b-af1a-7df5fa0e3701-ceilometer-tls-certs\") pod \"f6ca7ac0-4618-4a5b-af1a-7df5fa0e3701\" (UID: \"f6ca7ac0-4618-4a5b-af1a-7df5fa0e3701\") " Feb 27 08:12:33 crc kubenswrapper[4612]: I0227 08:12:33.314798 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f6ca7ac0-4618-4a5b-af1a-7df5fa0e3701-config-data\") pod \"f6ca7ac0-4618-4a5b-af1a-7df5fa0e3701\" (UID: \"f6ca7ac0-4618-4a5b-af1a-7df5fa0e3701\") " Feb 27 08:12:33 crc kubenswrapper[4612]: I0227 08:12:33.314834 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f6ca7ac0-4618-4a5b-af1a-7df5fa0e3701-log-httpd\") pod \"f6ca7ac0-4618-4a5b-af1a-7df5fa0e3701\" (UID: \"f6ca7ac0-4618-4a5b-af1a-7df5fa0e3701\") " Feb 27 08:12:33 crc kubenswrapper[4612]: I0227 08:12:33.315272 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f6ca7ac0-4618-4a5b-af1a-7df5fa0e3701-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "f6ca7ac0-4618-4a5b-af1a-7df5fa0e3701" (UID: "f6ca7ac0-4618-4a5b-af1a-7df5fa0e3701"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 08:12:33 crc kubenswrapper[4612]: I0227 08:12:33.315278 4612 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d6aaa653-6c5a-4809-957b-6bdefa994b16-config-data\") on node \"crc\" DevicePath \"\"" Feb 27 08:12:33 crc kubenswrapper[4612]: I0227 08:12:33.315514 4612 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d6aaa653-6c5a-4809-957b-6bdefa994b16-logs\") on node \"crc\" DevicePath \"\"" Feb 27 08:12:33 crc kubenswrapper[4612]: I0227 08:12:33.315543 4612 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6aaa653-6c5a-4809-957b-6bdefa994b16-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 08:12:33 crc kubenswrapper[4612]: I0227 08:12:33.315560 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-glztw\" (UniqueName: \"kubernetes.io/projected/d6aaa653-6c5a-4809-957b-6bdefa994b16-kube-api-access-glztw\") on node \"crc\" DevicePath \"\"" Feb 27 08:12:33 crc kubenswrapper[4612]: I0227 08:12:33.316158 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f6ca7ac0-4618-4a5b-af1a-7df5fa0e3701-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "f6ca7ac0-4618-4a5b-af1a-7df5fa0e3701" (UID: "f6ca7ac0-4618-4a5b-af1a-7df5fa0e3701"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 08:12:33 crc kubenswrapper[4612]: I0227 08:12:33.330993 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f6ca7ac0-4618-4a5b-af1a-7df5fa0e3701-scripts" (OuterVolumeSpecName: "scripts") pod "f6ca7ac0-4618-4a5b-af1a-7df5fa0e3701" (UID: "f6ca7ac0-4618-4a5b-af1a-7df5fa0e3701"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:12:33 crc kubenswrapper[4612]: I0227 08:12:33.342009 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f6ca7ac0-4618-4a5b-af1a-7df5fa0e3701-kube-api-access-z728g" (OuterVolumeSpecName: "kube-api-access-z728g") pod "f6ca7ac0-4618-4a5b-af1a-7df5fa0e3701" (UID: "f6ca7ac0-4618-4a5b-af1a-7df5fa0e3701"). InnerVolumeSpecName "kube-api-access-z728g". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 08:12:33 crc kubenswrapper[4612]: I0227 08:12:33.360585 4612 generic.go:334] "Generic (PLEG): container finished" podID="f6ca7ac0-4618-4a5b-af1a-7df5fa0e3701" containerID="01be7c2b40fa8ae2e3fc259c649bbe57b7ff91643eb9d4b0a6310dc040b990e8" exitCode=0 Feb 27 08:12:33 crc kubenswrapper[4612]: I0227 08:12:33.360643 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f6ca7ac0-4618-4a5b-af1a-7df5fa0e3701","Type":"ContainerDied","Data":"01be7c2b40fa8ae2e3fc259c649bbe57b7ff91643eb9d4b0a6310dc040b990e8"} Feb 27 08:12:33 crc kubenswrapper[4612]: I0227 08:12:33.360667 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f6ca7ac0-4618-4a5b-af1a-7df5fa0e3701","Type":"ContainerDied","Data":"df40f175d4e13aaf29051c5a619d63514d77a1c982646acb8391a1eed8cc83b5"} Feb 27 08:12:33 crc kubenswrapper[4612]: I0227 08:12:33.360683 4612 scope.go:117] "RemoveContainer" containerID="4ff36f9dedc579a142a58de2fc0e4ceb22548dad53a45b4ad10082fe2343907e" Feb 27 08:12:33 crc kubenswrapper[4612]: I0227 08:12:33.360780 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f6ca7ac0-4618-4a5b-af1a-7df5fa0e3701-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "f6ca7ac0-4618-4a5b-af1a-7df5fa0e3701" (UID: "f6ca7ac0-4618-4a5b-af1a-7df5fa0e3701"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:12:33 crc kubenswrapper[4612]: I0227 08:12:33.360838 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 27 08:12:33 crc kubenswrapper[4612]: I0227 08:12:33.368068 4612 generic.go:334] "Generic (PLEG): container finished" podID="d6aaa653-6c5a-4809-957b-6bdefa994b16" containerID="29b7410f9f42497aeb02a86bb006d6b74ad02aaf6af9b4aa7ad66a42e0b5faf8" exitCode=0 Feb 27 08:12:33 crc kubenswrapper[4612]: I0227 08:12:33.368609 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 27 08:12:33 crc kubenswrapper[4612]: I0227 08:12:33.369081 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d6aaa653-6c5a-4809-957b-6bdefa994b16","Type":"ContainerDied","Data":"29b7410f9f42497aeb02a86bb006d6b74ad02aaf6af9b4aa7ad66a42e0b5faf8"} Feb 27 08:12:33 crc kubenswrapper[4612]: I0227 08:12:33.369142 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d6aaa653-6c5a-4809-957b-6bdefa994b16","Type":"ContainerDied","Data":"3da8467bc27b960ab3f5c5e2ca8f4dbd708a2c9ae862bd6229372a0e8c31c5ba"} Feb 27 08:12:33 crc kubenswrapper[4612]: I0227 08:12:33.396582 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Feb 27 08:12:33 crc kubenswrapper[4612]: I0227 08:12:33.416288 4612 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f6ca7ac0-4618-4a5b-af1a-7df5fa0e3701-log-httpd\") on node \"crc\" DevicePath \"\"" Feb 27 08:12:33 crc kubenswrapper[4612]: I0227 08:12:33.416323 4612 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f6ca7ac0-4618-4a5b-af1a-7df5fa0e3701-run-httpd\") on node \"crc\" DevicePath \"\"" Feb 27 08:12:33 crc kubenswrapper[4612]: I0227 08:12:33.416336 4612 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f6ca7ac0-4618-4a5b-af1a-7df5fa0e3701-scripts\") on node \"crc\" DevicePath \"\"" Feb 27 08:12:33 crc kubenswrapper[4612]: I0227 08:12:33.416346 4612 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f6ca7ac0-4618-4a5b-af1a-7df5fa0e3701-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Feb 27 08:12:33 crc kubenswrapper[4612]: I0227 08:12:33.416355 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z728g\" (UniqueName: \"kubernetes.io/projected/f6ca7ac0-4618-4a5b-af1a-7df5fa0e3701-kube-api-access-z728g\") on node \"crc\" DevicePath \"\"" Feb 27 08:12:33 crc kubenswrapper[4612]: I0227 08:12:33.448618 4612 scope.go:117] "RemoveContainer" containerID="f70a60fbef6664a97cb3a6aea70caf097b359a8295be9fd0b3ce433a45ffd9db" Feb 27 08:12:33 crc kubenswrapper[4612]: I0227 08:12:33.458965 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Feb 27 08:12:33 crc kubenswrapper[4612]: I0227 08:12:33.469852 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Feb 27 08:12:33 crc kubenswrapper[4612]: I0227 08:12:33.484944 4612 scope.go:117] "RemoveContainer" containerID="01be7c2b40fa8ae2e3fc259c649bbe57b7ff91643eb9d4b0a6310dc040b990e8" Feb 27 08:12:33 crc kubenswrapper[4612]: I0227 08:12:33.496890 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f6ca7ac0-4618-4a5b-af1a-7df5fa0e3701-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "f6ca7ac0-4618-4a5b-af1a-7df5fa0e3701" (UID: "f6ca7ac0-4618-4a5b-af1a-7df5fa0e3701"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:12:33 crc kubenswrapper[4612]: I0227 08:12:33.499167 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f6ca7ac0-4618-4a5b-af1a-7df5fa0e3701-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f6ca7ac0-4618-4a5b-af1a-7df5fa0e3701" (UID: "f6ca7ac0-4618-4a5b-af1a-7df5fa0e3701"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:12:33 crc kubenswrapper[4612]: I0227 08:12:33.496624 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Feb 27 08:12:33 crc kubenswrapper[4612]: E0227 08:12:33.501323 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d6aaa653-6c5a-4809-957b-6bdefa994b16" containerName="nova-api-log" Feb 27 08:12:33 crc kubenswrapper[4612]: I0227 08:12:33.501342 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="d6aaa653-6c5a-4809-957b-6bdefa994b16" containerName="nova-api-log" Feb 27 08:12:33 crc kubenswrapper[4612]: E0227 08:12:33.501358 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f6ca7ac0-4618-4a5b-af1a-7df5fa0e3701" containerName="proxy-httpd" Feb 27 08:12:33 crc kubenswrapper[4612]: I0227 08:12:33.501366 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="f6ca7ac0-4618-4a5b-af1a-7df5fa0e3701" containerName="proxy-httpd" Feb 27 08:12:33 crc kubenswrapper[4612]: E0227 08:12:33.501387 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f6ca7ac0-4618-4a5b-af1a-7df5fa0e3701" containerName="sg-core" Feb 27 08:12:33 crc kubenswrapper[4612]: I0227 08:12:33.501395 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="f6ca7ac0-4618-4a5b-af1a-7df5fa0e3701" containerName="sg-core" Feb 27 08:12:33 crc kubenswrapper[4612]: E0227 08:12:33.501419 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f6ca7ac0-4618-4a5b-af1a-7df5fa0e3701" containerName="ceilometer-central-agent" Feb 27 08:12:33 crc kubenswrapper[4612]: I0227 08:12:33.501428 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="f6ca7ac0-4618-4a5b-af1a-7df5fa0e3701" containerName="ceilometer-central-agent" Feb 27 08:12:33 crc kubenswrapper[4612]: E0227 08:12:33.501453 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d6aaa653-6c5a-4809-957b-6bdefa994b16" containerName="nova-api-api" Feb 27 08:12:33 crc kubenswrapper[4612]: I0227 08:12:33.501460 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="d6aaa653-6c5a-4809-957b-6bdefa994b16" containerName="nova-api-api" Feb 27 08:12:33 crc kubenswrapper[4612]: E0227 08:12:33.501475 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f6ca7ac0-4618-4a5b-af1a-7df5fa0e3701" containerName="ceilometer-notification-agent" Feb 27 08:12:33 crc kubenswrapper[4612]: I0227 08:12:33.501483 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="f6ca7ac0-4618-4a5b-af1a-7df5fa0e3701" containerName="ceilometer-notification-agent" Feb 27 08:12:33 crc kubenswrapper[4612]: I0227 08:12:33.501671 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="d6aaa653-6c5a-4809-957b-6bdefa994b16" containerName="nova-api-api" Feb 27 08:12:33 crc kubenswrapper[4612]: I0227 08:12:33.501680 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="f6ca7ac0-4618-4a5b-af1a-7df5fa0e3701" containerName="sg-core" Feb 27 08:12:33 crc kubenswrapper[4612]: I0227 08:12:33.501702 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="d6aaa653-6c5a-4809-957b-6bdefa994b16" containerName="nova-api-log" Feb 27 08:12:33 crc kubenswrapper[4612]: I0227 08:12:33.501719 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="f6ca7ac0-4618-4a5b-af1a-7df5fa0e3701" containerName="proxy-httpd" Feb 27 08:12:33 crc kubenswrapper[4612]: I0227 08:12:33.501733 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="f6ca7ac0-4618-4a5b-af1a-7df5fa0e3701" containerName="ceilometer-notification-agent" Feb 27 08:12:33 crc kubenswrapper[4612]: I0227 08:12:33.501742 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="f6ca7ac0-4618-4a5b-af1a-7df5fa0e3701" containerName="ceilometer-central-agent" Feb 27 08:12:33 crc kubenswrapper[4612]: I0227 08:12:33.502678 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 27 08:12:33 crc kubenswrapper[4612]: I0227 08:12:33.508434 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Feb 27 08:12:33 crc kubenswrapper[4612]: I0227 08:12:33.508745 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Feb 27 08:12:33 crc kubenswrapper[4612]: I0227 08:12:33.509282 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Feb 27 08:12:33 crc kubenswrapper[4612]: I0227 08:12:33.519076 4612 scope.go:117] "RemoveContainer" containerID="dc0c3dd3d7b1c7ed0866ab1f050022d49c1a0f1c9020462c96b150e41116c369" Feb 27 08:12:33 crc kubenswrapper[4612]: I0227 08:12:33.519619 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f6ca7ac0-4618-4a5b-af1a-7df5fa0e3701-config-data" (OuterVolumeSpecName: "config-data") pod "f6ca7ac0-4618-4a5b-af1a-7df5fa0e3701" (UID: "f6ca7ac0-4618-4a5b-af1a-7df5fa0e3701"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:12:33 crc kubenswrapper[4612]: I0227 08:12:33.550195 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Feb 27 08:12:33 crc kubenswrapper[4612]: I0227 08:12:33.557336 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7690cb3e-e335-4dc2-bf76-0a3239bc6a5c-public-tls-certs\") pod \"nova-api-0\" (UID: \"7690cb3e-e335-4dc2-bf76-0a3239bc6a5c\") " pod="openstack/nova-api-0" Feb 27 08:12:33 crc kubenswrapper[4612]: I0227 08:12:33.557413 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7690cb3e-e335-4dc2-bf76-0a3239bc6a5c-logs\") pod \"nova-api-0\" (UID: \"7690cb3e-e335-4dc2-bf76-0a3239bc6a5c\") " pod="openstack/nova-api-0" Feb 27 08:12:33 crc kubenswrapper[4612]: I0227 08:12:33.557511 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7690cb3e-e335-4dc2-bf76-0a3239bc6a5c-internal-tls-certs\") pod \"nova-api-0\" (UID: \"7690cb3e-e335-4dc2-bf76-0a3239bc6a5c\") " pod="openstack/nova-api-0" Feb 27 08:12:33 crc kubenswrapper[4612]: I0227 08:12:33.557597 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7690cb3e-e335-4dc2-bf76-0a3239bc6a5c-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"7690cb3e-e335-4dc2-bf76-0a3239bc6a5c\") " pod="openstack/nova-api-0" Feb 27 08:12:33 crc kubenswrapper[4612]: I0227 08:12:33.557711 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xxh4z\" (UniqueName: \"kubernetes.io/projected/7690cb3e-e335-4dc2-bf76-0a3239bc6a5c-kube-api-access-xxh4z\") pod \"nova-api-0\" (UID: \"7690cb3e-e335-4dc2-bf76-0a3239bc6a5c\") " pod="openstack/nova-api-0" Feb 27 08:12:33 crc kubenswrapper[4612]: I0227 08:12:33.557765 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7690cb3e-e335-4dc2-bf76-0a3239bc6a5c-config-data\") pod \"nova-api-0\" (UID: \"7690cb3e-e335-4dc2-bf76-0a3239bc6a5c\") " pod="openstack/nova-api-0" Feb 27 08:12:33 crc kubenswrapper[4612]: I0227 08:12:33.557976 4612 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6ca7ac0-4618-4a5b-af1a-7df5fa0e3701-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 08:12:33 crc kubenswrapper[4612]: I0227 08:12:33.557994 4612 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/f6ca7ac0-4618-4a5b-af1a-7df5fa0e3701-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 27 08:12:33 crc kubenswrapper[4612]: I0227 08:12:33.558004 4612 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f6ca7ac0-4618-4a5b-af1a-7df5fa0e3701-config-data\") on node \"crc\" DevicePath \"\"" Feb 27 08:12:33 crc kubenswrapper[4612]: I0227 08:12:33.655764 4612 scope.go:117] "RemoveContainer" containerID="4ff36f9dedc579a142a58de2fc0e4ceb22548dad53a45b4ad10082fe2343907e" Feb 27 08:12:33 crc kubenswrapper[4612]: E0227 08:12:33.669670 4612 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4ff36f9dedc579a142a58de2fc0e4ceb22548dad53a45b4ad10082fe2343907e\": container with ID starting with 4ff36f9dedc579a142a58de2fc0e4ceb22548dad53a45b4ad10082fe2343907e not found: ID does not exist" containerID="4ff36f9dedc579a142a58de2fc0e4ceb22548dad53a45b4ad10082fe2343907e" Feb 27 08:12:33 crc kubenswrapper[4612]: I0227 08:12:33.669768 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4ff36f9dedc579a142a58de2fc0e4ceb22548dad53a45b4ad10082fe2343907e"} err="failed to get container status \"4ff36f9dedc579a142a58de2fc0e4ceb22548dad53a45b4ad10082fe2343907e\": rpc error: code = NotFound desc = could not find container \"4ff36f9dedc579a142a58de2fc0e4ceb22548dad53a45b4ad10082fe2343907e\": container with ID starting with 4ff36f9dedc579a142a58de2fc0e4ceb22548dad53a45b4ad10082fe2343907e not found: ID does not exist" Feb 27 08:12:33 crc kubenswrapper[4612]: I0227 08:12:33.669808 4612 scope.go:117] "RemoveContainer" containerID="f70a60fbef6664a97cb3a6aea70caf097b359a8295be9fd0b3ce433a45ffd9db" Feb 27 08:12:33 crc kubenswrapper[4612]: I0227 08:12:33.670615 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7690cb3e-e335-4dc2-bf76-0a3239bc6a5c-public-tls-certs\") pod \"nova-api-0\" (UID: \"7690cb3e-e335-4dc2-bf76-0a3239bc6a5c\") " pod="openstack/nova-api-0" Feb 27 08:12:33 crc kubenswrapper[4612]: I0227 08:12:33.670664 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7690cb3e-e335-4dc2-bf76-0a3239bc6a5c-logs\") pod \"nova-api-0\" (UID: \"7690cb3e-e335-4dc2-bf76-0a3239bc6a5c\") " pod="openstack/nova-api-0" Feb 27 08:12:33 crc kubenswrapper[4612]: I0227 08:12:33.670754 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7690cb3e-e335-4dc2-bf76-0a3239bc6a5c-internal-tls-certs\") pod \"nova-api-0\" (UID: \"7690cb3e-e335-4dc2-bf76-0a3239bc6a5c\") " pod="openstack/nova-api-0" Feb 27 08:12:33 crc kubenswrapper[4612]: I0227 08:12:33.670833 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7690cb3e-e335-4dc2-bf76-0a3239bc6a5c-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"7690cb3e-e335-4dc2-bf76-0a3239bc6a5c\") " pod="openstack/nova-api-0" Feb 27 08:12:33 crc kubenswrapper[4612]: I0227 08:12:33.670943 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xxh4z\" (UniqueName: \"kubernetes.io/projected/7690cb3e-e335-4dc2-bf76-0a3239bc6a5c-kube-api-access-xxh4z\") pod \"nova-api-0\" (UID: \"7690cb3e-e335-4dc2-bf76-0a3239bc6a5c\") " pod="openstack/nova-api-0" Feb 27 08:12:33 crc kubenswrapper[4612]: I0227 08:12:33.671026 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7690cb3e-e335-4dc2-bf76-0a3239bc6a5c-config-data\") pod \"nova-api-0\" (UID: \"7690cb3e-e335-4dc2-bf76-0a3239bc6a5c\") " pod="openstack/nova-api-0" Feb 27 08:12:33 crc kubenswrapper[4612]: E0227 08:12:33.672675 4612 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f70a60fbef6664a97cb3a6aea70caf097b359a8295be9fd0b3ce433a45ffd9db\": container with ID starting with f70a60fbef6664a97cb3a6aea70caf097b359a8295be9fd0b3ce433a45ffd9db not found: ID does not exist" containerID="f70a60fbef6664a97cb3a6aea70caf097b359a8295be9fd0b3ce433a45ffd9db" Feb 27 08:12:33 crc kubenswrapper[4612]: I0227 08:12:33.672732 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f70a60fbef6664a97cb3a6aea70caf097b359a8295be9fd0b3ce433a45ffd9db"} err="failed to get container status \"f70a60fbef6664a97cb3a6aea70caf097b359a8295be9fd0b3ce433a45ffd9db\": rpc error: code = NotFound desc = could not find container \"f70a60fbef6664a97cb3a6aea70caf097b359a8295be9fd0b3ce433a45ffd9db\": container with ID starting with f70a60fbef6664a97cb3a6aea70caf097b359a8295be9fd0b3ce433a45ffd9db not found: ID does not exist" Feb 27 08:12:33 crc kubenswrapper[4612]: I0227 08:12:33.672796 4612 scope.go:117] "RemoveContainer" containerID="01be7c2b40fa8ae2e3fc259c649bbe57b7ff91643eb9d4b0a6310dc040b990e8" Feb 27 08:12:33 crc kubenswrapper[4612]: E0227 08:12:33.674199 4612 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"01be7c2b40fa8ae2e3fc259c649bbe57b7ff91643eb9d4b0a6310dc040b990e8\": container with ID starting with 01be7c2b40fa8ae2e3fc259c649bbe57b7ff91643eb9d4b0a6310dc040b990e8 not found: ID does not exist" containerID="01be7c2b40fa8ae2e3fc259c649bbe57b7ff91643eb9d4b0a6310dc040b990e8" Feb 27 08:12:33 crc kubenswrapper[4612]: I0227 08:12:33.674235 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"01be7c2b40fa8ae2e3fc259c649bbe57b7ff91643eb9d4b0a6310dc040b990e8"} err="failed to get container status \"01be7c2b40fa8ae2e3fc259c649bbe57b7ff91643eb9d4b0a6310dc040b990e8\": rpc error: code = NotFound desc = could not find container \"01be7c2b40fa8ae2e3fc259c649bbe57b7ff91643eb9d4b0a6310dc040b990e8\": container with ID starting with 01be7c2b40fa8ae2e3fc259c649bbe57b7ff91643eb9d4b0a6310dc040b990e8 not found: ID does not exist" Feb 27 08:12:33 crc kubenswrapper[4612]: I0227 08:12:33.674255 4612 scope.go:117] "RemoveContainer" containerID="dc0c3dd3d7b1c7ed0866ab1f050022d49c1a0f1c9020462c96b150e41116c369" Feb 27 08:12:33 crc kubenswrapper[4612]: I0227 08:12:33.674869 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7690cb3e-e335-4dc2-bf76-0a3239bc6a5c-logs\") pod \"nova-api-0\" (UID: \"7690cb3e-e335-4dc2-bf76-0a3239bc6a5c\") " pod="openstack/nova-api-0" Feb 27 08:12:33 crc kubenswrapper[4612]: E0227 08:12:33.674983 4612 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dc0c3dd3d7b1c7ed0866ab1f050022d49c1a0f1c9020462c96b150e41116c369\": container with ID starting with dc0c3dd3d7b1c7ed0866ab1f050022d49c1a0f1c9020462c96b150e41116c369 not found: ID does not exist" containerID="dc0c3dd3d7b1c7ed0866ab1f050022d49c1a0f1c9020462c96b150e41116c369" Feb 27 08:12:33 crc kubenswrapper[4612]: I0227 08:12:33.675004 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dc0c3dd3d7b1c7ed0866ab1f050022d49c1a0f1c9020462c96b150e41116c369"} err="failed to get container status \"dc0c3dd3d7b1c7ed0866ab1f050022d49c1a0f1c9020462c96b150e41116c369\": rpc error: code = NotFound desc = could not find container \"dc0c3dd3d7b1c7ed0866ab1f050022d49c1a0f1c9020462c96b150e41116c369\": container with ID starting with dc0c3dd3d7b1c7ed0866ab1f050022d49c1a0f1c9020462c96b150e41116c369 not found: ID does not exist" Feb 27 08:12:33 crc kubenswrapper[4612]: I0227 08:12:33.675020 4612 scope.go:117] "RemoveContainer" containerID="29b7410f9f42497aeb02a86bb006d6b74ad02aaf6af9b4aa7ad66a42e0b5faf8" Feb 27 08:12:33 crc kubenswrapper[4612]: I0227 08:12:33.679268 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-7jx4l"] Feb 27 08:12:33 crc kubenswrapper[4612]: I0227 08:12:33.679302 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7690cb3e-e335-4dc2-bf76-0a3239bc6a5c-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"7690cb3e-e335-4dc2-bf76-0a3239bc6a5c\") " pod="openstack/nova-api-0" Feb 27 08:12:33 crc kubenswrapper[4612]: I0227 08:12:33.680547 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-7jx4l" Feb 27 08:12:33 crc kubenswrapper[4612]: I0227 08:12:33.689113 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Feb 27 08:12:33 crc kubenswrapper[4612]: I0227 08:12:33.689310 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Feb 27 08:12:33 crc kubenswrapper[4612]: I0227 08:12:33.699397 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7690cb3e-e335-4dc2-bf76-0a3239bc6a5c-config-data\") pod \"nova-api-0\" (UID: \"7690cb3e-e335-4dc2-bf76-0a3239bc6a5c\") " pod="openstack/nova-api-0" Feb 27 08:12:33 crc kubenswrapper[4612]: I0227 08:12:33.706374 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-7jx4l"] Feb 27 08:12:33 crc kubenswrapper[4612]: I0227 08:12:33.723350 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7690cb3e-e335-4dc2-bf76-0a3239bc6a5c-public-tls-certs\") pod \"nova-api-0\" (UID: \"7690cb3e-e335-4dc2-bf76-0a3239bc6a5c\") " pod="openstack/nova-api-0" Feb 27 08:12:33 crc kubenswrapper[4612]: I0227 08:12:33.723933 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7690cb3e-e335-4dc2-bf76-0a3239bc6a5c-internal-tls-certs\") pod \"nova-api-0\" (UID: \"7690cb3e-e335-4dc2-bf76-0a3239bc6a5c\") " pod="openstack/nova-api-0" Feb 27 08:12:33 crc kubenswrapper[4612]: I0227 08:12:33.724096 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xxh4z\" (UniqueName: \"kubernetes.io/projected/7690cb3e-e335-4dc2-bf76-0a3239bc6a5c-kube-api-access-xxh4z\") pod \"nova-api-0\" (UID: \"7690cb3e-e335-4dc2-bf76-0a3239bc6a5c\") " pod="openstack/nova-api-0" Feb 27 08:12:33 crc kubenswrapper[4612]: I0227 08:12:33.771829 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/16afe9fc-ff42-487d-9c1a-57f93c0a242c-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-7jx4l\" (UID: \"16afe9fc-ff42-487d-9c1a-57f93c0a242c\") " pod="openstack/nova-cell1-cell-mapping-7jx4l" Feb 27 08:12:33 crc kubenswrapper[4612]: I0227 08:12:33.772309 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ddl5r\" (UniqueName: \"kubernetes.io/projected/16afe9fc-ff42-487d-9c1a-57f93c0a242c-kube-api-access-ddl5r\") pod \"nova-cell1-cell-mapping-7jx4l\" (UID: \"16afe9fc-ff42-487d-9c1a-57f93c0a242c\") " pod="openstack/nova-cell1-cell-mapping-7jx4l" Feb 27 08:12:33 crc kubenswrapper[4612]: I0227 08:12:33.772452 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/16afe9fc-ff42-487d-9c1a-57f93c0a242c-scripts\") pod \"nova-cell1-cell-mapping-7jx4l\" (UID: \"16afe9fc-ff42-487d-9c1a-57f93c0a242c\") " pod="openstack/nova-cell1-cell-mapping-7jx4l" Feb 27 08:12:33 crc kubenswrapper[4612]: I0227 08:12:33.772502 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/16afe9fc-ff42-487d-9c1a-57f93c0a242c-config-data\") pod \"nova-cell1-cell-mapping-7jx4l\" (UID: \"16afe9fc-ff42-487d-9c1a-57f93c0a242c\") " pod="openstack/nova-cell1-cell-mapping-7jx4l" Feb 27 08:12:33 crc kubenswrapper[4612]: I0227 08:12:33.778880 4612 scope.go:117] "RemoveContainer" containerID="d78fb4b2a5a2ceeff13f9cc9454d569fedad601d2c42a08163d50af833bb6288" Feb 27 08:12:33 crc kubenswrapper[4612]: I0227 08:12:33.800777 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Feb 27 08:12:33 crc kubenswrapper[4612]: I0227 08:12:33.804497 4612 scope.go:117] "RemoveContainer" containerID="29b7410f9f42497aeb02a86bb006d6b74ad02aaf6af9b4aa7ad66a42e0b5faf8" Feb 27 08:12:33 crc kubenswrapper[4612]: E0227 08:12:33.804866 4612 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"29b7410f9f42497aeb02a86bb006d6b74ad02aaf6af9b4aa7ad66a42e0b5faf8\": container with ID starting with 29b7410f9f42497aeb02a86bb006d6b74ad02aaf6af9b4aa7ad66a42e0b5faf8 not found: ID does not exist" containerID="29b7410f9f42497aeb02a86bb006d6b74ad02aaf6af9b4aa7ad66a42e0b5faf8" Feb 27 08:12:33 crc kubenswrapper[4612]: I0227 08:12:33.804895 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"29b7410f9f42497aeb02a86bb006d6b74ad02aaf6af9b4aa7ad66a42e0b5faf8"} err="failed to get container status \"29b7410f9f42497aeb02a86bb006d6b74ad02aaf6af9b4aa7ad66a42e0b5faf8\": rpc error: code = NotFound desc = could not find container \"29b7410f9f42497aeb02a86bb006d6b74ad02aaf6af9b4aa7ad66a42e0b5faf8\": container with ID starting with 29b7410f9f42497aeb02a86bb006d6b74ad02aaf6af9b4aa7ad66a42e0b5faf8 not found: ID does not exist" Feb 27 08:12:33 crc kubenswrapper[4612]: I0227 08:12:33.804916 4612 scope.go:117] "RemoveContainer" containerID="d78fb4b2a5a2ceeff13f9cc9454d569fedad601d2c42a08163d50af833bb6288" Feb 27 08:12:33 crc kubenswrapper[4612]: E0227 08:12:33.805761 4612 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d78fb4b2a5a2ceeff13f9cc9454d569fedad601d2c42a08163d50af833bb6288\": container with ID starting with d78fb4b2a5a2ceeff13f9cc9454d569fedad601d2c42a08163d50af833bb6288 not found: ID does not exist" containerID="d78fb4b2a5a2ceeff13f9cc9454d569fedad601d2c42a08163d50af833bb6288" Feb 27 08:12:33 crc kubenswrapper[4612]: I0227 08:12:33.806226 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d78fb4b2a5a2ceeff13f9cc9454d569fedad601d2c42a08163d50af833bb6288"} err="failed to get container status \"d78fb4b2a5a2ceeff13f9cc9454d569fedad601d2c42a08163d50af833bb6288\": rpc error: code = NotFound desc = could not find container \"d78fb4b2a5a2ceeff13f9cc9454d569fedad601d2c42a08163d50af833bb6288\": container with ID starting with d78fb4b2a5a2ceeff13f9cc9454d569fedad601d2c42a08163d50af833bb6288 not found: ID does not exist" Feb 27 08:12:33 crc kubenswrapper[4612]: I0227 08:12:33.836627 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Feb 27 08:12:33 crc kubenswrapper[4612]: I0227 08:12:33.855756 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Feb 27 08:12:33 crc kubenswrapper[4612]: I0227 08:12:33.858136 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 27 08:12:33 crc kubenswrapper[4612]: I0227 08:12:33.858236 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 27 08:12:33 crc kubenswrapper[4612]: I0227 08:12:33.865585 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Feb 27 08:12:33 crc kubenswrapper[4612]: I0227 08:12:33.866382 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Feb 27 08:12:33 crc kubenswrapper[4612]: I0227 08:12:33.866523 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Feb 27 08:12:33 crc kubenswrapper[4612]: I0227 08:12:33.873601 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/16ab2b55-65e7-4d63-b521-0c6421ba52ea-log-httpd\") pod \"ceilometer-0\" (UID: \"16ab2b55-65e7-4d63-b521-0c6421ba52ea\") " pod="openstack/ceilometer-0" Feb 27 08:12:33 crc kubenswrapper[4612]: I0227 08:12:33.873646 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/16ab2b55-65e7-4d63-b521-0c6421ba52ea-config-data\") pod \"ceilometer-0\" (UID: \"16ab2b55-65e7-4d63-b521-0c6421ba52ea\") " pod="openstack/ceilometer-0" Feb 27 08:12:33 crc kubenswrapper[4612]: I0227 08:12:33.874680 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/16ab2b55-65e7-4d63-b521-0c6421ba52ea-scripts\") pod \"ceilometer-0\" (UID: \"16ab2b55-65e7-4d63-b521-0c6421ba52ea\") " pod="openstack/ceilometer-0" Feb 27 08:12:33 crc kubenswrapper[4612]: I0227 08:12:33.874735 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ddl5r\" (UniqueName: \"kubernetes.io/projected/16afe9fc-ff42-487d-9c1a-57f93c0a242c-kube-api-access-ddl5r\") pod \"nova-cell1-cell-mapping-7jx4l\" (UID: \"16afe9fc-ff42-487d-9c1a-57f93c0a242c\") " pod="openstack/nova-cell1-cell-mapping-7jx4l" Feb 27 08:12:33 crc kubenswrapper[4612]: I0227 08:12:33.874793 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/16afe9fc-ff42-487d-9c1a-57f93c0a242c-scripts\") pod \"nova-cell1-cell-mapping-7jx4l\" (UID: \"16afe9fc-ff42-487d-9c1a-57f93c0a242c\") " pod="openstack/nova-cell1-cell-mapping-7jx4l" Feb 27 08:12:33 crc kubenswrapper[4612]: I0227 08:12:33.874823 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/16ab2b55-65e7-4d63-b521-0c6421ba52ea-run-httpd\") pod \"ceilometer-0\" (UID: \"16ab2b55-65e7-4d63-b521-0c6421ba52ea\") " pod="openstack/ceilometer-0" Feb 27 08:12:33 crc kubenswrapper[4612]: I0227 08:12:33.874845 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/16afe9fc-ff42-487d-9c1a-57f93c0a242c-config-data\") pod \"nova-cell1-cell-mapping-7jx4l\" (UID: \"16afe9fc-ff42-487d-9c1a-57f93c0a242c\") " pod="openstack/nova-cell1-cell-mapping-7jx4l" Feb 27 08:12:33 crc kubenswrapper[4612]: I0227 08:12:33.874903 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/16afe9fc-ff42-487d-9c1a-57f93c0a242c-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-7jx4l\" (UID: \"16afe9fc-ff42-487d-9c1a-57f93c0a242c\") " pod="openstack/nova-cell1-cell-mapping-7jx4l" Feb 27 08:12:33 crc kubenswrapper[4612]: I0227 08:12:33.874926 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/16ab2b55-65e7-4d63-b521-0c6421ba52ea-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"16ab2b55-65e7-4d63-b521-0c6421ba52ea\") " pod="openstack/ceilometer-0" Feb 27 08:12:33 crc kubenswrapper[4612]: I0227 08:12:33.874942 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/16ab2b55-65e7-4d63-b521-0c6421ba52ea-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"16ab2b55-65e7-4d63-b521-0c6421ba52ea\") " pod="openstack/ceilometer-0" Feb 27 08:12:33 crc kubenswrapper[4612]: I0227 08:12:33.874961 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/16ab2b55-65e7-4d63-b521-0c6421ba52ea-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"16ab2b55-65e7-4d63-b521-0c6421ba52ea\") " pod="openstack/ceilometer-0" Feb 27 08:12:33 crc kubenswrapper[4612]: I0227 08:12:33.874987 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nkq6q\" (UniqueName: \"kubernetes.io/projected/16ab2b55-65e7-4d63-b521-0c6421ba52ea-kube-api-access-nkq6q\") pod \"ceilometer-0\" (UID: \"16ab2b55-65e7-4d63-b521-0c6421ba52ea\") " pod="openstack/ceilometer-0" Feb 27 08:12:33 crc kubenswrapper[4612]: I0227 08:12:33.878571 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/16afe9fc-ff42-487d-9c1a-57f93c0a242c-scripts\") pod \"nova-cell1-cell-mapping-7jx4l\" (UID: \"16afe9fc-ff42-487d-9c1a-57f93c0a242c\") " pod="openstack/nova-cell1-cell-mapping-7jx4l" Feb 27 08:12:33 crc kubenswrapper[4612]: I0227 08:12:33.886308 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/16afe9fc-ff42-487d-9c1a-57f93c0a242c-config-data\") pod \"nova-cell1-cell-mapping-7jx4l\" (UID: \"16afe9fc-ff42-487d-9c1a-57f93c0a242c\") " pod="openstack/nova-cell1-cell-mapping-7jx4l" Feb 27 08:12:33 crc kubenswrapper[4612]: I0227 08:12:33.894369 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ddl5r\" (UniqueName: \"kubernetes.io/projected/16afe9fc-ff42-487d-9c1a-57f93c0a242c-kube-api-access-ddl5r\") pod \"nova-cell1-cell-mapping-7jx4l\" (UID: \"16afe9fc-ff42-487d-9c1a-57f93c0a242c\") " pod="openstack/nova-cell1-cell-mapping-7jx4l" Feb 27 08:12:33 crc kubenswrapper[4612]: I0227 08:12:33.906504 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/16afe9fc-ff42-487d-9c1a-57f93c0a242c-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-7jx4l\" (UID: \"16afe9fc-ff42-487d-9c1a-57f93c0a242c\") " pod="openstack/nova-cell1-cell-mapping-7jx4l" Feb 27 08:12:33 crc kubenswrapper[4612]: I0227 08:12:33.956627 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 27 08:12:33 crc kubenswrapper[4612]: I0227 08:12:33.977055 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/16ab2b55-65e7-4d63-b521-0c6421ba52ea-run-httpd\") pod \"ceilometer-0\" (UID: \"16ab2b55-65e7-4d63-b521-0c6421ba52ea\") " pod="openstack/ceilometer-0" Feb 27 08:12:33 crc kubenswrapper[4612]: I0227 08:12:33.977141 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/16ab2b55-65e7-4d63-b521-0c6421ba52ea-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"16ab2b55-65e7-4d63-b521-0c6421ba52ea\") " pod="openstack/ceilometer-0" Feb 27 08:12:33 crc kubenswrapper[4612]: I0227 08:12:33.977163 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/16ab2b55-65e7-4d63-b521-0c6421ba52ea-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"16ab2b55-65e7-4d63-b521-0c6421ba52ea\") " pod="openstack/ceilometer-0" Feb 27 08:12:33 crc kubenswrapper[4612]: I0227 08:12:33.977181 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/16ab2b55-65e7-4d63-b521-0c6421ba52ea-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"16ab2b55-65e7-4d63-b521-0c6421ba52ea\") " pod="openstack/ceilometer-0" Feb 27 08:12:33 crc kubenswrapper[4612]: I0227 08:12:33.977212 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nkq6q\" (UniqueName: \"kubernetes.io/projected/16ab2b55-65e7-4d63-b521-0c6421ba52ea-kube-api-access-nkq6q\") pod \"ceilometer-0\" (UID: \"16ab2b55-65e7-4d63-b521-0c6421ba52ea\") " pod="openstack/ceilometer-0" Feb 27 08:12:33 crc kubenswrapper[4612]: I0227 08:12:33.977261 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/16ab2b55-65e7-4d63-b521-0c6421ba52ea-log-httpd\") pod \"ceilometer-0\" (UID: \"16ab2b55-65e7-4d63-b521-0c6421ba52ea\") " pod="openstack/ceilometer-0" Feb 27 08:12:33 crc kubenswrapper[4612]: I0227 08:12:33.977279 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/16ab2b55-65e7-4d63-b521-0c6421ba52ea-config-data\") pod \"ceilometer-0\" (UID: \"16ab2b55-65e7-4d63-b521-0c6421ba52ea\") " pod="openstack/ceilometer-0" Feb 27 08:12:33 crc kubenswrapper[4612]: I0227 08:12:33.977319 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/16ab2b55-65e7-4d63-b521-0c6421ba52ea-scripts\") pod \"ceilometer-0\" (UID: \"16ab2b55-65e7-4d63-b521-0c6421ba52ea\") " pod="openstack/ceilometer-0" Feb 27 08:12:33 crc kubenswrapper[4612]: I0227 08:12:33.977669 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/16ab2b55-65e7-4d63-b521-0c6421ba52ea-run-httpd\") pod \"ceilometer-0\" (UID: \"16ab2b55-65e7-4d63-b521-0c6421ba52ea\") " pod="openstack/ceilometer-0" Feb 27 08:12:33 crc kubenswrapper[4612]: I0227 08:12:33.977996 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/16ab2b55-65e7-4d63-b521-0c6421ba52ea-log-httpd\") pod \"ceilometer-0\" (UID: \"16ab2b55-65e7-4d63-b521-0c6421ba52ea\") " pod="openstack/ceilometer-0" Feb 27 08:12:33 crc kubenswrapper[4612]: I0227 08:12:33.981190 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/16ab2b55-65e7-4d63-b521-0c6421ba52ea-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"16ab2b55-65e7-4d63-b521-0c6421ba52ea\") " pod="openstack/ceilometer-0" Feb 27 08:12:33 crc kubenswrapper[4612]: I0227 08:12:33.981288 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/16ab2b55-65e7-4d63-b521-0c6421ba52ea-scripts\") pod \"ceilometer-0\" (UID: \"16ab2b55-65e7-4d63-b521-0c6421ba52ea\") " pod="openstack/ceilometer-0" Feb 27 08:12:33 crc kubenswrapper[4612]: I0227 08:12:33.983451 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/16ab2b55-65e7-4d63-b521-0c6421ba52ea-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"16ab2b55-65e7-4d63-b521-0c6421ba52ea\") " pod="openstack/ceilometer-0" Feb 27 08:12:33 crc kubenswrapper[4612]: I0227 08:12:33.984205 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/16ab2b55-65e7-4d63-b521-0c6421ba52ea-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"16ab2b55-65e7-4d63-b521-0c6421ba52ea\") " pod="openstack/ceilometer-0" Feb 27 08:12:33 crc kubenswrapper[4612]: I0227 08:12:33.984830 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/16ab2b55-65e7-4d63-b521-0c6421ba52ea-config-data\") pod \"ceilometer-0\" (UID: \"16ab2b55-65e7-4d63-b521-0c6421ba52ea\") " pod="openstack/ceilometer-0" Feb 27 08:12:33 crc kubenswrapper[4612]: I0227 08:12:33.996456 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nkq6q\" (UniqueName: \"kubernetes.io/projected/16ab2b55-65e7-4d63-b521-0c6421ba52ea-kube-api-access-nkq6q\") pod \"ceilometer-0\" (UID: \"16ab2b55-65e7-4d63-b521-0c6421ba52ea\") " pod="openstack/ceilometer-0" Feb 27 08:12:34 crc kubenswrapper[4612]: I0227 08:12:34.087945 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-7jx4l" Feb 27 08:12:34 crc kubenswrapper[4612]: I0227 08:12:34.226265 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Feb 27 08:12:34 crc kubenswrapper[4612]: I0227 08:12:34.438707 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Feb 27 08:12:34 crc kubenswrapper[4612]: I0227 08:12:34.601006 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-7jx4l"] Feb 27 08:12:34 crc kubenswrapper[4612]: W0227 08:12:34.607845 4612 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod16afe9fc_ff42_487d_9c1a_57f93c0a242c.slice/crio-d2a01edafeb68c3faa1030bad554b38fbadf0a878a5e7e0f71e2282bc804af55 WatchSource:0}: Error finding container d2a01edafeb68c3faa1030bad554b38fbadf0a878a5e7e0f71e2282bc804af55: Status 404 returned error can't find the container with id d2a01edafeb68c3faa1030bad554b38fbadf0a878a5e7e0f71e2282bc804af55 Feb 27 08:12:34 crc kubenswrapper[4612]: I0227 08:12:34.735724 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Feb 27 08:12:34 crc kubenswrapper[4612]: I0227 08:12:34.864873 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d6aaa653-6c5a-4809-957b-6bdefa994b16" path="/var/lib/kubelet/pods/d6aaa653-6c5a-4809-957b-6bdefa994b16/volumes" Feb 27 08:12:34 crc kubenswrapper[4612]: I0227 08:12:34.865768 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f6ca7ac0-4618-4a5b-af1a-7df5fa0e3701" path="/var/lib/kubelet/pods/f6ca7ac0-4618-4a5b-af1a-7df5fa0e3701/volumes" Feb 27 08:12:35 crc kubenswrapper[4612]: I0227 08:12:35.443348 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"16ab2b55-65e7-4d63-b521-0c6421ba52ea","Type":"ContainerStarted","Data":"81f89561ffec280ff86dfb4640a46b229c511b00c78c67b5f789117699dcb027"} Feb 27 08:12:35 crc kubenswrapper[4612]: I0227 08:12:35.443729 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"16ab2b55-65e7-4d63-b521-0c6421ba52ea","Type":"ContainerStarted","Data":"009e1cd84879d93475f93bb4db8b73c23f3624ad025152db0292e94f8fff25a7"} Feb 27 08:12:35 crc kubenswrapper[4612]: I0227 08:12:35.446430 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"7690cb3e-e335-4dc2-bf76-0a3239bc6a5c","Type":"ContainerStarted","Data":"69a882e4cb7dd05b7ecdd3b266255cd5975c6faa28760ebe00d8461124b1c87b"} Feb 27 08:12:35 crc kubenswrapper[4612]: I0227 08:12:35.446463 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"7690cb3e-e335-4dc2-bf76-0a3239bc6a5c","Type":"ContainerStarted","Data":"7c93b2b51dcc74913ce975a46c196bb10a5a0b39463a4799669c1c8957733d22"} Feb 27 08:12:35 crc kubenswrapper[4612]: I0227 08:12:35.446480 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"7690cb3e-e335-4dc2-bf76-0a3239bc6a5c","Type":"ContainerStarted","Data":"7e554d57b078dd5fe7a6f092dd79c1dda530d38fe949305e09c6c3378afe0bf2"} Feb 27 08:12:35 crc kubenswrapper[4612]: I0227 08:12:35.448189 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-7jx4l" event={"ID":"16afe9fc-ff42-487d-9c1a-57f93c0a242c","Type":"ContainerStarted","Data":"32e74328c187d5506986ecb007765a5d4157e22064c750d89488ce3ce250584c"} Feb 27 08:12:35 crc kubenswrapper[4612]: I0227 08:12:35.448323 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-7jx4l" event={"ID":"16afe9fc-ff42-487d-9c1a-57f93c0a242c","Type":"ContainerStarted","Data":"d2a01edafeb68c3faa1030bad554b38fbadf0a878a5e7e0f71e2282bc804af55"} Feb 27 08:12:35 crc kubenswrapper[4612]: I0227 08:12:35.471314 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.471297733 podStartE2EDuration="2.471297733s" podCreationTimestamp="2026-02-27 08:12:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 08:12:35.46456091 +0000 UTC m=+1413.318490908" watchObservedRunningTime="2026-02-27 08:12:35.471297733 +0000 UTC m=+1413.325227731" Feb 27 08:12:35 crc kubenswrapper[4612]: I0227 08:12:35.493994 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-7jx4l" podStartSLOduration=2.493976554 podStartE2EDuration="2.493976554s" podCreationTimestamp="2026-02-27 08:12:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 08:12:35.492887293 +0000 UTC m=+1413.346817291" watchObservedRunningTime="2026-02-27 08:12:35.493976554 +0000 UTC m=+1413.347906552" Feb 27 08:12:36 crc kubenswrapper[4612]: I0227 08:12:36.459519 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"16ab2b55-65e7-4d63-b521-0c6421ba52ea","Type":"ContainerStarted","Data":"1c39bd316dbfd41a026ef515586fb7685c7c4af7037f4691de95e45c6520b733"} Feb 27 08:12:36 crc kubenswrapper[4612]: I0227 08:12:36.891121 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-cd5cbd7b9-dczkk" Feb 27 08:12:36 crc kubenswrapper[4612]: I0227 08:12:36.951540 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-bccf8f775-kgn5g"] Feb 27 08:12:36 crc kubenswrapper[4612]: I0227 08:12:36.951828 4612 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-bccf8f775-kgn5g" podUID="35ff05eb-318f-455c-85a9-7fc07a444bf6" containerName="dnsmasq-dns" containerID="cri-o://49cf9107e3aaa1ff37311e366694fb3372a1bd51a3c8e0756d414d85f9f79070" gracePeriod=10 Feb 27 08:12:37 crc kubenswrapper[4612]: I0227 08:12:37.527994 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"16ab2b55-65e7-4d63-b521-0c6421ba52ea","Type":"ContainerStarted","Data":"8380dee2760b44b7691da0d09f25cad870c4aab1190a477c72c94528b57b5734"} Feb 27 08:12:37 crc kubenswrapper[4612]: I0227 08:12:37.536985 4612 generic.go:334] "Generic (PLEG): container finished" podID="35ff05eb-318f-455c-85a9-7fc07a444bf6" containerID="49cf9107e3aaa1ff37311e366694fb3372a1bd51a3c8e0756d414d85f9f79070" exitCode=0 Feb 27 08:12:37 crc kubenswrapper[4612]: I0227 08:12:37.537029 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bccf8f775-kgn5g" event={"ID":"35ff05eb-318f-455c-85a9-7fc07a444bf6","Type":"ContainerDied","Data":"49cf9107e3aaa1ff37311e366694fb3372a1bd51a3c8e0756d414d85f9f79070"} Feb 27 08:12:37 crc kubenswrapper[4612]: I0227 08:12:37.684268 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bccf8f775-kgn5g" Feb 27 08:12:37 crc kubenswrapper[4612]: I0227 08:12:37.769156 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/35ff05eb-318f-455c-85a9-7fc07a444bf6-config\") pod \"35ff05eb-318f-455c-85a9-7fc07a444bf6\" (UID: \"35ff05eb-318f-455c-85a9-7fc07a444bf6\") " Feb 27 08:12:37 crc kubenswrapper[4612]: I0227 08:12:37.769970 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/35ff05eb-318f-455c-85a9-7fc07a444bf6-ovsdbserver-nb\") pod \"35ff05eb-318f-455c-85a9-7fc07a444bf6\" (UID: \"35ff05eb-318f-455c-85a9-7fc07a444bf6\") " Feb 27 08:12:37 crc kubenswrapper[4612]: I0227 08:12:37.770055 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lj2n5\" (UniqueName: \"kubernetes.io/projected/35ff05eb-318f-455c-85a9-7fc07a444bf6-kube-api-access-lj2n5\") pod \"35ff05eb-318f-455c-85a9-7fc07a444bf6\" (UID: \"35ff05eb-318f-455c-85a9-7fc07a444bf6\") " Feb 27 08:12:37 crc kubenswrapper[4612]: I0227 08:12:37.770821 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/35ff05eb-318f-455c-85a9-7fc07a444bf6-dns-svc\") pod \"35ff05eb-318f-455c-85a9-7fc07a444bf6\" (UID: \"35ff05eb-318f-455c-85a9-7fc07a444bf6\") " Feb 27 08:12:37 crc kubenswrapper[4612]: I0227 08:12:37.770917 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/35ff05eb-318f-455c-85a9-7fc07a444bf6-ovsdbserver-sb\") pod \"35ff05eb-318f-455c-85a9-7fc07a444bf6\" (UID: \"35ff05eb-318f-455c-85a9-7fc07a444bf6\") " Feb 27 08:12:37 crc kubenswrapper[4612]: I0227 08:12:37.770952 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/35ff05eb-318f-455c-85a9-7fc07a444bf6-dns-swift-storage-0\") pod \"35ff05eb-318f-455c-85a9-7fc07a444bf6\" (UID: \"35ff05eb-318f-455c-85a9-7fc07a444bf6\") " Feb 27 08:12:37 crc kubenswrapper[4612]: I0227 08:12:37.784268 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/35ff05eb-318f-455c-85a9-7fc07a444bf6-kube-api-access-lj2n5" (OuterVolumeSpecName: "kube-api-access-lj2n5") pod "35ff05eb-318f-455c-85a9-7fc07a444bf6" (UID: "35ff05eb-318f-455c-85a9-7fc07a444bf6"). InnerVolumeSpecName "kube-api-access-lj2n5". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 08:12:37 crc kubenswrapper[4612]: I0227 08:12:37.873115 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/35ff05eb-318f-455c-85a9-7fc07a444bf6-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "35ff05eb-318f-455c-85a9-7fc07a444bf6" (UID: "35ff05eb-318f-455c-85a9-7fc07a444bf6"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 08:12:37 crc kubenswrapper[4612]: I0227 08:12:37.873399 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/35ff05eb-318f-455c-85a9-7fc07a444bf6-ovsdbserver-nb\") pod \"35ff05eb-318f-455c-85a9-7fc07a444bf6\" (UID: \"35ff05eb-318f-455c-85a9-7fc07a444bf6\") " Feb 27 08:12:37 crc kubenswrapper[4612]: I0227 08:12:37.874052 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lj2n5\" (UniqueName: \"kubernetes.io/projected/35ff05eb-318f-455c-85a9-7fc07a444bf6-kube-api-access-lj2n5\") on node \"crc\" DevicePath \"\"" Feb 27 08:12:37 crc kubenswrapper[4612]: W0227 08:12:37.874258 4612 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/35ff05eb-318f-455c-85a9-7fc07a444bf6/volumes/kubernetes.io~configmap/ovsdbserver-nb Feb 27 08:12:37 crc kubenswrapper[4612]: I0227 08:12:37.874281 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/35ff05eb-318f-455c-85a9-7fc07a444bf6-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "35ff05eb-318f-455c-85a9-7fc07a444bf6" (UID: "35ff05eb-318f-455c-85a9-7fc07a444bf6"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 08:12:37 crc kubenswrapper[4612]: I0227 08:12:37.874825 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/35ff05eb-318f-455c-85a9-7fc07a444bf6-config" (OuterVolumeSpecName: "config") pod "35ff05eb-318f-455c-85a9-7fc07a444bf6" (UID: "35ff05eb-318f-455c-85a9-7fc07a444bf6"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 08:12:37 crc kubenswrapper[4612]: I0227 08:12:37.905716 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/35ff05eb-318f-455c-85a9-7fc07a444bf6-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "35ff05eb-318f-455c-85a9-7fc07a444bf6" (UID: "35ff05eb-318f-455c-85a9-7fc07a444bf6"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 08:12:37 crc kubenswrapper[4612]: I0227 08:12:37.922523 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/35ff05eb-318f-455c-85a9-7fc07a444bf6-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "35ff05eb-318f-455c-85a9-7fc07a444bf6" (UID: "35ff05eb-318f-455c-85a9-7fc07a444bf6"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 08:12:37 crc kubenswrapper[4612]: I0227 08:12:37.943851 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/35ff05eb-318f-455c-85a9-7fc07a444bf6-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "35ff05eb-318f-455c-85a9-7fc07a444bf6" (UID: "35ff05eb-318f-455c-85a9-7fc07a444bf6"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 08:12:37 crc kubenswrapper[4612]: I0227 08:12:37.975572 4612 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/35ff05eb-318f-455c-85a9-7fc07a444bf6-config\") on node \"crc\" DevicePath \"\"" Feb 27 08:12:37 crc kubenswrapper[4612]: I0227 08:12:37.975605 4612 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/35ff05eb-318f-455c-85a9-7fc07a444bf6-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 27 08:12:37 crc kubenswrapper[4612]: I0227 08:12:37.975617 4612 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/35ff05eb-318f-455c-85a9-7fc07a444bf6-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 27 08:12:37 crc kubenswrapper[4612]: I0227 08:12:37.975624 4612 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/35ff05eb-318f-455c-85a9-7fc07a444bf6-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 27 08:12:37 crc kubenswrapper[4612]: I0227 08:12:37.975633 4612 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/35ff05eb-318f-455c-85a9-7fc07a444bf6-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Feb 27 08:12:38 crc kubenswrapper[4612]: I0227 08:12:38.549137 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bccf8f775-kgn5g" event={"ID":"35ff05eb-318f-455c-85a9-7fc07a444bf6","Type":"ContainerDied","Data":"a8c6e9a5808b6c72f39bf4d07cf08d027cee7be07395b714f4b359bffbb5486b"} Feb 27 08:12:38 crc kubenswrapper[4612]: I0227 08:12:38.549456 4612 scope.go:117] "RemoveContainer" containerID="49cf9107e3aaa1ff37311e366694fb3372a1bd51a3c8e0756d414d85f9f79070" Feb 27 08:12:38 crc kubenswrapper[4612]: I0227 08:12:38.549248 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bccf8f775-kgn5g" Feb 27 08:12:38 crc kubenswrapper[4612]: I0227 08:12:38.572139 4612 scope.go:117] "RemoveContainer" containerID="91cae3a3762855836c712786ebf97534e2857c9913837c47a6c8f9edc9488675" Feb 27 08:12:38 crc kubenswrapper[4612]: I0227 08:12:38.634877 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-bccf8f775-kgn5g"] Feb 27 08:12:38 crc kubenswrapper[4612]: I0227 08:12:38.646928 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-bccf8f775-kgn5g"] Feb 27 08:12:38 crc kubenswrapper[4612]: I0227 08:12:38.867630 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="35ff05eb-318f-455c-85a9-7fc07a444bf6" path="/var/lib/kubelet/pods/35ff05eb-318f-455c-85a9-7fc07a444bf6/volumes" Feb 27 08:12:39 crc kubenswrapper[4612]: E0227 08:12:39.163627 4612 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3c4d5bb6_8ca7_41d7_8e30_af438a0e267a.slice/crio-b20dc16d14af8f07dddc21f5ce3968bf2ad89f62cb1e10b67ad0a10affdda173.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3c4d5bb6_8ca7_41d7_8e30_af438a0e267a.slice/crio-conmon-b20dc16d14af8f07dddc21f5ce3968bf2ad89f62cb1e10b67ad0a10affdda173.scope\": RecentStats: unable to find data in memory cache]" Feb 27 08:12:39 crc kubenswrapper[4612]: I0227 08:12:39.575599 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"16ab2b55-65e7-4d63-b521-0c6421ba52ea","Type":"ContainerStarted","Data":"d1a9a7a5cd689192f49fbc25a8ecb33f3568b7ba4e0c08b052e10259f18bcc96"} Feb 27 08:12:39 crc kubenswrapper[4612]: I0227 08:12:39.576799 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Feb 27 08:12:39 crc kubenswrapper[4612]: I0227 08:12:39.607535 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.995094486 podStartE2EDuration="6.607512247s" podCreationTimestamp="2026-02-27 08:12:33 +0000 UTC" firstStartedPulling="2026-02-27 08:12:34.752325806 +0000 UTC m=+1412.606255804" lastFinishedPulling="2026-02-27 08:12:38.364743567 +0000 UTC m=+1416.218673565" observedRunningTime="2026-02-27 08:12:39.597282594 +0000 UTC m=+1417.451212582" watchObservedRunningTime="2026-02-27 08:12:39.607512247 +0000 UTC m=+1417.461442245" Feb 27 08:12:41 crc kubenswrapper[4612]: I0227 08:12:41.596423 4612 generic.go:334] "Generic (PLEG): container finished" podID="16afe9fc-ff42-487d-9c1a-57f93c0a242c" containerID="32e74328c187d5506986ecb007765a5d4157e22064c750d89488ce3ce250584c" exitCode=0 Feb 27 08:12:41 crc kubenswrapper[4612]: I0227 08:12:41.596544 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-7jx4l" event={"ID":"16afe9fc-ff42-487d-9c1a-57f93c0a242c","Type":"ContainerDied","Data":"32e74328c187d5506986ecb007765a5d4157e22064c750d89488ce3ce250584c"} Feb 27 08:12:42 crc kubenswrapper[4612]: I0227 08:12:42.998235 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-7jx4l" Feb 27 08:12:43 crc kubenswrapper[4612]: I0227 08:12:43.073259 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/16afe9fc-ff42-487d-9c1a-57f93c0a242c-scripts\") pod \"16afe9fc-ff42-487d-9c1a-57f93c0a242c\" (UID: \"16afe9fc-ff42-487d-9c1a-57f93c0a242c\") " Feb 27 08:12:43 crc kubenswrapper[4612]: I0227 08:12:43.073321 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ddl5r\" (UniqueName: \"kubernetes.io/projected/16afe9fc-ff42-487d-9c1a-57f93c0a242c-kube-api-access-ddl5r\") pod \"16afe9fc-ff42-487d-9c1a-57f93c0a242c\" (UID: \"16afe9fc-ff42-487d-9c1a-57f93c0a242c\") " Feb 27 08:12:43 crc kubenswrapper[4612]: I0227 08:12:43.073378 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/16afe9fc-ff42-487d-9c1a-57f93c0a242c-config-data\") pod \"16afe9fc-ff42-487d-9c1a-57f93c0a242c\" (UID: \"16afe9fc-ff42-487d-9c1a-57f93c0a242c\") " Feb 27 08:12:43 crc kubenswrapper[4612]: I0227 08:12:43.073561 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/16afe9fc-ff42-487d-9c1a-57f93c0a242c-combined-ca-bundle\") pod \"16afe9fc-ff42-487d-9c1a-57f93c0a242c\" (UID: \"16afe9fc-ff42-487d-9c1a-57f93c0a242c\") " Feb 27 08:12:43 crc kubenswrapper[4612]: I0227 08:12:43.078881 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/16afe9fc-ff42-487d-9c1a-57f93c0a242c-kube-api-access-ddl5r" (OuterVolumeSpecName: "kube-api-access-ddl5r") pod "16afe9fc-ff42-487d-9c1a-57f93c0a242c" (UID: "16afe9fc-ff42-487d-9c1a-57f93c0a242c"). InnerVolumeSpecName "kube-api-access-ddl5r". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 08:12:43 crc kubenswrapper[4612]: I0227 08:12:43.086932 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/16afe9fc-ff42-487d-9c1a-57f93c0a242c-scripts" (OuterVolumeSpecName: "scripts") pod "16afe9fc-ff42-487d-9c1a-57f93c0a242c" (UID: "16afe9fc-ff42-487d-9c1a-57f93c0a242c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:12:43 crc kubenswrapper[4612]: I0227 08:12:43.104829 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/16afe9fc-ff42-487d-9c1a-57f93c0a242c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "16afe9fc-ff42-487d-9c1a-57f93c0a242c" (UID: "16afe9fc-ff42-487d-9c1a-57f93c0a242c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:12:43 crc kubenswrapper[4612]: I0227 08:12:43.135623 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/16afe9fc-ff42-487d-9c1a-57f93c0a242c-config-data" (OuterVolumeSpecName: "config-data") pod "16afe9fc-ff42-487d-9c1a-57f93c0a242c" (UID: "16afe9fc-ff42-487d-9c1a-57f93c0a242c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:12:43 crc kubenswrapper[4612]: I0227 08:12:43.177106 4612 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/16afe9fc-ff42-487d-9c1a-57f93c0a242c-scripts\") on node \"crc\" DevicePath \"\"" Feb 27 08:12:43 crc kubenswrapper[4612]: I0227 08:12:43.177142 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ddl5r\" (UniqueName: \"kubernetes.io/projected/16afe9fc-ff42-487d-9c1a-57f93c0a242c-kube-api-access-ddl5r\") on node \"crc\" DevicePath \"\"" Feb 27 08:12:43 crc kubenswrapper[4612]: I0227 08:12:43.177160 4612 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/16afe9fc-ff42-487d-9c1a-57f93c0a242c-config-data\") on node \"crc\" DevicePath \"\"" Feb 27 08:12:43 crc kubenswrapper[4612]: I0227 08:12:43.177174 4612 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/16afe9fc-ff42-487d-9c1a-57f93c0a242c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 08:12:43 crc kubenswrapper[4612]: I0227 08:12:43.615097 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-7jx4l" event={"ID":"16afe9fc-ff42-487d-9c1a-57f93c0a242c","Type":"ContainerDied","Data":"d2a01edafeb68c3faa1030bad554b38fbadf0a878a5e7e0f71e2282bc804af55"} Feb 27 08:12:43 crc kubenswrapper[4612]: I0227 08:12:43.615408 4612 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d2a01edafeb68c3faa1030bad554b38fbadf0a878a5e7e0f71e2282bc804af55" Feb 27 08:12:43 crc kubenswrapper[4612]: I0227 08:12:43.615325 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-7jx4l" Feb 27 08:12:43 crc kubenswrapper[4612]: I0227 08:12:43.916413 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Feb 27 08:12:43 crc kubenswrapper[4612]: I0227 08:12:43.916611 4612 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="ff8b7423-9c8a-4c8a-8d7f-11e70ded2244" containerName="nova-scheduler-scheduler" containerID="cri-o://205e0e7c8976d96647c6a5c7b44e2fbf349cb1346ac71036d6ee20ad722ab20b" gracePeriod=30 Feb 27 08:12:43 crc kubenswrapper[4612]: I0227 08:12:43.960181 4612 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Feb 27 08:12:43 crc kubenswrapper[4612]: I0227 08:12:43.961166 4612 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Feb 27 08:12:43 crc kubenswrapper[4612]: I0227 08:12:43.965543 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Feb 27 08:12:44 crc kubenswrapper[4612]: I0227 08:12:44.010294 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Feb 27 08:12:44 crc kubenswrapper[4612]: I0227 08:12:44.010556 4612 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="59d65625-1a7d-4f75-95a3-82ce5976b27a" containerName="nova-metadata-log" containerID="cri-o://2f2643fefc567894c7e71a70aa2c5809517c1801c609eac1707c40bcf9c76d76" gracePeriod=30 Feb 27 08:12:44 crc kubenswrapper[4612]: I0227 08:12:44.010707 4612 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="59d65625-1a7d-4f75-95a3-82ce5976b27a" containerName="nova-metadata-metadata" containerID="cri-o://e04e3c34eac9644f55d0e1c95060a48f1ec164a86ffdcf54bf34f17667296e37" gracePeriod=30 Feb 27 08:12:44 crc kubenswrapper[4612]: E0227 08:12:44.372565 4612 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="205e0e7c8976d96647c6a5c7b44e2fbf349cb1346ac71036d6ee20ad722ab20b" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Feb 27 08:12:44 crc kubenswrapper[4612]: E0227 08:12:44.374468 4612 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="205e0e7c8976d96647c6a5c7b44e2fbf349cb1346ac71036d6ee20ad722ab20b" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Feb 27 08:12:44 crc kubenswrapper[4612]: E0227 08:12:44.382716 4612 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="205e0e7c8976d96647c6a5c7b44e2fbf349cb1346ac71036d6ee20ad722ab20b" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Feb 27 08:12:44 crc kubenswrapper[4612]: E0227 08:12:44.382801 4612 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="ff8b7423-9c8a-4c8a-8d7f-11e70ded2244" containerName="nova-scheduler-scheduler" Feb 27 08:12:44 crc kubenswrapper[4612]: I0227 08:12:44.625740 4612 generic.go:334] "Generic (PLEG): container finished" podID="59d65625-1a7d-4f75-95a3-82ce5976b27a" containerID="2f2643fefc567894c7e71a70aa2c5809517c1801c609eac1707c40bcf9c76d76" exitCode=143 Feb 27 08:12:44 crc kubenswrapper[4612]: I0227 08:12:44.625845 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"59d65625-1a7d-4f75-95a3-82ce5976b27a","Type":"ContainerDied","Data":"2f2643fefc567894c7e71a70aa2c5809517c1801c609eac1707c40bcf9c76d76"} Feb 27 08:12:45 crc kubenswrapper[4612]: I0227 08:12:45.004810 4612 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="7690cb3e-e335-4dc2-bf76-0a3239bc6a5c" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.208:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Feb 27 08:12:45 crc kubenswrapper[4612]: I0227 08:12:45.005261 4612 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="7690cb3e-e335-4dc2-bf76-0a3239bc6a5c" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.208:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Feb 27 08:12:45 crc kubenswrapper[4612]: I0227 08:12:45.635376 4612 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="7690cb3e-e335-4dc2-bf76-0a3239bc6a5c" containerName="nova-api-log" containerID="cri-o://7c93b2b51dcc74913ce975a46c196bb10a5a0b39463a4799669c1c8957733d22" gracePeriod=30 Feb 27 08:12:45 crc kubenswrapper[4612]: I0227 08:12:45.635406 4612 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="7690cb3e-e335-4dc2-bf76-0a3239bc6a5c" containerName="nova-api-api" containerID="cri-o://69a882e4cb7dd05b7ecdd3b266255cd5975c6faa28760ebe00d8461124b1c87b" gracePeriod=30 Feb 27 08:12:46 crc kubenswrapper[4612]: I0227 08:12:46.645352 4612 generic.go:334] "Generic (PLEG): container finished" podID="7690cb3e-e335-4dc2-bf76-0a3239bc6a5c" containerID="7c93b2b51dcc74913ce975a46c196bb10a5a0b39463a4799669c1c8957733d22" exitCode=143 Feb 27 08:12:46 crc kubenswrapper[4612]: I0227 08:12:46.645419 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"7690cb3e-e335-4dc2-bf76-0a3239bc6a5c","Type":"ContainerDied","Data":"7c93b2b51dcc74913ce975a46c196bb10a5a0b39463a4799669c1c8957733d22"} Feb 27 08:12:47 crc kubenswrapper[4612]: I0227 08:12:47.418571 4612 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="59d65625-1a7d-4f75-95a3-82ce5976b27a" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.203:8775/\": read tcp 10.217.0.2:48026->10.217.0.203:8775: read: connection reset by peer" Feb 27 08:12:47 crc kubenswrapper[4612]: I0227 08:12:47.418612 4612 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="59d65625-1a7d-4f75-95a3-82ce5976b27a" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.203:8775/\": read tcp 10.217.0.2:48028->10.217.0.203:8775: read: connection reset by peer" Feb 27 08:12:47 crc kubenswrapper[4612]: I0227 08:12:47.666184 4612 generic.go:334] "Generic (PLEG): container finished" podID="59d65625-1a7d-4f75-95a3-82ce5976b27a" containerID="e04e3c34eac9644f55d0e1c95060a48f1ec164a86ffdcf54bf34f17667296e37" exitCode=0 Feb 27 08:12:47 crc kubenswrapper[4612]: I0227 08:12:47.666447 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"59d65625-1a7d-4f75-95a3-82ce5976b27a","Type":"ContainerDied","Data":"e04e3c34eac9644f55d0e1c95060a48f1ec164a86ffdcf54bf34f17667296e37"} Feb 27 08:12:47 crc kubenswrapper[4612]: I0227 08:12:47.970597 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 27 08:12:48 crc kubenswrapper[4612]: I0227 08:12:48.083491 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/59d65625-1a7d-4f75-95a3-82ce5976b27a-config-data\") pod \"59d65625-1a7d-4f75-95a3-82ce5976b27a\" (UID: \"59d65625-1a7d-4f75-95a3-82ce5976b27a\") " Feb 27 08:12:48 crc kubenswrapper[4612]: I0227 08:12:48.083539 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9qnp2\" (UniqueName: \"kubernetes.io/projected/59d65625-1a7d-4f75-95a3-82ce5976b27a-kube-api-access-9qnp2\") pod \"59d65625-1a7d-4f75-95a3-82ce5976b27a\" (UID: \"59d65625-1a7d-4f75-95a3-82ce5976b27a\") " Feb 27 08:12:48 crc kubenswrapper[4612]: I0227 08:12:48.083660 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/59d65625-1a7d-4f75-95a3-82ce5976b27a-logs\") pod \"59d65625-1a7d-4f75-95a3-82ce5976b27a\" (UID: \"59d65625-1a7d-4f75-95a3-82ce5976b27a\") " Feb 27 08:12:48 crc kubenswrapper[4612]: I0227 08:12:48.083729 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/59d65625-1a7d-4f75-95a3-82ce5976b27a-nova-metadata-tls-certs\") pod \"59d65625-1a7d-4f75-95a3-82ce5976b27a\" (UID: \"59d65625-1a7d-4f75-95a3-82ce5976b27a\") " Feb 27 08:12:48 crc kubenswrapper[4612]: I0227 08:12:48.083751 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/59d65625-1a7d-4f75-95a3-82ce5976b27a-combined-ca-bundle\") pod \"59d65625-1a7d-4f75-95a3-82ce5976b27a\" (UID: \"59d65625-1a7d-4f75-95a3-82ce5976b27a\") " Feb 27 08:12:48 crc kubenswrapper[4612]: I0227 08:12:48.099004 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/59d65625-1a7d-4f75-95a3-82ce5976b27a-logs" (OuterVolumeSpecName: "logs") pod "59d65625-1a7d-4f75-95a3-82ce5976b27a" (UID: "59d65625-1a7d-4f75-95a3-82ce5976b27a"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 08:12:48 crc kubenswrapper[4612]: I0227 08:12:48.146187 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/59d65625-1a7d-4f75-95a3-82ce5976b27a-kube-api-access-9qnp2" (OuterVolumeSpecName: "kube-api-access-9qnp2") pod "59d65625-1a7d-4f75-95a3-82ce5976b27a" (UID: "59d65625-1a7d-4f75-95a3-82ce5976b27a"). InnerVolumeSpecName "kube-api-access-9qnp2". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 08:12:48 crc kubenswrapper[4612]: I0227 08:12:48.187195 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9qnp2\" (UniqueName: \"kubernetes.io/projected/59d65625-1a7d-4f75-95a3-82ce5976b27a-kube-api-access-9qnp2\") on node \"crc\" DevicePath \"\"" Feb 27 08:12:48 crc kubenswrapper[4612]: I0227 08:12:48.187221 4612 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/59d65625-1a7d-4f75-95a3-82ce5976b27a-logs\") on node \"crc\" DevicePath \"\"" Feb 27 08:12:48 crc kubenswrapper[4612]: I0227 08:12:48.225852 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/59d65625-1a7d-4f75-95a3-82ce5976b27a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "59d65625-1a7d-4f75-95a3-82ce5976b27a" (UID: "59d65625-1a7d-4f75-95a3-82ce5976b27a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:12:48 crc kubenswrapper[4612]: I0227 08:12:48.230765 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/59d65625-1a7d-4f75-95a3-82ce5976b27a-config-data" (OuterVolumeSpecName: "config-data") pod "59d65625-1a7d-4f75-95a3-82ce5976b27a" (UID: "59d65625-1a7d-4f75-95a3-82ce5976b27a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:12:48 crc kubenswrapper[4612]: I0227 08:12:48.290399 4612 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/59d65625-1a7d-4f75-95a3-82ce5976b27a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 08:12:48 crc kubenswrapper[4612]: I0227 08:12:48.290744 4612 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/59d65625-1a7d-4f75-95a3-82ce5976b27a-config-data\") on node \"crc\" DevicePath \"\"" Feb 27 08:12:48 crc kubenswrapper[4612]: I0227 08:12:48.294847 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/59d65625-1a7d-4f75-95a3-82ce5976b27a-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "59d65625-1a7d-4f75-95a3-82ce5976b27a" (UID: "59d65625-1a7d-4f75-95a3-82ce5976b27a"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:12:48 crc kubenswrapper[4612]: I0227 08:12:48.392238 4612 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/59d65625-1a7d-4f75-95a3-82ce5976b27a-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 27 08:12:48 crc kubenswrapper[4612]: I0227 08:12:48.679088 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"59d65625-1a7d-4f75-95a3-82ce5976b27a","Type":"ContainerDied","Data":"1ce7ea348234bd6cfcc81b61b16af5d3eabed4cab4d7609771b3f81ecc8e2606"} Feb 27 08:12:48 crc kubenswrapper[4612]: I0227 08:12:48.679139 4612 scope.go:117] "RemoveContainer" containerID="e04e3c34eac9644f55d0e1c95060a48f1ec164a86ffdcf54bf34f17667296e37" Feb 27 08:12:48 crc kubenswrapper[4612]: I0227 08:12:48.679264 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 27 08:12:48 crc kubenswrapper[4612]: I0227 08:12:48.694926 4612 generic.go:334] "Generic (PLEG): container finished" podID="ff8b7423-9c8a-4c8a-8d7f-11e70ded2244" containerID="205e0e7c8976d96647c6a5c7b44e2fbf349cb1346ac71036d6ee20ad722ab20b" exitCode=0 Feb 27 08:12:48 crc kubenswrapper[4612]: I0227 08:12:48.694975 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"ff8b7423-9c8a-4c8a-8d7f-11e70ded2244","Type":"ContainerDied","Data":"205e0e7c8976d96647c6a5c7b44e2fbf349cb1346ac71036d6ee20ad722ab20b"} Feb 27 08:12:48 crc kubenswrapper[4612]: I0227 08:12:48.726463 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Feb 27 08:12:48 crc kubenswrapper[4612]: I0227 08:12:48.740799 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Feb 27 08:12:48 crc kubenswrapper[4612]: I0227 08:12:48.755087 4612 scope.go:117] "RemoveContainer" containerID="2f2643fefc567894c7e71a70aa2c5809517c1801c609eac1707c40bcf9c76d76" Feb 27 08:12:48 crc kubenswrapper[4612]: I0227 08:12:48.760845 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Feb 27 08:12:48 crc kubenswrapper[4612]: E0227 08:12:48.761326 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="35ff05eb-318f-455c-85a9-7fc07a444bf6" containerName="dnsmasq-dns" Feb 27 08:12:48 crc kubenswrapper[4612]: I0227 08:12:48.761349 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="35ff05eb-318f-455c-85a9-7fc07a444bf6" containerName="dnsmasq-dns" Feb 27 08:12:48 crc kubenswrapper[4612]: E0227 08:12:48.761366 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="59d65625-1a7d-4f75-95a3-82ce5976b27a" containerName="nova-metadata-log" Feb 27 08:12:48 crc kubenswrapper[4612]: I0227 08:12:48.761374 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="59d65625-1a7d-4f75-95a3-82ce5976b27a" containerName="nova-metadata-log" Feb 27 08:12:48 crc kubenswrapper[4612]: E0227 08:12:48.761393 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="59d65625-1a7d-4f75-95a3-82ce5976b27a" containerName="nova-metadata-metadata" Feb 27 08:12:48 crc kubenswrapper[4612]: I0227 08:12:48.761402 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="59d65625-1a7d-4f75-95a3-82ce5976b27a" containerName="nova-metadata-metadata" Feb 27 08:12:48 crc kubenswrapper[4612]: E0227 08:12:48.761421 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="35ff05eb-318f-455c-85a9-7fc07a444bf6" containerName="init" Feb 27 08:12:48 crc kubenswrapper[4612]: I0227 08:12:48.761428 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="35ff05eb-318f-455c-85a9-7fc07a444bf6" containerName="init" Feb 27 08:12:48 crc kubenswrapper[4612]: E0227 08:12:48.761445 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="16afe9fc-ff42-487d-9c1a-57f93c0a242c" containerName="nova-manage" Feb 27 08:12:48 crc kubenswrapper[4612]: I0227 08:12:48.761452 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="16afe9fc-ff42-487d-9c1a-57f93c0a242c" containerName="nova-manage" Feb 27 08:12:48 crc kubenswrapper[4612]: I0227 08:12:48.761652 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="35ff05eb-318f-455c-85a9-7fc07a444bf6" containerName="dnsmasq-dns" Feb 27 08:12:48 crc kubenswrapper[4612]: I0227 08:12:48.761668 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="59d65625-1a7d-4f75-95a3-82ce5976b27a" containerName="nova-metadata-log" Feb 27 08:12:48 crc kubenswrapper[4612]: I0227 08:12:48.762926 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="16afe9fc-ff42-487d-9c1a-57f93c0a242c" containerName="nova-manage" Feb 27 08:12:48 crc kubenswrapper[4612]: I0227 08:12:48.762955 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="59d65625-1a7d-4f75-95a3-82ce5976b27a" containerName="nova-metadata-metadata" Feb 27 08:12:48 crc kubenswrapper[4612]: I0227 08:12:48.764109 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 27 08:12:48 crc kubenswrapper[4612]: I0227 08:12:48.772624 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Feb 27 08:12:48 crc kubenswrapper[4612]: I0227 08:12:48.773275 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Feb 27 08:12:48 crc kubenswrapper[4612]: I0227 08:12:48.780813 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Feb 27 08:12:48 crc kubenswrapper[4612]: I0227 08:12:48.876480 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="59d65625-1a7d-4f75-95a3-82ce5976b27a" path="/var/lib/kubelet/pods/59d65625-1a7d-4f75-95a3-82ce5976b27a/volumes" Feb 27 08:12:48 crc kubenswrapper[4612]: I0227 08:12:48.902173 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/24c836ae-ba59-4432-acee-fd17b60c8236-config-data\") pod \"nova-metadata-0\" (UID: \"24c836ae-ba59-4432-acee-fd17b60c8236\") " pod="openstack/nova-metadata-0" Feb 27 08:12:48 crc kubenswrapper[4612]: I0227 08:12:48.902209 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/24c836ae-ba59-4432-acee-fd17b60c8236-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"24c836ae-ba59-4432-acee-fd17b60c8236\") " pod="openstack/nova-metadata-0" Feb 27 08:12:48 crc kubenswrapper[4612]: I0227 08:12:48.902270 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/24c836ae-ba59-4432-acee-fd17b60c8236-logs\") pod \"nova-metadata-0\" (UID: \"24c836ae-ba59-4432-acee-fd17b60c8236\") " pod="openstack/nova-metadata-0" Feb 27 08:12:48 crc kubenswrapper[4612]: I0227 08:12:48.902342 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24c836ae-ba59-4432-acee-fd17b60c8236-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"24c836ae-ba59-4432-acee-fd17b60c8236\") " pod="openstack/nova-metadata-0" Feb 27 08:12:48 crc kubenswrapper[4612]: I0227 08:12:48.902388 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xxt4d\" (UniqueName: \"kubernetes.io/projected/24c836ae-ba59-4432-acee-fd17b60c8236-kube-api-access-xxt4d\") pod \"nova-metadata-0\" (UID: \"24c836ae-ba59-4432-acee-fd17b60c8236\") " pod="openstack/nova-metadata-0" Feb 27 08:12:49 crc kubenswrapper[4612]: I0227 08:12:49.003625 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/24c836ae-ba59-4432-acee-fd17b60c8236-logs\") pod \"nova-metadata-0\" (UID: \"24c836ae-ba59-4432-acee-fd17b60c8236\") " pod="openstack/nova-metadata-0" Feb 27 08:12:49 crc kubenswrapper[4612]: I0227 08:12:49.003748 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24c836ae-ba59-4432-acee-fd17b60c8236-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"24c836ae-ba59-4432-acee-fd17b60c8236\") " pod="openstack/nova-metadata-0" Feb 27 08:12:49 crc kubenswrapper[4612]: I0227 08:12:49.003784 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xxt4d\" (UniqueName: \"kubernetes.io/projected/24c836ae-ba59-4432-acee-fd17b60c8236-kube-api-access-xxt4d\") pod \"nova-metadata-0\" (UID: \"24c836ae-ba59-4432-acee-fd17b60c8236\") " pod="openstack/nova-metadata-0" Feb 27 08:12:49 crc kubenswrapper[4612]: I0227 08:12:49.003868 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/24c836ae-ba59-4432-acee-fd17b60c8236-config-data\") pod \"nova-metadata-0\" (UID: \"24c836ae-ba59-4432-acee-fd17b60c8236\") " pod="openstack/nova-metadata-0" Feb 27 08:12:49 crc kubenswrapper[4612]: I0227 08:12:49.003884 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/24c836ae-ba59-4432-acee-fd17b60c8236-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"24c836ae-ba59-4432-acee-fd17b60c8236\") " pod="openstack/nova-metadata-0" Feb 27 08:12:49 crc kubenswrapper[4612]: I0227 08:12:49.004361 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/24c836ae-ba59-4432-acee-fd17b60c8236-logs\") pod \"nova-metadata-0\" (UID: \"24c836ae-ba59-4432-acee-fd17b60c8236\") " pod="openstack/nova-metadata-0" Feb 27 08:12:49 crc kubenswrapper[4612]: I0227 08:12:49.010551 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24c836ae-ba59-4432-acee-fd17b60c8236-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"24c836ae-ba59-4432-acee-fd17b60c8236\") " pod="openstack/nova-metadata-0" Feb 27 08:12:49 crc kubenswrapper[4612]: I0227 08:12:49.011338 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/24c836ae-ba59-4432-acee-fd17b60c8236-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"24c836ae-ba59-4432-acee-fd17b60c8236\") " pod="openstack/nova-metadata-0" Feb 27 08:12:49 crc kubenswrapper[4612]: I0227 08:12:49.013248 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/24c836ae-ba59-4432-acee-fd17b60c8236-config-data\") pod \"nova-metadata-0\" (UID: \"24c836ae-ba59-4432-acee-fd17b60c8236\") " pod="openstack/nova-metadata-0" Feb 27 08:12:49 crc kubenswrapper[4612]: I0227 08:12:49.030374 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xxt4d\" (UniqueName: \"kubernetes.io/projected/24c836ae-ba59-4432-acee-fd17b60c8236-kube-api-access-xxt4d\") pod \"nova-metadata-0\" (UID: \"24c836ae-ba59-4432-acee-fd17b60c8236\") " pod="openstack/nova-metadata-0" Feb 27 08:12:49 crc kubenswrapper[4612]: I0227 08:12:49.100745 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 27 08:12:49 crc kubenswrapper[4612]: I0227 08:12:49.108203 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Feb 27 08:12:49 crc kubenswrapper[4612]: I0227 08:12:49.205629 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff8b7423-9c8a-4c8a-8d7f-11e70ded2244-combined-ca-bundle\") pod \"ff8b7423-9c8a-4c8a-8d7f-11e70ded2244\" (UID: \"ff8b7423-9c8a-4c8a-8d7f-11e70ded2244\") " Feb 27 08:12:49 crc kubenswrapper[4612]: I0227 08:12:49.205757 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rtkd2\" (UniqueName: \"kubernetes.io/projected/ff8b7423-9c8a-4c8a-8d7f-11e70ded2244-kube-api-access-rtkd2\") pod \"ff8b7423-9c8a-4c8a-8d7f-11e70ded2244\" (UID: \"ff8b7423-9c8a-4c8a-8d7f-11e70ded2244\") " Feb 27 08:12:49 crc kubenswrapper[4612]: I0227 08:12:49.205878 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ff8b7423-9c8a-4c8a-8d7f-11e70ded2244-config-data\") pod \"ff8b7423-9c8a-4c8a-8d7f-11e70ded2244\" (UID: \"ff8b7423-9c8a-4c8a-8d7f-11e70ded2244\") " Feb 27 08:12:49 crc kubenswrapper[4612]: I0227 08:12:49.216988 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ff8b7423-9c8a-4c8a-8d7f-11e70ded2244-kube-api-access-rtkd2" (OuterVolumeSpecName: "kube-api-access-rtkd2") pod "ff8b7423-9c8a-4c8a-8d7f-11e70ded2244" (UID: "ff8b7423-9c8a-4c8a-8d7f-11e70ded2244"). InnerVolumeSpecName "kube-api-access-rtkd2". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 08:12:49 crc kubenswrapper[4612]: I0227 08:12:49.269044 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ff8b7423-9c8a-4c8a-8d7f-11e70ded2244-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ff8b7423-9c8a-4c8a-8d7f-11e70ded2244" (UID: "ff8b7423-9c8a-4c8a-8d7f-11e70ded2244"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:12:49 crc kubenswrapper[4612]: I0227 08:12:49.273817 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ff8b7423-9c8a-4c8a-8d7f-11e70ded2244-config-data" (OuterVolumeSpecName: "config-data") pod "ff8b7423-9c8a-4c8a-8d7f-11e70ded2244" (UID: "ff8b7423-9c8a-4c8a-8d7f-11e70ded2244"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:12:49 crc kubenswrapper[4612]: I0227 08:12:49.308145 4612 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff8b7423-9c8a-4c8a-8d7f-11e70ded2244-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 08:12:49 crc kubenswrapper[4612]: I0227 08:12:49.308168 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rtkd2\" (UniqueName: \"kubernetes.io/projected/ff8b7423-9c8a-4c8a-8d7f-11e70ded2244-kube-api-access-rtkd2\") on node \"crc\" DevicePath \"\"" Feb 27 08:12:49 crc kubenswrapper[4612]: I0227 08:12:49.308179 4612 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ff8b7423-9c8a-4c8a-8d7f-11e70ded2244-config-data\") on node \"crc\" DevicePath \"\"" Feb 27 08:12:49 crc kubenswrapper[4612]: E0227 08:12:49.548641 4612 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3c4d5bb6_8ca7_41d7_8e30_af438a0e267a.slice/crio-conmon-b20dc16d14af8f07dddc21f5ce3968bf2ad89f62cb1e10b67ad0a10affdda173.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3c4d5bb6_8ca7_41d7_8e30_af438a0e267a.slice/crio-b20dc16d14af8f07dddc21f5ce3968bf2ad89f62cb1e10b67ad0a10affdda173.scope\": RecentStats: unable to find data in memory cache]" Feb 27 08:12:49 crc kubenswrapper[4612]: I0227 08:12:49.570649 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Feb 27 08:12:49 crc kubenswrapper[4612]: I0227 08:12:49.767001 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"ff8b7423-9c8a-4c8a-8d7f-11e70ded2244","Type":"ContainerDied","Data":"83e9f6a60dbda9b9f4133346657757915edd38f1e46079776f3938e91d88f745"} Feb 27 08:12:49 crc kubenswrapper[4612]: I0227 08:12:49.767051 4612 scope.go:117] "RemoveContainer" containerID="205e0e7c8976d96647c6a5c7b44e2fbf349cb1346ac71036d6ee20ad722ab20b" Feb 27 08:12:49 crc kubenswrapper[4612]: I0227 08:12:49.767124 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 27 08:12:49 crc kubenswrapper[4612]: I0227 08:12:49.768528 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"24c836ae-ba59-4432-acee-fd17b60c8236","Type":"ContainerStarted","Data":"74a68007e307113a98c29c69c75669615adeda4c0d4f9dfff4c8cb68043d797e"} Feb 27 08:12:49 crc kubenswrapper[4612]: I0227 08:12:49.852023 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Feb 27 08:12:49 crc kubenswrapper[4612]: I0227 08:12:49.860770 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Feb 27 08:12:49 crc kubenswrapper[4612]: I0227 08:12:49.901178 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Feb 27 08:12:49 crc kubenswrapper[4612]: E0227 08:12:49.901598 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff8b7423-9c8a-4c8a-8d7f-11e70ded2244" containerName="nova-scheduler-scheduler" Feb 27 08:12:49 crc kubenswrapper[4612]: I0227 08:12:49.901614 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff8b7423-9c8a-4c8a-8d7f-11e70ded2244" containerName="nova-scheduler-scheduler" Feb 27 08:12:49 crc kubenswrapper[4612]: I0227 08:12:49.901833 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="ff8b7423-9c8a-4c8a-8d7f-11e70ded2244" containerName="nova-scheduler-scheduler" Feb 27 08:12:49 crc kubenswrapper[4612]: I0227 08:12:49.902420 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 27 08:12:49 crc kubenswrapper[4612]: I0227 08:12:49.904309 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Feb 27 08:12:49 crc kubenswrapper[4612]: I0227 08:12:49.917609 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Feb 27 08:12:50 crc kubenswrapper[4612]: I0227 08:12:50.023531 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/43270ba1-f597-4ca7-bc1e-741d7b915940-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"43270ba1-f597-4ca7-bc1e-741d7b915940\") " pod="openstack/nova-scheduler-0" Feb 27 08:12:50 crc kubenswrapper[4612]: I0227 08:12:50.024077 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xx2t8\" (UniqueName: \"kubernetes.io/projected/43270ba1-f597-4ca7-bc1e-741d7b915940-kube-api-access-xx2t8\") pod \"nova-scheduler-0\" (UID: \"43270ba1-f597-4ca7-bc1e-741d7b915940\") " pod="openstack/nova-scheduler-0" Feb 27 08:12:50 crc kubenswrapper[4612]: I0227 08:12:50.024145 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/43270ba1-f597-4ca7-bc1e-741d7b915940-config-data\") pod \"nova-scheduler-0\" (UID: \"43270ba1-f597-4ca7-bc1e-741d7b915940\") " pod="openstack/nova-scheduler-0" Feb 27 08:12:50 crc kubenswrapper[4612]: I0227 08:12:50.126355 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/43270ba1-f597-4ca7-bc1e-741d7b915940-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"43270ba1-f597-4ca7-bc1e-741d7b915940\") " pod="openstack/nova-scheduler-0" Feb 27 08:12:50 crc kubenswrapper[4612]: I0227 08:12:50.126483 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xx2t8\" (UniqueName: \"kubernetes.io/projected/43270ba1-f597-4ca7-bc1e-741d7b915940-kube-api-access-xx2t8\") pod \"nova-scheduler-0\" (UID: \"43270ba1-f597-4ca7-bc1e-741d7b915940\") " pod="openstack/nova-scheduler-0" Feb 27 08:12:50 crc kubenswrapper[4612]: I0227 08:12:50.126506 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/43270ba1-f597-4ca7-bc1e-741d7b915940-config-data\") pod \"nova-scheduler-0\" (UID: \"43270ba1-f597-4ca7-bc1e-741d7b915940\") " pod="openstack/nova-scheduler-0" Feb 27 08:12:50 crc kubenswrapper[4612]: I0227 08:12:50.130118 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/43270ba1-f597-4ca7-bc1e-741d7b915940-config-data\") pod \"nova-scheduler-0\" (UID: \"43270ba1-f597-4ca7-bc1e-741d7b915940\") " pod="openstack/nova-scheduler-0" Feb 27 08:12:50 crc kubenswrapper[4612]: I0227 08:12:50.130228 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/43270ba1-f597-4ca7-bc1e-741d7b915940-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"43270ba1-f597-4ca7-bc1e-741d7b915940\") " pod="openstack/nova-scheduler-0" Feb 27 08:12:50 crc kubenswrapper[4612]: I0227 08:12:50.141062 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xx2t8\" (UniqueName: \"kubernetes.io/projected/43270ba1-f597-4ca7-bc1e-741d7b915940-kube-api-access-xx2t8\") pod \"nova-scheduler-0\" (UID: \"43270ba1-f597-4ca7-bc1e-741d7b915940\") " pod="openstack/nova-scheduler-0" Feb 27 08:12:50 crc kubenswrapper[4612]: I0227 08:12:50.301306 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Feb 27 08:12:50 crc kubenswrapper[4612]: I0227 08:12:50.421437 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 27 08:12:50 crc kubenswrapper[4612]: I0227 08:12:50.534150 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7690cb3e-e335-4dc2-bf76-0a3239bc6a5c-combined-ca-bundle\") pod \"7690cb3e-e335-4dc2-bf76-0a3239bc6a5c\" (UID: \"7690cb3e-e335-4dc2-bf76-0a3239bc6a5c\") " Feb 27 08:12:50 crc kubenswrapper[4612]: I0227 08:12:50.534233 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7690cb3e-e335-4dc2-bf76-0a3239bc6a5c-public-tls-certs\") pod \"7690cb3e-e335-4dc2-bf76-0a3239bc6a5c\" (UID: \"7690cb3e-e335-4dc2-bf76-0a3239bc6a5c\") " Feb 27 08:12:50 crc kubenswrapper[4612]: I0227 08:12:50.534321 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7690cb3e-e335-4dc2-bf76-0a3239bc6a5c-config-data\") pod \"7690cb3e-e335-4dc2-bf76-0a3239bc6a5c\" (UID: \"7690cb3e-e335-4dc2-bf76-0a3239bc6a5c\") " Feb 27 08:12:50 crc kubenswrapper[4612]: I0227 08:12:50.534353 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xxh4z\" (UniqueName: \"kubernetes.io/projected/7690cb3e-e335-4dc2-bf76-0a3239bc6a5c-kube-api-access-xxh4z\") pod \"7690cb3e-e335-4dc2-bf76-0a3239bc6a5c\" (UID: \"7690cb3e-e335-4dc2-bf76-0a3239bc6a5c\") " Feb 27 08:12:50 crc kubenswrapper[4612]: I0227 08:12:50.534410 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7690cb3e-e335-4dc2-bf76-0a3239bc6a5c-logs\") pod \"7690cb3e-e335-4dc2-bf76-0a3239bc6a5c\" (UID: \"7690cb3e-e335-4dc2-bf76-0a3239bc6a5c\") " Feb 27 08:12:50 crc kubenswrapper[4612]: I0227 08:12:50.534435 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7690cb3e-e335-4dc2-bf76-0a3239bc6a5c-internal-tls-certs\") pod \"7690cb3e-e335-4dc2-bf76-0a3239bc6a5c\" (UID: \"7690cb3e-e335-4dc2-bf76-0a3239bc6a5c\") " Feb 27 08:12:50 crc kubenswrapper[4612]: I0227 08:12:50.535979 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7690cb3e-e335-4dc2-bf76-0a3239bc6a5c-logs" (OuterVolumeSpecName: "logs") pod "7690cb3e-e335-4dc2-bf76-0a3239bc6a5c" (UID: "7690cb3e-e335-4dc2-bf76-0a3239bc6a5c"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 08:12:50 crc kubenswrapper[4612]: I0227 08:12:50.540441 4612 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7690cb3e-e335-4dc2-bf76-0a3239bc6a5c-logs\") on node \"crc\" DevicePath \"\"" Feb 27 08:12:50 crc kubenswrapper[4612]: I0227 08:12:50.547838 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7690cb3e-e335-4dc2-bf76-0a3239bc6a5c-kube-api-access-xxh4z" (OuterVolumeSpecName: "kube-api-access-xxh4z") pod "7690cb3e-e335-4dc2-bf76-0a3239bc6a5c" (UID: "7690cb3e-e335-4dc2-bf76-0a3239bc6a5c"). InnerVolumeSpecName "kube-api-access-xxh4z". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 08:12:50 crc kubenswrapper[4612]: I0227 08:12:50.577824 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7690cb3e-e335-4dc2-bf76-0a3239bc6a5c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7690cb3e-e335-4dc2-bf76-0a3239bc6a5c" (UID: "7690cb3e-e335-4dc2-bf76-0a3239bc6a5c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:12:50 crc kubenswrapper[4612]: I0227 08:12:50.584025 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7690cb3e-e335-4dc2-bf76-0a3239bc6a5c-config-data" (OuterVolumeSpecName: "config-data") pod "7690cb3e-e335-4dc2-bf76-0a3239bc6a5c" (UID: "7690cb3e-e335-4dc2-bf76-0a3239bc6a5c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:12:50 crc kubenswrapper[4612]: I0227 08:12:50.610294 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7690cb3e-e335-4dc2-bf76-0a3239bc6a5c-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "7690cb3e-e335-4dc2-bf76-0a3239bc6a5c" (UID: "7690cb3e-e335-4dc2-bf76-0a3239bc6a5c"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:12:50 crc kubenswrapper[4612]: I0227 08:12:50.623154 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7690cb3e-e335-4dc2-bf76-0a3239bc6a5c-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "7690cb3e-e335-4dc2-bf76-0a3239bc6a5c" (UID: "7690cb3e-e335-4dc2-bf76-0a3239bc6a5c"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:12:50 crc kubenswrapper[4612]: I0227 08:12:50.644649 4612 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7690cb3e-e335-4dc2-bf76-0a3239bc6a5c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 08:12:50 crc kubenswrapper[4612]: I0227 08:12:50.644702 4612 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7690cb3e-e335-4dc2-bf76-0a3239bc6a5c-public-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 27 08:12:50 crc kubenswrapper[4612]: I0227 08:12:50.644713 4612 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7690cb3e-e335-4dc2-bf76-0a3239bc6a5c-config-data\") on node \"crc\" DevicePath \"\"" Feb 27 08:12:50 crc kubenswrapper[4612]: I0227 08:12:50.644723 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xxh4z\" (UniqueName: \"kubernetes.io/projected/7690cb3e-e335-4dc2-bf76-0a3239bc6a5c-kube-api-access-xxh4z\") on node \"crc\" DevicePath \"\"" Feb 27 08:12:50 crc kubenswrapper[4612]: I0227 08:12:50.644738 4612 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7690cb3e-e335-4dc2-bf76-0a3239bc6a5c-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Feb 27 08:12:50 crc kubenswrapper[4612]: I0227 08:12:50.754369 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Feb 27 08:12:50 crc kubenswrapper[4612]: I0227 08:12:50.778834 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"43270ba1-f597-4ca7-bc1e-741d7b915940","Type":"ContainerStarted","Data":"ac8911cad8b68a87e8cc8c026003cef0d1f3093fbb0c01809c2a46b29b60f8f2"} Feb 27 08:12:50 crc kubenswrapper[4612]: I0227 08:12:50.788597 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"24c836ae-ba59-4432-acee-fd17b60c8236","Type":"ContainerStarted","Data":"ecdee75ecfc362e2e46374abfdf52315680975bea0e58ab994bc156cef2cbc97"} Feb 27 08:12:50 crc kubenswrapper[4612]: I0227 08:12:50.788758 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"24c836ae-ba59-4432-acee-fd17b60c8236","Type":"ContainerStarted","Data":"7d40ae5ea5eef994843a52dcfbad2e43a0ec2bdbb28c353ad0a5e4658929e3c1"} Feb 27 08:12:50 crc kubenswrapper[4612]: I0227 08:12:50.793211 4612 generic.go:334] "Generic (PLEG): container finished" podID="7690cb3e-e335-4dc2-bf76-0a3239bc6a5c" containerID="69a882e4cb7dd05b7ecdd3b266255cd5975c6faa28760ebe00d8461124b1c87b" exitCode=0 Feb 27 08:12:50 crc kubenswrapper[4612]: I0227 08:12:50.793287 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"7690cb3e-e335-4dc2-bf76-0a3239bc6a5c","Type":"ContainerDied","Data":"69a882e4cb7dd05b7ecdd3b266255cd5975c6faa28760ebe00d8461124b1c87b"} Feb 27 08:12:50 crc kubenswrapper[4612]: I0227 08:12:50.793317 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"7690cb3e-e335-4dc2-bf76-0a3239bc6a5c","Type":"ContainerDied","Data":"7e554d57b078dd5fe7a6f092dd79c1dda530d38fe949305e09c6c3378afe0bf2"} Feb 27 08:12:50 crc kubenswrapper[4612]: I0227 08:12:50.793338 4612 scope.go:117] "RemoveContainer" containerID="69a882e4cb7dd05b7ecdd3b266255cd5975c6faa28760ebe00d8461124b1c87b" Feb 27 08:12:50 crc kubenswrapper[4612]: I0227 08:12:50.793525 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 27 08:12:50 crc kubenswrapper[4612]: I0227 08:12:50.866375 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.866356594 podStartE2EDuration="2.866356594s" podCreationTimestamp="2026-02-27 08:12:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 08:12:50.824127581 +0000 UTC m=+1428.678057589" watchObservedRunningTime="2026-02-27 08:12:50.866356594 +0000 UTC m=+1428.720286592" Feb 27 08:12:50 crc kubenswrapper[4612]: I0227 08:12:50.905020 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ff8b7423-9c8a-4c8a-8d7f-11e70ded2244" path="/var/lib/kubelet/pods/ff8b7423-9c8a-4c8a-8d7f-11e70ded2244/volumes" Feb 27 08:12:50 crc kubenswrapper[4612]: I0227 08:12:50.907888 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Feb 27 08:12:50 crc kubenswrapper[4612]: I0227 08:12:50.910215 4612 scope.go:117] "RemoveContainer" containerID="7c93b2b51dcc74913ce975a46c196bb10a5a0b39463a4799669c1c8957733d22" Feb 27 08:12:50 crc kubenswrapper[4612]: I0227 08:12:50.928698 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Feb 27 08:12:50 crc kubenswrapper[4612]: I0227 08:12:50.928749 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Feb 27 08:12:50 crc kubenswrapper[4612]: E0227 08:12:50.929109 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7690cb3e-e335-4dc2-bf76-0a3239bc6a5c" containerName="nova-api-api" Feb 27 08:12:50 crc kubenswrapper[4612]: I0227 08:12:50.929121 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="7690cb3e-e335-4dc2-bf76-0a3239bc6a5c" containerName="nova-api-api" Feb 27 08:12:50 crc kubenswrapper[4612]: E0227 08:12:50.929140 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7690cb3e-e335-4dc2-bf76-0a3239bc6a5c" containerName="nova-api-log" Feb 27 08:12:50 crc kubenswrapper[4612]: I0227 08:12:50.929146 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="7690cb3e-e335-4dc2-bf76-0a3239bc6a5c" containerName="nova-api-log" Feb 27 08:12:50 crc kubenswrapper[4612]: I0227 08:12:50.929320 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="7690cb3e-e335-4dc2-bf76-0a3239bc6a5c" containerName="nova-api-api" Feb 27 08:12:50 crc kubenswrapper[4612]: I0227 08:12:50.929337 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="7690cb3e-e335-4dc2-bf76-0a3239bc6a5c" containerName="nova-api-log" Feb 27 08:12:50 crc kubenswrapper[4612]: I0227 08:12:50.930456 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 27 08:12:50 crc kubenswrapper[4612]: I0227 08:12:50.939954 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Feb 27 08:12:50 crc kubenswrapper[4612]: I0227 08:12:50.964358 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Feb 27 08:12:50 crc kubenswrapper[4612]: I0227 08:12:50.964562 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Feb 27 08:12:50 crc kubenswrapper[4612]: I0227 08:12:50.964672 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Feb 27 08:12:51 crc kubenswrapper[4612]: I0227 08:12:51.013455 4612 scope.go:117] "RemoveContainer" containerID="69a882e4cb7dd05b7ecdd3b266255cd5975c6faa28760ebe00d8461124b1c87b" Feb 27 08:12:51 crc kubenswrapper[4612]: E0227 08:12:51.014812 4612 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"69a882e4cb7dd05b7ecdd3b266255cd5975c6faa28760ebe00d8461124b1c87b\": container with ID starting with 69a882e4cb7dd05b7ecdd3b266255cd5975c6faa28760ebe00d8461124b1c87b not found: ID does not exist" containerID="69a882e4cb7dd05b7ecdd3b266255cd5975c6faa28760ebe00d8461124b1c87b" Feb 27 08:12:51 crc kubenswrapper[4612]: I0227 08:12:51.014878 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"69a882e4cb7dd05b7ecdd3b266255cd5975c6faa28760ebe00d8461124b1c87b"} err="failed to get container status \"69a882e4cb7dd05b7ecdd3b266255cd5975c6faa28760ebe00d8461124b1c87b\": rpc error: code = NotFound desc = could not find container \"69a882e4cb7dd05b7ecdd3b266255cd5975c6faa28760ebe00d8461124b1c87b\": container with ID starting with 69a882e4cb7dd05b7ecdd3b266255cd5975c6faa28760ebe00d8461124b1c87b not found: ID does not exist" Feb 27 08:12:51 crc kubenswrapper[4612]: I0227 08:12:51.014907 4612 scope.go:117] "RemoveContainer" containerID="7c93b2b51dcc74913ce975a46c196bb10a5a0b39463a4799669c1c8957733d22" Feb 27 08:12:51 crc kubenswrapper[4612]: E0227 08:12:51.015406 4612 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7c93b2b51dcc74913ce975a46c196bb10a5a0b39463a4799669c1c8957733d22\": container with ID starting with 7c93b2b51dcc74913ce975a46c196bb10a5a0b39463a4799669c1c8957733d22 not found: ID does not exist" containerID="7c93b2b51dcc74913ce975a46c196bb10a5a0b39463a4799669c1c8957733d22" Feb 27 08:12:51 crc kubenswrapper[4612]: I0227 08:12:51.015446 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7c93b2b51dcc74913ce975a46c196bb10a5a0b39463a4799669c1c8957733d22"} err="failed to get container status \"7c93b2b51dcc74913ce975a46c196bb10a5a0b39463a4799669c1c8957733d22\": rpc error: code = NotFound desc = could not find container \"7c93b2b51dcc74913ce975a46c196bb10a5a0b39463a4799669c1c8957733d22\": container with ID starting with 7c93b2b51dcc74913ce975a46c196bb10a5a0b39463a4799669c1c8957733d22 not found: ID does not exist" Feb 27 08:12:51 crc kubenswrapper[4612]: I0227 08:12:51.168854 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e492f61b-94d4-4c4e-ac47-168fdb1cd997-public-tls-certs\") pod \"nova-api-0\" (UID: \"e492f61b-94d4-4c4e-ac47-168fdb1cd997\") " pod="openstack/nova-api-0" Feb 27 08:12:51 crc kubenswrapper[4612]: I0227 08:12:51.169185 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e492f61b-94d4-4c4e-ac47-168fdb1cd997-internal-tls-certs\") pod \"nova-api-0\" (UID: \"e492f61b-94d4-4c4e-ac47-168fdb1cd997\") " pod="openstack/nova-api-0" Feb 27 08:12:51 crc kubenswrapper[4612]: I0227 08:12:51.169214 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bdswf\" (UniqueName: \"kubernetes.io/projected/e492f61b-94d4-4c4e-ac47-168fdb1cd997-kube-api-access-bdswf\") pod \"nova-api-0\" (UID: \"e492f61b-94d4-4c4e-ac47-168fdb1cd997\") " pod="openstack/nova-api-0" Feb 27 08:12:51 crc kubenswrapper[4612]: I0227 08:12:51.169251 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e492f61b-94d4-4c4e-ac47-168fdb1cd997-logs\") pod \"nova-api-0\" (UID: \"e492f61b-94d4-4c4e-ac47-168fdb1cd997\") " pod="openstack/nova-api-0" Feb 27 08:12:51 crc kubenswrapper[4612]: I0227 08:12:51.169322 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e492f61b-94d4-4c4e-ac47-168fdb1cd997-config-data\") pod \"nova-api-0\" (UID: \"e492f61b-94d4-4c4e-ac47-168fdb1cd997\") " pod="openstack/nova-api-0" Feb 27 08:12:51 crc kubenswrapper[4612]: I0227 08:12:51.169339 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e492f61b-94d4-4c4e-ac47-168fdb1cd997-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"e492f61b-94d4-4c4e-ac47-168fdb1cd997\") " pod="openstack/nova-api-0" Feb 27 08:12:51 crc kubenswrapper[4612]: I0227 08:12:51.270591 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e492f61b-94d4-4c4e-ac47-168fdb1cd997-logs\") pod \"nova-api-0\" (UID: \"e492f61b-94d4-4c4e-ac47-168fdb1cd997\") " pod="openstack/nova-api-0" Feb 27 08:12:51 crc kubenswrapper[4612]: I0227 08:12:51.270707 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e492f61b-94d4-4c4e-ac47-168fdb1cd997-config-data\") pod \"nova-api-0\" (UID: \"e492f61b-94d4-4c4e-ac47-168fdb1cd997\") " pod="openstack/nova-api-0" Feb 27 08:12:51 crc kubenswrapper[4612]: I0227 08:12:51.270745 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e492f61b-94d4-4c4e-ac47-168fdb1cd997-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"e492f61b-94d4-4c4e-ac47-168fdb1cd997\") " pod="openstack/nova-api-0" Feb 27 08:12:51 crc kubenswrapper[4612]: I0227 08:12:51.270857 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e492f61b-94d4-4c4e-ac47-168fdb1cd997-public-tls-certs\") pod \"nova-api-0\" (UID: \"e492f61b-94d4-4c4e-ac47-168fdb1cd997\") " pod="openstack/nova-api-0" Feb 27 08:12:51 crc kubenswrapper[4612]: I0227 08:12:51.270911 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e492f61b-94d4-4c4e-ac47-168fdb1cd997-internal-tls-certs\") pod \"nova-api-0\" (UID: \"e492f61b-94d4-4c4e-ac47-168fdb1cd997\") " pod="openstack/nova-api-0" Feb 27 08:12:51 crc kubenswrapper[4612]: I0227 08:12:51.270953 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bdswf\" (UniqueName: \"kubernetes.io/projected/e492f61b-94d4-4c4e-ac47-168fdb1cd997-kube-api-access-bdswf\") pod \"nova-api-0\" (UID: \"e492f61b-94d4-4c4e-ac47-168fdb1cd997\") " pod="openstack/nova-api-0" Feb 27 08:12:51 crc kubenswrapper[4612]: I0227 08:12:51.272022 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e492f61b-94d4-4c4e-ac47-168fdb1cd997-logs\") pod \"nova-api-0\" (UID: \"e492f61b-94d4-4c4e-ac47-168fdb1cd997\") " pod="openstack/nova-api-0" Feb 27 08:12:51 crc kubenswrapper[4612]: I0227 08:12:51.275667 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e492f61b-94d4-4c4e-ac47-168fdb1cd997-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"e492f61b-94d4-4c4e-ac47-168fdb1cd997\") " pod="openstack/nova-api-0" Feb 27 08:12:51 crc kubenswrapper[4612]: I0227 08:12:51.277472 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e492f61b-94d4-4c4e-ac47-168fdb1cd997-public-tls-certs\") pod \"nova-api-0\" (UID: \"e492f61b-94d4-4c4e-ac47-168fdb1cd997\") " pod="openstack/nova-api-0" Feb 27 08:12:51 crc kubenswrapper[4612]: I0227 08:12:51.278483 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e492f61b-94d4-4c4e-ac47-168fdb1cd997-config-data\") pod \"nova-api-0\" (UID: \"e492f61b-94d4-4c4e-ac47-168fdb1cd997\") " pod="openstack/nova-api-0" Feb 27 08:12:51 crc kubenswrapper[4612]: I0227 08:12:51.278913 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e492f61b-94d4-4c4e-ac47-168fdb1cd997-internal-tls-certs\") pod \"nova-api-0\" (UID: \"e492f61b-94d4-4c4e-ac47-168fdb1cd997\") " pod="openstack/nova-api-0" Feb 27 08:12:51 crc kubenswrapper[4612]: I0227 08:12:51.291317 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bdswf\" (UniqueName: \"kubernetes.io/projected/e492f61b-94d4-4c4e-ac47-168fdb1cd997-kube-api-access-bdswf\") pod \"nova-api-0\" (UID: \"e492f61b-94d4-4c4e-ac47-168fdb1cd997\") " pod="openstack/nova-api-0" Feb 27 08:12:51 crc kubenswrapper[4612]: I0227 08:12:51.589151 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Feb 27 08:12:51 crc kubenswrapper[4612]: I0227 08:12:51.818847 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"43270ba1-f597-4ca7-bc1e-741d7b915940","Type":"ContainerStarted","Data":"5a17778f2e8d6746dcc2cbe9e980c039a7d30e77b8be59112c51806fe4dc3e27"} Feb 27 08:12:51 crc kubenswrapper[4612]: I0227 08:12:51.840807 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.8407841940000003 podStartE2EDuration="2.840784194s" podCreationTimestamp="2026-02-27 08:12:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 08:12:51.834842943 +0000 UTC m=+1429.688772941" watchObservedRunningTime="2026-02-27 08:12:51.840784194 +0000 UTC m=+1429.694714202" Feb 27 08:12:52 crc kubenswrapper[4612]: I0227 08:12:52.109421 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Feb 27 08:12:52 crc kubenswrapper[4612]: I0227 08:12:52.835415 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"e492f61b-94d4-4c4e-ac47-168fdb1cd997","Type":"ContainerStarted","Data":"7f702f4fc6e22839d53ec44306cfb88f5cda0288c46e6009aa0df3f206837994"} Feb 27 08:12:52 crc kubenswrapper[4612]: I0227 08:12:52.835770 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"e492f61b-94d4-4c4e-ac47-168fdb1cd997","Type":"ContainerStarted","Data":"25392075a7929f99bada92235fb8a72450ab3309b7b457b9f170c9aa3f5fc442"} Feb 27 08:12:52 crc kubenswrapper[4612]: I0227 08:12:52.835784 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"e492f61b-94d4-4c4e-ac47-168fdb1cd997","Type":"ContainerStarted","Data":"76d92b5b47ec71b9359fca26e94455407c2a88251ded555c0c9646037782bb88"} Feb 27 08:12:52 crc kubenswrapper[4612]: I0227 08:12:52.871414 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.871387326 podStartE2EDuration="2.871387326s" podCreationTimestamp="2026-02-27 08:12:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 08:12:52.855361026 +0000 UTC m=+1430.709291044" watchObservedRunningTime="2026-02-27 08:12:52.871387326 +0000 UTC m=+1430.725317344" Feb 27 08:12:52 crc kubenswrapper[4612]: I0227 08:12:52.885631 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7690cb3e-e335-4dc2-bf76-0a3239bc6a5c" path="/var/lib/kubelet/pods/7690cb3e-e335-4dc2-bf76-0a3239bc6a5c/volumes" Feb 27 08:12:54 crc kubenswrapper[4612]: I0227 08:12:54.108903 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Feb 27 08:12:54 crc kubenswrapper[4612]: I0227 08:12:54.109298 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Feb 27 08:12:54 crc kubenswrapper[4612]: I0227 08:12:54.773647 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-qvxtt"] Feb 27 08:12:54 crc kubenswrapper[4612]: I0227 08:12:54.775482 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qvxtt" Feb 27 08:12:54 crc kubenswrapper[4612]: I0227 08:12:54.830281 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-qvxtt"] Feb 27 08:12:54 crc kubenswrapper[4612]: I0227 08:12:54.951565 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d73fa1d-60eb-4c32-a82d-96f26011b04d-catalog-content\") pod \"redhat-operators-qvxtt\" (UID: \"1d73fa1d-60eb-4c32-a82d-96f26011b04d\") " pod="openshift-marketplace/redhat-operators-qvxtt" Feb 27 08:12:54 crc kubenswrapper[4612]: I0227 08:12:54.952020 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2n9dl\" (UniqueName: \"kubernetes.io/projected/1d73fa1d-60eb-4c32-a82d-96f26011b04d-kube-api-access-2n9dl\") pod \"redhat-operators-qvxtt\" (UID: \"1d73fa1d-60eb-4c32-a82d-96f26011b04d\") " pod="openshift-marketplace/redhat-operators-qvxtt" Feb 27 08:12:54 crc kubenswrapper[4612]: I0227 08:12:54.952097 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d73fa1d-60eb-4c32-a82d-96f26011b04d-utilities\") pod \"redhat-operators-qvxtt\" (UID: \"1d73fa1d-60eb-4c32-a82d-96f26011b04d\") " pod="openshift-marketplace/redhat-operators-qvxtt" Feb 27 08:12:55 crc kubenswrapper[4612]: I0227 08:12:55.054029 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d73fa1d-60eb-4c32-a82d-96f26011b04d-catalog-content\") pod \"redhat-operators-qvxtt\" (UID: \"1d73fa1d-60eb-4c32-a82d-96f26011b04d\") " pod="openshift-marketplace/redhat-operators-qvxtt" Feb 27 08:12:55 crc kubenswrapper[4612]: I0227 08:12:55.054135 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2n9dl\" (UniqueName: \"kubernetes.io/projected/1d73fa1d-60eb-4c32-a82d-96f26011b04d-kube-api-access-2n9dl\") pod \"redhat-operators-qvxtt\" (UID: \"1d73fa1d-60eb-4c32-a82d-96f26011b04d\") " pod="openshift-marketplace/redhat-operators-qvxtt" Feb 27 08:12:55 crc kubenswrapper[4612]: I0227 08:12:55.054250 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d73fa1d-60eb-4c32-a82d-96f26011b04d-utilities\") pod \"redhat-operators-qvxtt\" (UID: \"1d73fa1d-60eb-4c32-a82d-96f26011b04d\") " pod="openshift-marketplace/redhat-operators-qvxtt" Feb 27 08:12:55 crc kubenswrapper[4612]: I0227 08:12:55.054540 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d73fa1d-60eb-4c32-a82d-96f26011b04d-catalog-content\") pod \"redhat-operators-qvxtt\" (UID: \"1d73fa1d-60eb-4c32-a82d-96f26011b04d\") " pod="openshift-marketplace/redhat-operators-qvxtt" Feb 27 08:12:55 crc kubenswrapper[4612]: I0227 08:12:55.054919 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d73fa1d-60eb-4c32-a82d-96f26011b04d-utilities\") pod \"redhat-operators-qvxtt\" (UID: \"1d73fa1d-60eb-4c32-a82d-96f26011b04d\") " pod="openshift-marketplace/redhat-operators-qvxtt" Feb 27 08:12:55 crc kubenswrapper[4612]: I0227 08:12:55.073948 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2n9dl\" (UniqueName: \"kubernetes.io/projected/1d73fa1d-60eb-4c32-a82d-96f26011b04d-kube-api-access-2n9dl\") pod \"redhat-operators-qvxtt\" (UID: \"1d73fa1d-60eb-4c32-a82d-96f26011b04d\") " pod="openshift-marketplace/redhat-operators-qvxtt" Feb 27 08:12:55 crc kubenswrapper[4612]: I0227 08:12:55.125216 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qvxtt" Feb 27 08:12:55 crc kubenswrapper[4612]: I0227 08:12:55.311995 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Feb 27 08:12:55 crc kubenswrapper[4612]: I0227 08:12:55.616915 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-qvxtt"] Feb 27 08:12:55 crc kubenswrapper[4612]: I0227 08:12:55.890561 4612 generic.go:334] "Generic (PLEG): container finished" podID="1d73fa1d-60eb-4c32-a82d-96f26011b04d" containerID="ae32025fde9b0cf665bb5327e5651f0c68488f2101ddc8578c862c314de37807" exitCode=0 Feb 27 08:12:55 crc kubenswrapper[4612]: I0227 08:12:55.890638 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qvxtt" event={"ID":"1d73fa1d-60eb-4c32-a82d-96f26011b04d","Type":"ContainerDied","Data":"ae32025fde9b0cf665bb5327e5651f0c68488f2101ddc8578c862c314de37807"} Feb 27 08:12:55 crc kubenswrapper[4612]: I0227 08:12:55.891018 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qvxtt" event={"ID":"1d73fa1d-60eb-4c32-a82d-96f26011b04d","Type":"ContainerStarted","Data":"5dbdb6991e292e05257ebb41b982169110d884f2736c6e613a1921cf15789e83"} Feb 27 08:12:56 crc kubenswrapper[4612]: I0227 08:12:56.901310 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qvxtt" event={"ID":"1d73fa1d-60eb-4c32-a82d-96f26011b04d","Type":"ContainerStarted","Data":"f79143078a8cbf26ec7fdff20b8ad5880be4facea8dc60f5969d90525543ffc4"} Feb 27 08:12:59 crc kubenswrapper[4612]: I0227 08:12:59.108701 4612 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Feb 27 08:12:59 crc kubenswrapper[4612]: I0227 08:12:59.109080 4612 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Feb 27 08:12:59 crc kubenswrapper[4612]: E0227 08:12:59.810341 4612 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3c4d5bb6_8ca7_41d7_8e30_af438a0e267a.slice/crio-conmon-b20dc16d14af8f07dddc21f5ce3968bf2ad89f62cb1e10b67ad0a10affdda173.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3c4d5bb6_8ca7_41d7_8e30_af438a0e267a.slice/crio-b20dc16d14af8f07dddc21f5ce3968bf2ad89f62cb1e10b67ad0a10affdda173.scope\": RecentStats: unable to find data in memory cache]" Feb 27 08:13:00 crc kubenswrapper[4612]: I0227 08:13:00.120883 4612 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="24c836ae-ba59-4432-acee-fd17b60c8236" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.211:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Feb 27 08:13:00 crc kubenswrapper[4612]: I0227 08:13:00.120905 4612 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="24c836ae-ba59-4432-acee-fd17b60c8236" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.211:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Feb 27 08:13:00 crc kubenswrapper[4612]: I0227 08:13:00.304003 4612 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Feb 27 08:13:00 crc kubenswrapper[4612]: I0227 08:13:00.342576 4612 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Feb 27 08:13:00 crc kubenswrapper[4612]: I0227 08:13:00.984236 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Feb 27 08:13:01 crc kubenswrapper[4612]: I0227 08:13:01.589892 4612 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Feb 27 08:13:01 crc kubenswrapper[4612]: I0227 08:13:01.589939 4612 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Feb 27 08:13:02 crc kubenswrapper[4612]: I0227 08:13:02.604894 4612 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="e492f61b-94d4-4c4e-ac47-168fdb1cd997" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.213:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Feb 27 08:13:02 crc kubenswrapper[4612]: I0227 08:13:02.604915 4612 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="e492f61b-94d4-4c4e-ac47-168fdb1cd997" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.213:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Feb 27 08:13:03 crc kubenswrapper[4612]: I0227 08:13:03.985100 4612 generic.go:334] "Generic (PLEG): container finished" podID="1d73fa1d-60eb-4c32-a82d-96f26011b04d" containerID="f79143078a8cbf26ec7fdff20b8ad5880be4facea8dc60f5969d90525543ffc4" exitCode=0 Feb 27 08:13:03 crc kubenswrapper[4612]: I0227 08:13:03.985409 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qvxtt" event={"ID":"1d73fa1d-60eb-4c32-a82d-96f26011b04d","Type":"ContainerDied","Data":"f79143078a8cbf26ec7fdff20b8ad5880be4facea8dc60f5969d90525543ffc4"} Feb 27 08:13:04 crc kubenswrapper[4612]: I0227 08:13:04.234043 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Feb 27 08:13:05 crc kubenswrapper[4612]: I0227 08:13:05.020072 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qvxtt" event={"ID":"1d73fa1d-60eb-4c32-a82d-96f26011b04d","Type":"ContainerStarted","Data":"4df1e0965c1632a4513b96f78bb286899b2034df53742ac3953ed35fe0e59ed8"} Feb 27 08:13:05 crc kubenswrapper[4612]: I0227 08:13:05.057054 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-qvxtt" podStartSLOduration=2.584294938 podStartE2EDuration="11.057031789s" podCreationTimestamp="2026-02-27 08:12:54 +0000 UTC" firstStartedPulling="2026-02-27 08:12:55.892673051 +0000 UTC m=+1433.746603069" lastFinishedPulling="2026-02-27 08:13:04.365409922 +0000 UTC m=+1442.219339920" observedRunningTime="2026-02-27 08:13:05.039266728 +0000 UTC m=+1442.893196726" watchObservedRunningTime="2026-02-27 08:13:05.057031789 +0000 UTC m=+1442.910961797" Feb 27 08:13:05 crc kubenswrapper[4612]: I0227 08:13:05.127445 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-qvxtt" Feb 27 08:13:05 crc kubenswrapper[4612]: I0227 08:13:05.127755 4612 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-qvxtt" Feb 27 08:13:06 crc kubenswrapper[4612]: I0227 08:13:06.301747 4612 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-qvxtt" podUID="1d73fa1d-60eb-4c32-a82d-96f26011b04d" containerName="registry-server" probeResult="failure" output=< Feb 27 08:13:06 crc kubenswrapper[4612]: timeout: failed to connect service ":50051" within 1s Feb 27 08:13:06 crc kubenswrapper[4612]: > Feb 27 08:13:09 crc kubenswrapper[4612]: I0227 08:13:09.116912 4612 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Feb 27 08:13:09 crc kubenswrapper[4612]: I0227 08:13:09.129115 4612 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Feb 27 08:13:09 crc kubenswrapper[4612]: I0227 08:13:09.131933 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Feb 27 08:13:10 crc kubenswrapper[4612]: I0227 08:13:10.077022 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Feb 27 08:13:11 crc kubenswrapper[4612]: I0227 08:13:11.598241 4612 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Feb 27 08:13:11 crc kubenswrapper[4612]: I0227 08:13:11.598313 4612 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Feb 27 08:13:11 crc kubenswrapper[4612]: I0227 08:13:11.598880 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Feb 27 08:13:11 crc kubenswrapper[4612]: I0227 08:13:11.598936 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Feb 27 08:13:11 crc kubenswrapper[4612]: I0227 08:13:11.607776 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Feb 27 08:13:11 crc kubenswrapper[4612]: I0227 08:13:11.609435 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Feb 27 08:13:14 crc kubenswrapper[4612]: I0227 08:13:14.275028 4612 scope.go:117] "RemoveContainer" containerID="a0cfe92dc9f1a9bddbcb67a210c2a83330e1fdb4e95bccc2b9d8a7c4b47dd5aa" Feb 27 08:13:14 crc kubenswrapper[4612]: I0227 08:13:14.476000 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-nvkhs"] Feb 27 08:13:14 crc kubenswrapper[4612]: I0227 08:13:14.478874 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nvkhs" Feb 27 08:13:14 crc kubenswrapper[4612]: I0227 08:13:14.498443 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-nvkhs"] Feb 27 08:13:14 crc kubenswrapper[4612]: I0227 08:13:14.620791 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vmj7v\" (UniqueName: \"kubernetes.io/projected/3485c18d-1e56-4c62-b27e-609ee08ab06f-kube-api-access-vmj7v\") pod \"community-operators-nvkhs\" (UID: \"3485c18d-1e56-4c62-b27e-609ee08ab06f\") " pod="openshift-marketplace/community-operators-nvkhs" Feb 27 08:13:14 crc kubenswrapper[4612]: I0227 08:13:14.620902 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3485c18d-1e56-4c62-b27e-609ee08ab06f-catalog-content\") pod \"community-operators-nvkhs\" (UID: \"3485c18d-1e56-4c62-b27e-609ee08ab06f\") " pod="openshift-marketplace/community-operators-nvkhs" Feb 27 08:13:14 crc kubenswrapper[4612]: I0227 08:13:14.621002 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3485c18d-1e56-4c62-b27e-609ee08ab06f-utilities\") pod \"community-operators-nvkhs\" (UID: \"3485c18d-1e56-4c62-b27e-609ee08ab06f\") " pod="openshift-marketplace/community-operators-nvkhs" Feb 27 08:13:14 crc kubenswrapper[4612]: I0227 08:13:14.723041 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3485c18d-1e56-4c62-b27e-609ee08ab06f-utilities\") pod \"community-operators-nvkhs\" (UID: \"3485c18d-1e56-4c62-b27e-609ee08ab06f\") " pod="openshift-marketplace/community-operators-nvkhs" Feb 27 08:13:14 crc kubenswrapper[4612]: I0227 08:13:14.723113 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vmj7v\" (UniqueName: \"kubernetes.io/projected/3485c18d-1e56-4c62-b27e-609ee08ab06f-kube-api-access-vmj7v\") pod \"community-operators-nvkhs\" (UID: \"3485c18d-1e56-4c62-b27e-609ee08ab06f\") " pod="openshift-marketplace/community-operators-nvkhs" Feb 27 08:13:14 crc kubenswrapper[4612]: I0227 08:13:14.723180 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3485c18d-1e56-4c62-b27e-609ee08ab06f-catalog-content\") pod \"community-operators-nvkhs\" (UID: \"3485c18d-1e56-4c62-b27e-609ee08ab06f\") " pod="openshift-marketplace/community-operators-nvkhs" Feb 27 08:13:14 crc kubenswrapper[4612]: I0227 08:13:14.723620 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3485c18d-1e56-4c62-b27e-609ee08ab06f-catalog-content\") pod \"community-operators-nvkhs\" (UID: \"3485c18d-1e56-4c62-b27e-609ee08ab06f\") " pod="openshift-marketplace/community-operators-nvkhs" Feb 27 08:13:14 crc kubenswrapper[4612]: I0227 08:13:14.723840 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3485c18d-1e56-4c62-b27e-609ee08ab06f-utilities\") pod \"community-operators-nvkhs\" (UID: \"3485c18d-1e56-4c62-b27e-609ee08ab06f\") " pod="openshift-marketplace/community-operators-nvkhs" Feb 27 08:13:14 crc kubenswrapper[4612]: I0227 08:13:14.745887 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vmj7v\" (UniqueName: \"kubernetes.io/projected/3485c18d-1e56-4c62-b27e-609ee08ab06f-kube-api-access-vmj7v\") pod \"community-operators-nvkhs\" (UID: \"3485c18d-1e56-4c62-b27e-609ee08ab06f\") " pod="openshift-marketplace/community-operators-nvkhs" Feb 27 08:13:14 crc kubenswrapper[4612]: I0227 08:13:14.797843 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nvkhs" Feb 27 08:13:15 crc kubenswrapper[4612]: I0227 08:13:15.363343 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-nvkhs"] Feb 27 08:13:16 crc kubenswrapper[4612]: I0227 08:13:16.143961 4612 generic.go:334] "Generic (PLEG): container finished" podID="3485c18d-1e56-4c62-b27e-609ee08ab06f" containerID="57694b1cb9056ecc305864d2b80b28321b96260b03532d2f92d6308a31d91632" exitCode=0 Feb 27 08:13:16 crc kubenswrapper[4612]: I0227 08:13:16.144259 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nvkhs" event={"ID":"3485c18d-1e56-4c62-b27e-609ee08ab06f","Type":"ContainerDied","Data":"57694b1cb9056ecc305864d2b80b28321b96260b03532d2f92d6308a31d91632"} Feb 27 08:13:16 crc kubenswrapper[4612]: I0227 08:13:16.144284 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nvkhs" event={"ID":"3485c18d-1e56-4c62-b27e-609ee08ab06f","Type":"ContainerStarted","Data":"e7154d0ed03bfa21ed4df7ef658ab204bfdebb7248b134369ed681ba77950dd3"} Feb 27 08:13:16 crc kubenswrapper[4612]: I0227 08:13:16.146036 4612 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 27 08:13:16 crc kubenswrapper[4612]: I0227 08:13:16.178891 4612 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-qvxtt" podUID="1d73fa1d-60eb-4c32-a82d-96f26011b04d" containerName="registry-server" probeResult="failure" output=< Feb 27 08:13:16 crc kubenswrapper[4612]: timeout: failed to connect service ":50051" within 1s Feb 27 08:13:16 crc kubenswrapper[4612]: > Feb 27 08:13:17 crc kubenswrapper[4612]: I0227 08:13:17.158544 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nvkhs" event={"ID":"3485c18d-1e56-4c62-b27e-609ee08ab06f","Type":"ContainerStarted","Data":"08e8efea5e358f6e6010676ad3d80b7e2d0ece31f38fa9eb5f45ad3728fa44f5"} Feb 27 08:13:19 crc kubenswrapper[4612]: I0227 08:13:19.177764 4612 generic.go:334] "Generic (PLEG): container finished" podID="3485c18d-1e56-4c62-b27e-609ee08ab06f" containerID="08e8efea5e358f6e6010676ad3d80b7e2d0ece31f38fa9eb5f45ad3728fa44f5" exitCode=0 Feb 27 08:13:19 crc kubenswrapper[4612]: I0227 08:13:19.177847 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nvkhs" event={"ID":"3485c18d-1e56-4c62-b27e-609ee08ab06f","Type":"ContainerDied","Data":"08e8efea5e358f6e6010676ad3d80b7e2d0ece31f38fa9eb5f45ad3728fa44f5"} Feb 27 08:13:19 crc kubenswrapper[4612]: I0227 08:13:19.402604 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Feb 27 08:13:20 crc kubenswrapper[4612]: I0227 08:13:20.187884 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nvkhs" event={"ID":"3485c18d-1e56-4c62-b27e-609ee08ab06f","Type":"ContainerStarted","Data":"ed88b222ece5c62b9fe1d6a3abb48a5ae8499490751152a7f44fe3b59b44b23e"} Feb 27 08:13:20 crc kubenswrapper[4612]: I0227 08:13:20.217746 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-nvkhs" podStartSLOduration=2.795666111 podStartE2EDuration="6.217728627s" podCreationTimestamp="2026-02-27 08:13:14 +0000 UTC" firstStartedPulling="2026-02-27 08:13:16.145846825 +0000 UTC m=+1453.999776823" lastFinishedPulling="2026-02-27 08:13:19.567909341 +0000 UTC m=+1457.421839339" observedRunningTime="2026-02-27 08:13:20.217528651 +0000 UTC m=+1458.071458649" watchObservedRunningTime="2026-02-27 08:13:20.217728627 +0000 UTC m=+1458.071658625" Feb 27 08:13:20 crc kubenswrapper[4612]: I0227 08:13:20.461971 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Feb 27 08:13:23 crc kubenswrapper[4612]: I0227 08:13:23.790287 4612 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="0da436b1-803d-41bb-94ab-9cfd6fa769eb" containerName="rabbitmq" containerID="cri-o://edd43cdd132a714448da88e522b53d539c8bad75e13f40e8d8ad5319c1936c48" gracePeriod=604796 Feb 27 08:13:24 crc kubenswrapper[4612]: I0227 08:13:24.799026 4612 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-nvkhs" Feb 27 08:13:24 crc kubenswrapper[4612]: I0227 08:13:24.799277 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-nvkhs" Feb 27 08:13:24 crc kubenswrapper[4612]: I0227 08:13:24.864529 4612 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-nvkhs" Feb 27 08:13:25 crc kubenswrapper[4612]: I0227 08:13:25.188418 4612 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="0da436b1-803d-41bb-94ab-9cfd6fa769eb" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.100:5671: connect: connection refused" Feb 27 08:13:25 crc kubenswrapper[4612]: I0227 08:13:25.321280 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-nvkhs" Feb 27 08:13:25 crc kubenswrapper[4612]: I0227 08:13:25.508071 4612 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="6d1ba6db-1cae-461e-9716-5d5b36350218" containerName="rabbitmq" containerID="cri-o://d4bbac865331908abcaa583aff895c0dfe51bae301a0ff6fa4b2062f92b46c45" gracePeriod=604795 Feb 27 08:13:25 crc kubenswrapper[4612]: I0227 08:13:25.982546 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-nvkhs"] Feb 27 08:13:26 crc kubenswrapper[4612]: I0227 08:13:26.169232 4612 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-qvxtt" podUID="1d73fa1d-60eb-4c32-a82d-96f26011b04d" containerName="registry-server" probeResult="failure" output=< Feb 27 08:13:26 crc kubenswrapper[4612]: timeout: failed to connect service ":50051" within 1s Feb 27 08:13:26 crc kubenswrapper[4612]: > Feb 27 08:13:27 crc kubenswrapper[4612]: I0227 08:13:27.251100 4612 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-nvkhs" podUID="3485c18d-1e56-4c62-b27e-609ee08ab06f" containerName="registry-server" containerID="cri-o://ed88b222ece5c62b9fe1d6a3abb48a5ae8499490751152a7f44fe3b59b44b23e" gracePeriod=2 Feb 27 08:13:27 crc kubenswrapper[4612]: I0227 08:13:27.839962 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nvkhs" Feb 27 08:13:27 crc kubenswrapper[4612]: I0227 08:13:27.990148 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3485c18d-1e56-4c62-b27e-609ee08ab06f-catalog-content\") pod \"3485c18d-1e56-4c62-b27e-609ee08ab06f\" (UID: \"3485c18d-1e56-4c62-b27e-609ee08ab06f\") " Feb 27 08:13:27 crc kubenswrapper[4612]: I0227 08:13:27.990281 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vmj7v\" (UniqueName: \"kubernetes.io/projected/3485c18d-1e56-4c62-b27e-609ee08ab06f-kube-api-access-vmj7v\") pod \"3485c18d-1e56-4c62-b27e-609ee08ab06f\" (UID: \"3485c18d-1e56-4c62-b27e-609ee08ab06f\") " Feb 27 08:13:27 crc kubenswrapper[4612]: I0227 08:13:27.990399 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3485c18d-1e56-4c62-b27e-609ee08ab06f-utilities\") pod \"3485c18d-1e56-4c62-b27e-609ee08ab06f\" (UID: \"3485c18d-1e56-4c62-b27e-609ee08ab06f\") " Feb 27 08:13:27 crc kubenswrapper[4612]: I0227 08:13:27.991197 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3485c18d-1e56-4c62-b27e-609ee08ab06f-utilities" (OuterVolumeSpecName: "utilities") pod "3485c18d-1e56-4c62-b27e-609ee08ab06f" (UID: "3485c18d-1e56-4c62-b27e-609ee08ab06f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 08:13:27 crc kubenswrapper[4612]: I0227 08:13:27.998857 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3485c18d-1e56-4c62-b27e-609ee08ab06f-kube-api-access-vmj7v" (OuterVolumeSpecName: "kube-api-access-vmj7v") pod "3485c18d-1e56-4c62-b27e-609ee08ab06f" (UID: "3485c18d-1e56-4c62-b27e-609ee08ab06f"). InnerVolumeSpecName "kube-api-access-vmj7v". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 08:13:28 crc kubenswrapper[4612]: I0227 08:13:28.038563 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3485c18d-1e56-4c62-b27e-609ee08ab06f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3485c18d-1e56-4c62-b27e-609ee08ab06f" (UID: "3485c18d-1e56-4c62-b27e-609ee08ab06f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 08:13:28 crc kubenswrapper[4612]: I0227 08:13:28.092399 4612 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3485c18d-1e56-4c62-b27e-609ee08ab06f-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 27 08:13:28 crc kubenswrapper[4612]: I0227 08:13:28.092643 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vmj7v\" (UniqueName: \"kubernetes.io/projected/3485c18d-1e56-4c62-b27e-609ee08ab06f-kube-api-access-vmj7v\") on node \"crc\" DevicePath \"\"" Feb 27 08:13:28 crc kubenswrapper[4612]: I0227 08:13:28.092654 4612 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3485c18d-1e56-4c62-b27e-609ee08ab06f-utilities\") on node \"crc\" DevicePath \"\"" Feb 27 08:13:28 crc kubenswrapper[4612]: I0227 08:13:28.263212 4612 generic.go:334] "Generic (PLEG): container finished" podID="3485c18d-1e56-4c62-b27e-609ee08ab06f" containerID="ed88b222ece5c62b9fe1d6a3abb48a5ae8499490751152a7f44fe3b59b44b23e" exitCode=0 Feb 27 08:13:28 crc kubenswrapper[4612]: I0227 08:13:28.263283 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nvkhs" event={"ID":"3485c18d-1e56-4c62-b27e-609ee08ab06f","Type":"ContainerDied","Data":"ed88b222ece5c62b9fe1d6a3abb48a5ae8499490751152a7f44fe3b59b44b23e"} Feb 27 08:13:28 crc kubenswrapper[4612]: I0227 08:13:28.263316 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nvkhs" Feb 27 08:13:28 crc kubenswrapper[4612]: I0227 08:13:28.263334 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nvkhs" event={"ID":"3485c18d-1e56-4c62-b27e-609ee08ab06f","Type":"ContainerDied","Data":"e7154d0ed03bfa21ed4df7ef658ab204bfdebb7248b134369ed681ba77950dd3"} Feb 27 08:13:28 crc kubenswrapper[4612]: I0227 08:13:28.263366 4612 scope.go:117] "RemoveContainer" containerID="ed88b222ece5c62b9fe1d6a3abb48a5ae8499490751152a7f44fe3b59b44b23e" Feb 27 08:13:28 crc kubenswrapper[4612]: I0227 08:13:28.293572 4612 scope.go:117] "RemoveContainer" containerID="08e8efea5e358f6e6010676ad3d80b7e2d0ece31f38fa9eb5f45ad3728fa44f5" Feb 27 08:13:28 crc kubenswrapper[4612]: I0227 08:13:28.312962 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-nvkhs"] Feb 27 08:13:28 crc kubenswrapper[4612]: I0227 08:13:28.324823 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-nvkhs"] Feb 27 08:13:28 crc kubenswrapper[4612]: I0227 08:13:28.326796 4612 scope.go:117] "RemoveContainer" containerID="57694b1cb9056ecc305864d2b80b28321b96260b03532d2f92d6308a31d91632" Feb 27 08:13:28 crc kubenswrapper[4612]: I0227 08:13:28.372208 4612 scope.go:117] "RemoveContainer" containerID="ed88b222ece5c62b9fe1d6a3abb48a5ae8499490751152a7f44fe3b59b44b23e" Feb 27 08:13:28 crc kubenswrapper[4612]: E0227 08:13:28.372661 4612 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ed88b222ece5c62b9fe1d6a3abb48a5ae8499490751152a7f44fe3b59b44b23e\": container with ID starting with ed88b222ece5c62b9fe1d6a3abb48a5ae8499490751152a7f44fe3b59b44b23e not found: ID does not exist" containerID="ed88b222ece5c62b9fe1d6a3abb48a5ae8499490751152a7f44fe3b59b44b23e" Feb 27 08:13:28 crc kubenswrapper[4612]: I0227 08:13:28.372716 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ed88b222ece5c62b9fe1d6a3abb48a5ae8499490751152a7f44fe3b59b44b23e"} err="failed to get container status \"ed88b222ece5c62b9fe1d6a3abb48a5ae8499490751152a7f44fe3b59b44b23e\": rpc error: code = NotFound desc = could not find container \"ed88b222ece5c62b9fe1d6a3abb48a5ae8499490751152a7f44fe3b59b44b23e\": container with ID starting with ed88b222ece5c62b9fe1d6a3abb48a5ae8499490751152a7f44fe3b59b44b23e not found: ID does not exist" Feb 27 08:13:28 crc kubenswrapper[4612]: I0227 08:13:28.372752 4612 scope.go:117] "RemoveContainer" containerID="08e8efea5e358f6e6010676ad3d80b7e2d0ece31f38fa9eb5f45ad3728fa44f5" Feb 27 08:13:28 crc kubenswrapper[4612]: E0227 08:13:28.373053 4612 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"08e8efea5e358f6e6010676ad3d80b7e2d0ece31f38fa9eb5f45ad3728fa44f5\": container with ID starting with 08e8efea5e358f6e6010676ad3d80b7e2d0ece31f38fa9eb5f45ad3728fa44f5 not found: ID does not exist" containerID="08e8efea5e358f6e6010676ad3d80b7e2d0ece31f38fa9eb5f45ad3728fa44f5" Feb 27 08:13:28 crc kubenswrapper[4612]: I0227 08:13:28.373114 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"08e8efea5e358f6e6010676ad3d80b7e2d0ece31f38fa9eb5f45ad3728fa44f5"} err="failed to get container status \"08e8efea5e358f6e6010676ad3d80b7e2d0ece31f38fa9eb5f45ad3728fa44f5\": rpc error: code = NotFound desc = could not find container \"08e8efea5e358f6e6010676ad3d80b7e2d0ece31f38fa9eb5f45ad3728fa44f5\": container with ID starting with 08e8efea5e358f6e6010676ad3d80b7e2d0ece31f38fa9eb5f45ad3728fa44f5 not found: ID does not exist" Feb 27 08:13:28 crc kubenswrapper[4612]: I0227 08:13:28.373140 4612 scope.go:117] "RemoveContainer" containerID="57694b1cb9056ecc305864d2b80b28321b96260b03532d2f92d6308a31d91632" Feb 27 08:13:28 crc kubenswrapper[4612]: E0227 08:13:28.373393 4612 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"57694b1cb9056ecc305864d2b80b28321b96260b03532d2f92d6308a31d91632\": container with ID starting with 57694b1cb9056ecc305864d2b80b28321b96260b03532d2f92d6308a31d91632 not found: ID does not exist" containerID="57694b1cb9056ecc305864d2b80b28321b96260b03532d2f92d6308a31d91632" Feb 27 08:13:28 crc kubenswrapper[4612]: I0227 08:13:28.373413 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"57694b1cb9056ecc305864d2b80b28321b96260b03532d2f92d6308a31d91632"} err="failed to get container status \"57694b1cb9056ecc305864d2b80b28321b96260b03532d2f92d6308a31d91632\": rpc error: code = NotFound desc = could not find container \"57694b1cb9056ecc305864d2b80b28321b96260b03532d2f92d6308a31d91632\": container with ID starting with 57694b1cb9056ecc305864d2b80b28321b96260b03532d2f92d6308a31d91632 not found: ID does not exist" Feb 27 08:13:28 crc kubenswrapper[4612]: I0227 08:13:28.863302 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3485c18d-1e56-4c62-b27e-609ee08ab06f" path="/var/lib/kubelet/pods/3485c18d-1e56-4c62-b27e-609ee08ab06f/volumes" Feb 27 08:13:30 crc kubenswrapper[4612]: I0227 08:13:30.287112 4612 generic.go:334] "Generic (PLEG): container finished" podID="0da436b1-803d-41bb-94ab-9cfd6fa769eb" containerID="edd43cdd132a714448da88e522b53d539c8bad75e13f40e8d8ad5319c1936c48" exitCode=0 Feb 27 08:13:30 crc kubenswrapper[4612]: I0227 08:13:30.288440 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"0da436b1-803d-41bb-94ab-9cfd6fa769eb","Type":"ContainerDied","Data":"edd43cdd132a714448da88e522b53d539c8bad75e13f40e8d8ad5319c1936c48"} Feb 27 08:13:30 crc kubenswrapper[4612]: I0227 08:13:30.288573 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"0da436b1-803d-41bb-94ab-9cfd6fa769eb","Type":"ContainerDied","Data":"b0d2efa8e535b610aa785074d361a1c941c0b5146f868873b1ed12043662d340"} Feb 27 08:13:30 crc kubenswrapper[4612]: I0227 08:13:30.288646 4612 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b0d2efa8e535b610aa785074d361a1c941c0b5146f868873b1ed12043662d340" Feb 27 08:13:30 crc kubenswrapper[4612]: I0227 08:13:30.385854 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Feb 27 08:13:30 crc kubenswrapper[4612]: I0227 08:13:30.557286 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/0da436b1-803d-41bb-94ab-9cfd6fa769eb-rabbitmq-plugins\") pod \"0da436b1-803d-41bb-94ab-9cfd6fa769eb\" (UID: \"0da436b1-803d-41bb-94ab-9cfd6fa769eb\") " Feb 27 08:13:30 crc kubenswrapper[4612]: I0227 08:13:30.557671 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/0da436b1-803d-41bb-94ab-9cfd6fa769eb-plugins-conf\") pod \"0da436b1-803d-41bb-94ab-9cfd6fa769eb\" (UID: \"0da436b1-803d-41bb-94ab-9cfd6fa769eb\") " Feb 27 08:13:30 crc kubenswrapper[4612]: I0227 08:13:30.557781 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/0da436b1-803d-41bb-94ab-9cfd6fa769eb-server-conf\") pod \"0da436b1-803d-41bb-94ab-9cfd6fa769eb\" (UID: \"0da436b1-803d-41bb-94ab-9cfd6fa769eb\") " Feb 27 08:13:30 crc kubenswrapper[4612]: I0227 08:13:30.557833 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/0da436b1-803d-41bb-94ab-9cfd6fa769eb-rabbitmq-tls\") pod \"0da436b1-803d-41bb-94ab-9cfd6fa769eb\" (UID: \"0da436b1-803d-41bb-94ab-9cfd6fa769eb\") " Feb 27 08:13:30 crc kubenswrapper[4612]: I0227 08:13:30.557897 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0da436b1-803d-41bb-94ab-9cfd6fa769eb-config-data\") pod \"0da436b1-803d-41bb-94ab-9cfd6fa769eb\" (UID: \"0da436b1-803d-41bb-94ab-9cfd6fa769eb\") " Feb 27 08:13:30 crc kubenswrapper[4612]: I0227 08:13:30.557929 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/0da436b1-803d-41bb-94ab-9cfd6fa769eb-rabbitmq-confd\") pod \"0da436b1-803d-41bb-94ab-9cfd6fa769eb\" (UID: \"0da436b1-803d-41bb-94ab-9cfd6fa769eb\") " Feb 27 08:13:30 crc kubenswrapper[4612]: I0227 08:13:30.557973 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vjjpx\" (UniqueName: \"kubernetes.io/projected/0da436b1-803d-41bb-94ab-9cfd6fa769eb-kube-api-access-vjjpx\") pod \"0da436b1-803d-41bb-94ab-9cfd6fa769eb\" (UID: \"0da436b1-803d-41bb-94ab-9cfd6fa769eb\") " Feb 27 08:13:30 crc kubenswrapper[4612]: I0227 08:13:30.558003 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/0da436b1-803d-41bb-94ab-9cfd6fa769eb-pod-info\") pod \"0da436b1-803d-41bb-94ab-9cfd6fa769eb\" (UID: \"0da436b1-803d-41bb-94ab-9cfd6fa769eb\") " Feb 27 08:13:30 crc kubenswrapper[4612]: I0227 08:13:30.558032 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"0da436b1-803d-41bb-94ab-9cfd6fa769eb\" (UID: \"0da436b1-803d-41bb-94ab-9cfd6fa769eb\") " Feb 27 08:13:30 crc kubenswrapper[4612]: I0227 08:13:30.558092 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/0da436b1-803d-41bb-94ab-9cfd6fa769eb-erlang-cookie-secret\") pod \"0da436b1-803d-41bb-94ab-9cfd6fa769eb\" (UID: \"0da436b1-803d-41bb-94ab-9cfd6fa769eb\") " Feb 27 08:13:30 crc kubenswrapper[4612]: I0227 08:13:30.558139 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/0da436b1-803d-41bb-94ab-9cfd6fa769eb-rabbitmq-erlang-cookie\") pod \"0da436b1-803d-41bb-94ab-9cfd6fa769eb\" (UID: \"0da436b1-803d-41bb-94ab-9cfd6fa769eb\") " Feb 27 08:13:30 crc kubenswrapper[4612]: I0227 08:13:30.560384 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0da436b1-803d-41bb-94ab-9cfd6fa769eb-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "0da436b1-803d-41bb-94ab-9cfd6fa769eb" (UID: "0da436b1-803d-41bb-94ab-9cfd6fa769eb"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 08:13:30 crc kubenswrapper[4612]: I0227 08:13:30.574456 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0da436b1-803d-41bb-94ab-9cfd6fa769eb-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "0da436b1-803d-41bb-94ab-9cfd6fa769eb" (UID: "0da436b1-803d-41bb-94ab-9cfd6fa769eb"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 08:13:30 crc kubenswrapper[4612]: I0227 08:13:30.579988 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0da436b1-803d-41bb-94ab-9cfd6fa769eb-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "0da436b1-803d-41bb-94ab-9cfd6fa769eb" (UID: "0da436b1-803d-41bb-94ab-9cfd6fa769eb"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 08:13:30 crc kubenswrapper[4612]: I0227 08:13:30.589373 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0da436b1-803d-41bb-94ab-9cfd6fa769eb-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "0da436b1-803d-41bb-94ab-9cfd6fa769eb" (UID: "0da436b1-803d-41bb-94ab-9cfd6fa769eb"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 08:13:30 crc kubenswrapper[4612]: I0227 08:13:30.591588 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage02-crc" (OuterVolumeSpecName: "persistence") pod "0da436b1-803d-41bb-94ab-9cfd6fa769eb" (UID: "0da436b1-803d-41bb-94ab-9cfd6fa769eb"). InnerVolumeSpecName "local-storage02-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Feb 27 08:13:30 crc kubenswrapper[4612]: I0227 08:13:30.598922 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/0da436b1-803d-41bb-94ab-9cfd6fa769eb-pod-info" (OuterVolumeSpecName: "pod-info") pod "0da436b1-803d-41bb-94ab-9cfd6fa769eb" (UID: "0da436b1-803d-41bb-94ab-9cfd6fa769eb"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Feb 27 08:13:30 crc kubenswrapper[4612]: I0227 08:13:30.603052 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0da436b1-803d-41bb-94ab-9cfd6fa769eb-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "0da436b1-803d-41bb-94ab-9cfd6fa769eb" (UID: "0da436b1-803d-41bb-94ab-9cfd6fa769eb"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:13:30 crc kubenswrapper[4612]: I0227 08:13:30.603682 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0da436b1-803d-41bb-94ab-9cfd6fa769eb-kube-api-access-vjjpx" (OuterVolumeSpecName: "kube-api-access-vjjpx") pod "0da436b1-803d-41bb-94ab-9cfd6fa769eb" (UID: "0da436b1-803d-41bb-94ab-9cfd6fa769eb"). InnerVolumeSpecName "kube-api-access-vjjpx". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 08:13:30 crc kubenswrapper[4612]: I0227 08:13:30.637110 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0da436b1-803d-41bb-94ab-9cfd6fa769eb-config-data" (OuterVolumeSpecName: "config-data") pod "0da436b1-803d-41bb-94ab-9cfd6fa769eb" (UID: "0da436b1-803d-41bb-94ab-9cfd6fa769eb"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 08:13:30 crc kubenswrapper[4612]: I0227 08:13:30.660731 4612 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/0da436b1-803d-41bb-94ab-9cfd6fa769eb-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Feb 27 08:13:30 crc kubenswrapper[4612]: I0227 08:13:30.660767 4612 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/0da436b1-803d-41bb-94ab-9cfd6fa769eb-plugins-conf\") on node \"crc\" DevicePath \"\"" Feb 27 08:13:30 crc kubenswrapper[4612]: I0227 08:13:30.660779 4612 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/0da436b1-803d-41bb-94ab-9cfd6fa769eb-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Feb 27 08:13:30 crc kubenswrapper[4612]: I0227 08:13:30.660791 4612 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0da436b1-803d-41bb-94ab-9cfd6fa769eb-config-data\") on node \"crc\" DevicePath \"\"" Feb 27 08:13:30 crc kubenswrapper[4612]: I0227 08:13:30.660803 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vjjpx\" (UniqueName: \"kubernetes.io/projected/0da436b1-803d-41bb-94ab-9cfd6fa769eb-kube-api-access-vjjpx\") on node \"crc\" DevicePath \"\"" Feb 27 08:13:30 crc kubenswrapper[4612]: I0227 08:13:30.660817 4612 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/0da436b1-803d-41bb-94ab-9cfd6fa769eb-pod-info\") on node \"crc\" DevicePath \"\"" Feb 27 08:13:30 crc kubenswrapper[4612]: I0227 08:13:30.660845 4612 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" " Feb 27 08:13:30 crc kubenswrapper[4612]: I0227 08:13:30.660858 4612 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/0da436b1-803d-41bb-94ab-9cfd6fa769eb-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Feb 27 08:13:30 crc kubenswrapper[4612]: I0227 08:13:30.660870 4612 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/0da436b1-803d-41bb-94ab-9cfd6fa769eb-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Feb 27 08:13:30 crc kubenswrapper[4612]: I0227 08:13:30.677274 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0da436b1-803d-41bb-94ab-9cfd6fa769eb-server-conf" (OuterVolumeSpecName: "server-conf") pod "0da436b1-803d-41bb-94ab-9cfd6fa769eb" (UID: "0da436b1-803d-41bb-94ab-9cfd6fa769eb"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 08:13:30 crc kubenswrapper[4612]: I0227 08:13:30.695045 4612 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage02-crc" (UniqueName: "kubernetes.io/local-volume/local-storage02-crc") on node "crc" Feb 27 08:13:30 crc kubenswrapper[4612]: I0227 08:13:30.762549 4612 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/0da436b1-803d-41bb-94ab-9cfd6fa769eb-server-conf\") on node \"crc\" DevicePath \"\"" Feb 27 08:13:30 crc kubenswrapper[4612]: I0227 08:13:30.762581 4612 reconciler_common.go:293] "Volume detached for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" DevicePath \"\"" Feb 27 08:13:30 crc kubenswrapper[4612]: I0227 08:13:30.779437 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0da436b1-803d-41bb-94ab-9cfd6fa769eb-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "0da436b1-803d-41bb-94ab-9cfd6fa769eb" (UID: "0da436b1-803d-41bb-94ab-9cfd6fa769eb"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 08:13:30 crc kubenswrapper[4612]: I0227 08:13:30.863821 4612 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/0da436b1-803d-41bb-94ab-9cfd6fa769eb-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Feb 27 08:13:31 crc kubenswrapper[4612]: I0227 08:13:31.296449 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Feb 27 08:13:31 crc kubenswrapper[4612]: I0227 08:13:31.329825 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Feb 27 08:13:31 crc kubenswrapper[4612]: I0227 08:13:31.345433 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Feb 27 08:13:31 crc kubenswrapper[4612]: I0227 08:13:31.364522 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Feb 27 08:13:31 crc kubenswrapper[4612]: E0227 08:13:31.365068 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3485c18d-1e56-4c62-b27e-609ee08ab06f" containerName="extract-content" Feb 27 08:13:31 crc kubenswrapper[4612]: I0227 08:13:31.365093 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="3485c18d-1e56-4c62-b27e-609ee08ab06f" containerName="extract-content" Feb 27 08:13:31 crc kubenswrapper[4612]: E0227 08:13:31.365123 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0da436b1-803d-41bb-94ab-9cfd6fa769eb" containerName="rabbitmq" Feb 27 08:13:31 crc kubenswrapper[4612]: I0227 08:13:31.365132 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="0da436b1-803d-41bb-94ab-9cfd6fa769eb" containerName="rabbitmq" Feb 27 08:13:31 crc kubenswrapper[4612]: E0227 08:13:31.365158 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0da436b1-803d-41bb-94ab-9cfd6fa769eb" containerName="setup-container" Feb 27 08:13:31 crc kubenswrapper[4612]: I0227 08:13:31.365165 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="0da436b1-803d-41bb-94ab-9cfd6fa769eb" containerName="setup-container" Feb 27 08:13:31 crc kubenswrapper[4612]: E0227 08:13:31.365175 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3485c18d-1e56-4c62-b27e-609ee08ab06f" containerName="registry-server" Feb 27 08:13:31 crc kubenswrapper[4612]: I0227 08:13:31.365182 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="3485c18d-1e56-4c62-b27e-609ee08ab06f" containerName="registry-server" Feb 27 08:13:31 crc kubenswrapper[4612]: E0227 08:13:31.365192 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3485c18d-1e56-4c62-b27e-609ee08ab06f" containerName="extract-utilities" Feb 27 08:13:31 crc kubenswrapper[4612]: I0227 08:13:31.365201 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="3485c18d-1e56-4c62-b27e-609ee08ab06f" containerName="extract-utilities" Feb 27 08:13:31 crc kubenswrapper[4612]: I0227 08:13:31.365431 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="3485c18d-1e56-4c62-b27e-609ee08ab06f" containerName="registry-server" Feb 27 08:13:31 crc kubenswrapper[4612]: I0227 08:13:31.365455 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="0da436b1-803d-41bb-94ab-9cfd6fa769eb" containerName="rabbitmq" Feb 27 08:13:31 crc kubenswrapper[4612]: I0227 08:13:31.366635 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Feb 27 08:13:31 crc kubenswrapper[4612]: I0227 08:13:31.371403 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Feb 27 08:13:31 crc kubenswrapper[4612]: I0227 08:13:31.371449 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Feb 27 08:13:31 crc kubenswrapper[4612]: I0227 08:13:31.371704 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Feb 27 08:13:31 crc kubenswrapper[4612]: I0227 08:13:31.371828 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Feb 27 08:13:31 crc kubenswrapper[4612]: I0227 08:13:31.372046 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-r2sww" Feb 27 08:13:31 crc kubenswrapper[4612]: I0227 08:13:31.372090 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Feb 27 08:13:31 crc kubenswrapper[4612]: I0227 08:13:31.374470 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Feb 27 08:13:31 crc kubenswrapper[4612]: I0227 08:13:31.390879 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Feb 27 08:13:31 crc kubenswrapper[4612]: I0227 08:13:31.515228 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/ba8e827f-294b-4532-8636-c81eeb511ba4-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"ba8e827f-294b-4532-8636-c81eeb511ba4\") " pod="openstack/rabbitmq-server-0" Feb 27 08:13:31 crc kubenswrapper[4612]: I0227 08:13:31.515283 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/ba8e827f-294b-4532-8636-c81eeb511ba4-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"ba8e827f-294b-4532-8636-c81eeb511ba4\") " pod="openstack/rabbitmq-server-0" Feb 27 08:13:31 crc kubenswrapper[4612]: I0227 08:13:31.515331 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/ba8e827f-294b-4532-8636-c81eeb511ba4-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"ba8e827f-294b-4532-8636-c81eeb511ba4\") " pod="openstack/rabbitmq-server-0" Feb 27 08:13:31 crc kubenswrapper[4612]: I0227 08:13:31.515360 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xg9rh\" (UniqueName: \"kubernetes.io/projected/ba8e827f-294b-4532-8636-c81eeb511ba4-kube-api-access-xg9rh\") pod \"rabbitmq-server-0\" (UID: \"ba8e827f-294b-4532-8636-c81eeb511ba4\") " pod="openstack/rabbitmq-server-0" Feb 27 08:13:31 crc kubenswrapper[4612]: I0227 08:13:31.515397 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ba8e827f-294b-4532-8636-c81eeb511ba4-config-data\") pod \"rabbitmq-server-0\" (UID: \"ba8e827f-294b-4532-8636-c81eeb511ba4\") " pod="openstack/rabbitmq-server-0" Feb 27 08:13:31 crc kubenswrapper[4612]: I0227 08:13:31.515413 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/ba8e827f-294b-4532-8636-c81eeb511ba4-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"ba8e827f-294b-4532-8636-c81eeb511ba4\") " pod="openstack/rabbitmq-server-0" Feb 27 08:13:31 crc kubenswrapper[4612]: I0227 08:13:31.515427 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/ba8e827f-294b-4532-8636-c81eeb511ba4-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"ba8e827f-294b-4532-8636-c81eeb511ba4\") " pod="openstack/rabbitmq-server-0" Feb 27 08:13:31 crc kubenswrapper[4612]: I0227 08:13:31.515449 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/ba8e827f-294b-4532-8636-c81eeb511ba4-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"ba8e827f-294b-4532-8636-c81eeb511ba4\") " pod="openstack/rabbitmq-server-0" Feb 27 08:13:31 crc kubenswrapper[4612]: I0227 08:13:31.515479 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/ba8e827f-294b-4532-8636-c81eeb511ba4-pod-info\") pod \"rabbitmq-server-0\" (UID: \"ba8e827f-294b-4532-8636-c81eeb511ba4\") " pod="openstack/rabbitmq-server-0" Feb 27 08:13:31 crc kubenswrapper[4612]: I0227 08:13:31.515509 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-server-0\" (UID: \"ba8e827f-294b-4532-8636-c81eeb511ba4\") " pod="openstack/rabbitmq-server-0" Feb 27 08:13:31 crc kubenswrapper[4612]: I0227 08:13:31.515535 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/ba8e827f-294b-4532-8636-c81eeb511ba4-server-conf\") pod \"rabbitmq-server-0\" (UID: \"ba8e827f-294b-4532-8636-c81eeb511ba4\") " pod="openstack/rabbitmq-server-0" Feb 27 08:13:31 crc kubenswrapper[4612]: I0227 08:13:31.617518 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/ba8e827f-294b-4532-8636-c81eeb511ba4-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"ba8e827f-294b-4532-8636-c81eeb511ba4\") " pod="openstack/rabbitmq-server-0" Feb 27 08:13:31 crc kubenswrapper[4612]: I0227 08:13:31.617603 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/ba8e827f-294b-4532-8636-c81eeb511ba4-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"ba8e827f-294b-4532-8636-c81eeb511ba4\") " pod="openstack/rabbitmq-server-0" Feb 27 08:13:31 crc kubenswrapper[4612]: I0227 08:13:31.617641 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xg9rh\" (UniqueName: \"kubernetes.io/projected/ba8e827f-294b-4532-8636-c81eeb511ba4-kube-api-access-xg9rh\") pod \"rabbitmq-server-0\" (UID: \"ba8e827f-294b-4532-8636-c81eeb511ba4\") " pod="openstack/rabbitmq-server-0" Feb 27 08:13:31 crc kubenswrapper[4612]: I0227 08:13:31.617678 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ba8e827f-294b-4532-8636-c81eeb511ba4-config-data\") pod \"rabbitmq-server-0\" (UID: \"ba8e827f-294b-4532-8636-c81eeb511ba4\") " pod="openstack/rabbitmq-server-0" Feb 27 08:13:31 crc kubenswrapper[4612]: I0227 08:13:31.617716 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/ba8e827f-294b-4532-8636-c81eeb511ba4-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"ba8e827f-294b-4532-8636-c81eeb511ba4\") " pod="openstack/rabbitmq-server-0" Feb 27 08:13:31 crc kubenswrapper[4612]: I0227 08:13:31.617740 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/ba8e827f-294b-4532-8636-c81eeb511ba4-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"ba8e827f-294b-4532-8636-c81eeb511ba4\") " pod="openstack/rabbitmq-server-0" Feb 27 08:13:31 crc kubenswrapper[4612]: I0227 08:13:31.617770 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/ba8e827f-294b-4532-8636-c81eeb511ba4-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"ba8e827f-294b-4532-8636-c81eeb511ba4\") " pod="openstack/rabbitmq-server-0" Feb 27 08:13:31 crc kubenswrapper[4612]: I0227 08:13:31.617809 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/ba8e827f-294b-4532-8636-c81eeb511ba4-pod-info\") pod \"rabbitmq-server-0\" (UID: \"ba8e827f-294b-4532-8636-c81eeb511ba4\") " pod="openstack/rabbitmq-server-0" Feb 27 08:13:31 crc kubenswrapper[4612]: I0227 08:13:31.617854 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-server-0\" (UID: \"ba8e827f-294b-4532-8636-c81eeb511ba4\") " pod="openstack/rabbitmq-server-0" Feb 27 08:13:31 crc kubenswrapper[4612]: I0227 08:13:31.617885 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/ba8e827f-294b-4532-8636-c81eeb511ba4-server-conf\") pod \"rabbitmq-server-0\" (UID: \"ba8e827f-294b-4532-8636-c81eeb511ba4\") " pod="openstack/rabbitmq-server-0" Feb 27 08:13:31 crc kubenswrapper[4612]: I0227 08:13:31.617947 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/ba8e827f-294b-4532-8636-c81eeb511ba4-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"ba8e827f-294b-4532-8636-c81eeb511ba4\") " pod="openstack/rabbitmq-server-0" Feb 27 08:13:31 crc kubenswrapper[4612]: I0227 08:13:31.618319 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/ba8e827f-294b-4532-8636-c81eeb511ba4-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"ba8e827f-294b-4532-8636-c81eeb511ba4\") " pod="openstack/rabbitmq-server-0" Feb 27 08:13:31 crc kubenswrapper[4612]: I0227 08:13:31.618517 4612 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-server-0\" (UID: \"ba8e827f-294b-4532-8636-c81eeb511ba4\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/rabbitmq-server-0" Feb 27 08:13:31 crc kubenswrapper[4612]: I0227 08:13:31.618747 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ba8e827f-294b-4532-8636-c81eeb511ba4-config-data\") pod \"rabbitmq-server-0\" (UID: \"ba8e827f-294b-4532-8636-c81eeb511ba4\") " pod="openstack/rabbitmq-server-0" Feb 27 08:13:31 crc kubenswrapper[4612]: I0227 08:13:31.618964 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/ba8e827f-294b-4532-8636-c81eeb511ba4-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"ba8e827f-294b-4532-8636-c81eeb511ba4\") " pod="openstack/rabbitmq-server-0" Feb 27 08:13:31 crc kubenswrapper[4612]: I0227 08:13:31.619017 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/ba8e827f-294b-4532-8636-c81eeb511ba4-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"ba8e827f-294b-4532-8636-c81eeb511ba4\") " pod="openstack/rabbitmq-server-0" Feb 27 08:13:31 crc kubenswrapper[4612]: I0227 08:13:31.619626 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/ba8e827f-294b-4532-8636-c81eeb511ba4-server-conf\") pod \"rabbitmq-server-0\" (UID: \"ba8e827f-294b-4532-8636-c81eeb511ba4\") " pod="openstack/rabbitmq-server-0" Feb 27 08:13:31 crc kubenswrapper[4612]: I0227 08:13:31.630477 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/ba8e827f-294b-4532-8636-c81eeb511ba4-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"ba8e827f-294b-4532-8636-c81eeb511ba4\") " pod="openstack/rabbitmq-server-0" Feb 27 08:13:31 crc kubenswrapper[4612]: I0227 08:13:31.630638 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/ba8e827f-294b-4532-8636-c81eeb511ba4-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"ba8e827f-294b-4532-8636-c81eeb511ba4\") " pod="openstack/rabbitmq-server-0" Feb 27 08:13:31 crc kubenswrapper[4612]: I0227 08:13:31.630889 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/ba8e827f-294b-4532-8636-c81eeb511ba4-pod-info\") pod \"rabbitmq-server-0\" (UID: \"ba8e827f-294b-4532-8636-c81eeb511ba4\") " pod="openstack/rabbitmq-server-0" Feb 27 08:13:31 crc kubenswrapper[4612]: I0227 08:13:31.640089 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/ba8e827f-294b-4532-8636-c81eeb511ba4-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"ba8e827f-294b-4532-8636-c81eeb511ba4\") " pod="openstack/rabbitmq-server-0" Feb 27 08:13:31 crc kubenswrapper[4612]: I0227 08:13:31.643527 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xg9rh\" (UniqueName: \"kubernetes.io/projected/ba8e827f-294b-4532-8636-c81eeb511ba4-kube-api-access-xg9rh\") pod \"rabbitmq-server-0\" (UID: \"ba8e827f-294b-4532-8636-c81eeb511ba4\") " pod="openstack/rabbitmq-server-0" Feb 27 08:13:31 crc kubenswrapper[4612]: I0227 08:13:31.667946 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-server-0\" (UID: \"ba8e827f-294b-4532-8636-c81eeb511ba4\") " pod="openstack/rabbitmq-server-0" Feb 27 08:13:31 crc kubenswrapper[4612]: I0227 08:13:31.692310 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Feb 27 08:13:32 crc kubenswrapper[4612]: I0227 08:13:32.044723 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Feb 27 08:13:32 crc kubenswrapper[4612]: I0227 08:13:32.313684 4612 generic.go:334] "Generic (PLEG): container finished" podID="6d1ba6db-1cae-461e-9716-5d5b36350218" containerID="d4bbac865331908abcaa583aff895c0dfe51bae301a0ff6fa4b2062f92b46c45" exitCode=0 Feb 27 08:13:32 crc kubenswrapper[4612]: I0227 08:13:32.313732 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"6d1ba6db-1cae-461e-9716-5d5b36350218","Type":"ContainerDied","Data":"d4bbac865331908abcaa583aff895c0dfe51bae301a0ff6fa4b2062f92b46c45"} Feb 27 08:13:32 crc kubenswrapper[4612]: I0227 08:13:32.314175 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"6d1ba6db-1cae-461e-9716-5d5b36350218","Type":"ContainerDied","Data":"152b8e50f185ca631e828e8d4886ab57c0abda78df5a4124b28847d2a6cd7111"} Feb 27 08:13:32 crc kubenswrapper[4612]: I0227 08:13:32.314191 4612 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="152b8e50f185ca631e828e8d4886ab57c0abda78df5a4124b28847d2a6cd7111" Feb 27 08:13:32 crc kubenswrapper[4612]: I0227 08:13:32.316976 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"ba8e827f-294b-4532-8636-c81eeb511ba4","Type":"ContainerStarted","Data":"d6d0f4188352e3f8f18108d5ce88a6b6208f3ae45076e43f5d7ab9ac1d243da3"} Feb 27 08:13:32 crc kubenswrapper[4612]: I0227 08:13:32.329047 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Feb 27 08:13:32 crc kubenswrapper[4612]: I0227 08:13:32.441742 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/6d1ba6db-1cae-461e-9716-5d5b36350218-plugins-conf\") pod \"6d1ba6db-1cae-461e-9716-5d5b36350218\" (UID: \"6d1ba6db-1cae-461e-9716-5d5b36350218\") " Feb 27 08:13:32 crc kubenswrapper[4612]: I0227 08:13:32.441806 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"6d1ba6db-1cae-461e-9716-5d5b36350218\" (UID: \"6d1ba6db-1cae-461e-9716-5d5b36350218\") " Feb 27 08:13:32 crc kubenswrapper[4612]: I0227 08:13:32.441862 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/6d1ba6db-1cae-461e-9716-5d5b36350218-rabbitmq-erlang-cookie\") pod \"6d1ba6db-1cae-461e-9716-5d5b36350218\" (UID: \"6d1ba6db-1cae-461e-9716-5d5b36350218\") " Feb 27 08:13:32 crc kubenswrapper[4612]: I0227 08:13:32.441911 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-86pqk\" (UniqueName: \"kubernetes.io/projected/6d1ba6db-1cae-461e-9716-5d5b36350218-kube-api-access-86pqk\") pod \"6d1ba6db-1cae-461e-9716-5d5b36350218\" (UID: \"6d1ba6db-1cae-461e-9716-5d5b36350218\") " Feb 27 08:13:32 crc kubenswrapper[4612]: I0227 08:13:32.441988 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/6d1ba6db-1cae-461e-9716-5d5b36350218-rabbitmq-confd\") pod \"6d1ba6db-1cae-461e-9716-5d5b36350218\" (UID: \"6d1ba6db-1cae-461e-9716-5d5b36350218\") " Feb 27 08:13:32 crc kubenswrapper[4612]: I0227 08:13:32.442006 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/6d1ba6db-1cae-461e-9716-5d5b36350218-server-conf\") pod \"6d1ba6db-1cae-461e-9716-5d5b36350218\" (UID: \"6d1ba6db-1cae-461e-9716-5d5b36350218\") " Feb 27 08:13:32 crc kubenswrapper[4612]: I0227 08:13:32.442084 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/6d1ba6db-1cae-461e-9716-5d5b36350218-rabbitmq-plugins\") pod \"6d1ba6db-1cae-461e-9716-5d5b36350218\" (UID: \"6d1ba6db-1cae-461e-9716-5d5b36350218\") " Feb 27 08:13:32 crc kubenswrapper[4612]: I0227 08:13:32.442109 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/6d1ba6db-1cae-461e-9716-5d5b36350218-pod-info\") pod \"6d1ba6db-1cae-461e-9716-5d5b36350218\" (UID: \"6d1ba6db-1cae-461e-9716-5d5b36350218\") " Feb 27 08:13:32 crc kubenswrapper[4612]: I0227 08:13:32.442127 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6d1ba6db-1cae-461e-9716-5d5b36350218-config-data\") pod \"6d1ba6db-1cae-461e-9716-5d5b36350218\" (UID: \"6d1ba6db-1cae-461e-9716-5d5b36350218\") " Feb 27 08:13:32 crc kubenswrapper[4612]: I0227 08:13:32.442148 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/6d1ba6db-1cae-461e-9716-5d5b36350218-rabbitmq-tls\") pod \"6d1ba6db-1cae-461e-9716-5d5b36350218\" (UID: \"6d1ba6db-1cae-461e-9716-5d5b36350218\") " Feb 27 08:13:32 crc kubenswrapper[4612]: I0227 08:13:32.442174 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/6d1ba6db-1cae-461e-9716-5d5b36350218-erlang-cookie-secret\") pod \"6d1ba6db-1cae-461e-9716-5d5b36350218\" (UID: \"6d1ba6db-1cae-461e-9716-5d5b36350218\") " Feb 27 08:13:32 crc kubenswrapper[4612]: I0227 08:13:32.443313 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6d1ba6db-1cae-461e-9716-5d5b36350218-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "6d1ba6db-1cae-461e-9716-5d5b36350218" (UID: "6d1ba6db-1cae-461e-9716-5d5b36350218"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 08:13:32 crc kubenswrapper[4612]: I0227 08:13:32.443389 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6d1ba6db-1cae-461e-9716-5d5b36350218-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "6d1ba6db-1cae-461e-9716-5d5b36350218" (UID: "6d1ba6db-1cae-461e-9716-5d5b36350218"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 08:13:32 crc kubenswrapper[4612]: I0227 08:13:32.443885 4612 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/6d1ba6db-1cae-461e-9716-5d5b36350218-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Feb 27 08:13:32 crc kubenswrapper[4612]: I0227 08:13:32.443901 4612 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/6d1ba6db-1cae-461e-9716-5d5b36350218-plugins-conf\") on node \"crc\" DevicePath \"\"" Feb 27 08:13:32 crc kubenswrapper[4612]: I0227 08:13:32.450953 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6d1ba6db-1cae-461e-9716-5d5b36350218-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "6d1ba6db-1cae-461e-9716-5d5b36350218" (UID: "6d1ba6db-1cae-461e-9716-5d5b36350218"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 08:13:32 crc kubenswrapper[4612]: I0227 08:13:32.462152 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6d1ba6db-1cae-461e-9716-5d5b36350218-kube-api-access-86pqk" (OuterVolumeSpecName: "kube-api-access-86pqk") pod "6d1ba6db-1cae-461e-9716-5d5b36350218" (UID: "6d1ba6db-1cae-461e-9716-5d5b36350218"). InnerVolumeSpecName "kube-api-access-86pqk". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 08:13:32 crc kubenswrapper[4612]: I0227 08:13:32.462190 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/6d1ba6db-1cae-461e-9716-5d5b36350218-pod-info" (OuterVolumeSpecName: "pod-info") pod "6d1ba6db-1cae-461e-9716-5d5b36350218" (UID: "6d1ba6db-1cae-461e-9716-5d5b36350218"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Feb 27 08:13:32 crc kubenswrapper[4612]: I0227 08:13:32.462270 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage10-crc" (OuterVolumeSpecName: "persistence") pod "6d1ba6db-1cae-461e-9716-5d5b36350218" (UID: "6d1ba6db-1cae-461e-9716-5d5b36350218"). InnerVolumeSpecName "local-storage10-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Feb 27 08:13:32 crc kubenswrapper[4612]: I0227 08:13:32.463088 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6d1ba6db-1cae-461e-9716-5d5b36350218-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "6d1ba6db-1cae-461e-9716-5d5b36350218" (UID: "6d1ba6db-1cae-461e-9716-5d5b36350218"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 08:13:32 crc kubenswrapper[4612]: I0227 08:13:32.466766 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6d1ba6db-1cae-461e-9716-5d5b36350218-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "6d1ba6db-1cae-461e-9716-5d5b36350218" (UID: "6d1ba6db-1cae-461e-9716-5d5b36350218"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:13:32 crc kubenswrapper[4612]: I0227 08:13:32.486631 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6d1ba6db-1cae-461e-9716-5d5b36350218-config-data" (OuterVolumeSpecName: "config-data") pod "6d1ba6db-1cae-461e-9716-5d5b36350218" (UID: "6d1ba6db-1cae-461e-9716-5d5b36350218"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 08:13:32 crc kubenswrapper[4612]: I0227 08:13:32.517144 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6d1ba6db-1cae-461e-9716-5d5b36350218-server-conf" (OuterVolumeSpecName: "server-conf") pod "6d1ba6db-1cae-461e-9716-5d5b36350218" (UID: "6d1ba6db-1cae-461e-9716-5d5b36350218"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 08:13:32 crc kubenswrapper[4612]: I0227 08:13:32.545126 4612 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/6d1ba6db-1cae-461e-9716-5d5b36350218-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Feb 27 08:13:32 crc kubenswrapper[4612]: I0227 08:13:32.545158 4612 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/6d1ba6db-1cae-461e-9716-5d5b36350218-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Feb 27 08:13:32 crc kubenswrapper[4612]: I0227 08:13:32.545187 4612 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" " Feb 27 08:13:32 crc kubenswrapper[4612]: I0227 08:13:32.545197 4612 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/6d1ba6db-1cae-461e-9716-5d5b36350218-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Feb 27 08:13:32 crc kubenswrapper[4612]: I0227 08:13:32.545207 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-86pqk\" (UniqueName: \"kubernetes.io/projected/6d1ba6db-1cae-461e-9716-5d5b36350218-kube-api-access-86pqk\") on node \"crc\" DevicePath \"\"" Feb 27 08:13:32 crc kubenswrapper[4612]: I0227 08:13:32.545215 4612 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/6d1ba6db-1cae-461e-9716-5d5b36350218-server-conf\") on node \"crc\" DevicePath \"\"" Feb 27 08:13:32 crc kubenswrapper[4612]: I0227 08:13:32.545225 4612 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/6d1ba6db-1cae-461e-9716-5d5b36350218-pod-info\") on node \"crc\" DevicePath \"\"" Feb 27 08:13:32 crc kubenswrapper[4612]: I0227 08:13:32.545233 4612 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6d1ba6db-1cae-461e-9716-5d5b36350218-config-data\") on node \"crc\" DevicePath \"\"" Feb 27 08:13:32 crc kubenswrapper[4612]: I0227 08:13:32.566856 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6d1ba6db-1cae-461e-9716-5d5b36350218-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "6d1ba6db-1cae-461e-9716-5d5b36350218" (UID: "6d1ba6db-1cae-461e-9716-5d5b36350218"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 08:13:32 crc kubenswrapper[4612]: I0227 08:13:32.570561 4612 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage10-crc" (UniqueName: "kubernetes.io/local-volume/local-storage10-crc") on node "crc" Feb 27 08:13:32 crc kubenswrapper[4612]: I0227 08:13:32.646513 4612 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/6d1ba6db-1cae-461e-9716-5d5b36350218-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Feb 27 08:13:32 crc kubenswrapper[4612]: I0227 08:13:32.646783 4612 reconciler_common.go:293] "Volume detached for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" DevicePath \"\"" Feb 27 08:13:32 crc kubenswrapper[4612]: I0227 08:13:32.870504 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0da436b1-803d-41bb-94ab-9cfd6fa769eb" path="/var/lib/kubelet/pods/0da436b1-803d-41bb-94ab-9cfd6fa769eb/volumes" Feb 27 08:13:32 crc kubenswrapper[4612]: I0227 08:13:32.988139 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-d558885bc-fx2zp"] Feb 27 08:13:32 crc kubenswrapper[4612]: E0227 08:13:32.988515 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d1ba6db-1cae-461e-9716-5d5b36350218" containerName="setup-container" Feb 27 08:13:32 crc kubenswrapper[4612]: I0227 08:13:32.988533 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d1ba6db-1cae-461e-9716-5d5b36350218" containerName="setup-container" Feb 27 08:13:32 crc kubenswrapper[4612]: E0227 08:13:32.988560 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d1ba6db-1cae-461e-9716-5d5b36350218" containerName="rabbitmq" Feb 27 08:13:32 crc kubenswrapper[4612]: I0227 08:13:32.988568 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d1ba6db-1cae-461e-9716-5d5b36350218" containerName="rabbitmq" Feb 27 08:13:32 crc kubenswrapper[4612]: I0227 08:13:32.988750 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="6d1ba6db-1cae-461e-9716-5d5b36350218" containerName="rabbitmq" Feb 27 08:13:32 crc kubenswrapper[4612]: I0227 08:13:32.989731 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-d558885bc-fx2zp" Feb 27 08:13:32 crc kubenswrapper[4612]: I0227 08:13:32.992213 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-edpm-ipam" Feb 27 08:13:33 crc kubenswrapper[4612]: I0227 08:13:33.003959 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-d558885bc-fx2zp"] Feb 27 08:13:33 crc kubenswrapper[4612]: I0227 08:13:33.157882 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/86fd845c-88df-4e7e-8442-89a9f3dddac0-config\") pod \"dnsmasq-dns-d558885bc-fx2zp\" (UID: \"86fd845c-88df-4e7e-8442-89a9f3dddac0\") " pod="openstack/dnsmasq-dns-d558885bc-fx2zp" Feb 27 08:13:33 crc kubenswrapper[4612]: I0227 08:13:33.157946 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/86fd845c-88df-4e7e-8442-89a9f3dddac0-dns-svc\") pod \"dnsmasq-dns-d558885bc-fx2zp\" (UID: \"86fd845c-88df-4e7e-8442-89a9f3dddac0\") " pod="openstack/dnsmasq-dns-d558885bc-fx2zp" Feb 27 08:13:33 crc kubenswrapper[4612]: I0227 08:13:33.158011 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/86fd845c-88df-4e7e-8442-89a9f3dddac0-openstack-edpm-ipam\") pod \"dnsmasq-dns-d558885bc-fx2zp\" (UID: \"86fd845c-88df-4e7e-8442-89a9f3dddac0\") " pod="openstack/dnsmasq-dns-d558885bc-fx2zp" Feb 27 08:13:33 crc kubenswrapper[4612]: I0227 08:13:33.158034 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qrgpd\" (UniqueName: \"kubernetes.io/projected/86fd845c-88df-4e7e-8442-89a9f3dddac0-kube-api-access-qrgpd\") pod \"dnsmasq-dns-d558885bc-fx2zp\" (UID: \"86fd845c-88df-4e7e-8442-89a9f3dddac0\") " pod="openstack/dnsmasq-dns-d558885bc-fx2zp" Feb 27 08:13:33 crc kubenswrapper[4612]: I0227 08:13:33.158056 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/86fd845c-88df-4e7e-8442-89a9f3dddac0-ovsdbserver-sb\") pod \"dnsmasq-dns-d558885bc-fx2zp\" (UID: \"86fd845c-88df-4e7e-8442-89a9f3dddac0\") " pod="openstack/dnsmasq-dns-d558885bc-fx2zp" Feb 27 08:13:33 crc kubenswrapper[4612]: I0227 08:13:33.158115 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/86fd845c-88df-4e7e-8442-89a9f3dddac0-ovsdbserver-nb\") pod \"dnsmasq-dns-d558885bc-fx2zp\" (UID: \"86fd845c-88df-4e7e-8442-89a9f3dddac0\") " pod="openstack/dnsmasq-dns-d558885bc-fx2zp" Feb 27 08:13:33 crc kubenswrapper[4612]: I0227 08:13:33.158158 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/86fd845c-88df-4e7e-8442-89a9f3dddac0-dns-swift-storage-0\") pod \"dnsmasq-dns-d558885bc-fx2zp\" (UID: \"86fd845c-88df-4e7e-8442-89a9f3dddac0\") " pod="openstack/dnsmasq-dns-d558885bc-fx2zp" Feb 27 08:13:33 crc kubenswrapper[4612]: I0227 08:13:33.260155 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/86fd845c-88df-4e7e-8442-89a9f3dddac0-ovsdbserver-nb\") pod \"dnsmasq-dns-d558885bc-fx2zp\" (UID: \"86fd845c-88df-4e7e-8442-89a9f3dddac0\") " pod="openstack/dnsmasq-dns-d558885bc-fx2zp" Feb 27 08:13:33 crc kubenswrapper[4612]: I0227 08:13:33.260216 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/86fd845c-88df-4e7e-8442-89a9f3dddac0-dns-swift-storage-0\") pod \"dnsmasq-dns-d558885bc-fx2zp\" (UID: \"86fd845c-88df-4e7e-8442-89a9f3dddac0\") " pod="openstack/dnsmasq-dns-d558885bc-fx2zp" Feb 27 08:13:33 crc kubenswrapper[4612]: I0227 08:13:33.260274 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/86fd845c-88df-4e7e-8442-89a9f3dddac0-config\") pod \"dnsmasq-dns-d558885bc-fx2zp\" (UID: \"86fd845c-88df-4e7e-8442-89a9f3dddac0\") " pod="openstack/dnsmasq-dns-d558885bc-fx2zp" Feb 27 08:13:33 crc kubenswrapper[4612]: I0227 08:13:33.260304 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/86fd845c-88df-4e7e-8442-89a9f3dddac0-dns-svc\") pod \"dnsmasq-dns-d558885bc-fx2zp\" (UID: \"86fd845c-88df-4e7e-8442-89a9f3dddac0\") " pod="openstack/dnsmasq-dns-d558885bc-fx2zp" Feb 27 08:13:33 crc kubenswrapper[4612]: I0227 08:13:33.260351 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/86fd845c-88df-4e7e-8442-89a9f3dddac0-openstack-edpm-ipam\") pod \"dnsmasq-dns-d558885bc-fx2zp\" (UID: \"86fd845c-88df-4e7e-8442-89a9f3dddac0\") " pod="openstack/dnsmasq-dns-d558885bc-fx2zp" Feb 27 08:13:33 crc kubenswrapper[4612]: I0227 08:13:33.260366 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qrgpd\" (UniqueName: \"kubernetes.io/projected/86fd845c-88df-4e7e-8442-89a9f3dddac0-kube-api-access-qrgpd\") pod \"dnsmasq-dns-d558885bc-fx2zp\" (UID: \"86fd845c-88df-4e7e-8442-89a9f3dddac0\") " pod="openstack/dnsmasq-dns-d558885bc-fx2zp" Feb 27 08:13:33 crc kubenswrapper[4612]: I0227 08:13:33.260382 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/86fd845c-88df-4e7e-8442-89a9f3dddac0-ovsdbserver-sb\") pod \"dnsmasq-dns-d558885bc-fx2zp\" (UID: \"86fd845c-88df-4e7e-8442-89a9f3dddac0\") " pod="openstack/dnsmasq-dns-d558885bc-fx2zp" Feb 27 08:13:33 crc kubenswrapper[4612]: I0227 08:13:33.261120 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/86fd845c-88df-4e7e-8442-89a9f3dddac0-ovsdbserver-nb\") pod \"dnsmasq-dns-d558885bc-fx2zp\" (UID: \"86fd845c-88df-4e7e-8442-89a9f3dddac0\") " pod="openstack/dnsmasq-dns-d558885bc-fx2zp" Feb 27 08:13:33 crc kubenswrapper[4612]: I0227 08:13:33.261277 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/86fd845c-88df-4e7e-8442-89a9f3dddac0-ovsdbserver-sb\") pod \"dnsmasq-dns-d558885bc-fx2zp\" (UID: \"86fd845c-88df-4e7e-8442-89a9f3dddac0\") " pod="openstack/dnsmasq-dns-d558885bc-fx2zp" Feb 27 08:13:33 crc kubenswrapper[4612]: I0227 08:13:33.262095 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/86fd845c-88df-4e7e-8442-89a9f3dddac0-dns-swift-storage-0\") pod \"dnsmasq-dns-d558885bc-fx2zp\" (UID: \"86fd845c-88df-4e7e-8442-89a9f3dddac0\") " pod="openstack/dnsmasq-dns-d558885bc-fx2zp" Feb 27 08:13:33 crc kubenswrapper[4612]: I0227 08:13:33.262241 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/86fd845c-88df-4e7e-8442-89a9f3dddac0-dns-svc\") pod \"dnsmasq-dns-d558885bc-fx2zp\" (UID: \"86fd845c-88df-4e7e-8442-89a9f3dddac0\") " pod="openstack/dnsmasq-dns-d558885bc-fx2zp" Feb 27 08:13:33 crc kubenswrapper[4612]: I0227 08:13:33.262252 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/86fd845c-88df-4e7e-8442-89a9f3dddac0-config\") pod \"dnsmasq-dns-d558885bc-fx2zp\" (UID: \"86fd845c-88df-4e7e-8442-89a9f3dddac0\") " pod="openstack/dnsmasq-dns-d558885bc-fx2zp" Feb 27 08:13:33 crc kubenswrapper[4612]: I0227 08:13:33.262595 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/86fd845c-88df-4e7e-8442-89a9f3dddac0-openstack-edpm-ipam\") pod \"dnsmasq-dns-d558885bc-fx2zp\" (UID: \"86fd845c-88df-4e7e-8442-89a9f3dddac0\") " pod="openstack/dnsmasq-dns-d558885bc-fx2zp" Feb 27 08:13:33 crc kubenswrapper[4612]: I0227 08:13:33.296118 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qrgpd\" (UniqueName: \"kubernetes.io/projected/86fd845c-88df-4e7e-8442-89a9f3dddac0-kube-api-access-qrgpd\") pod \"dnsmasq-dns-d558885bc-fx2zp\" (UID: \"86fd845c-88df-4e7e-8442-89a9f3dddac0\") " pod="openstack/dnsmasq-dns-d558885bc-fx2zp" Feb 27 08:13:33 crc kubenswrapper[4612]: I0227 08:13:33.307742 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-d558885bc-fx2zp" Feb 27 08:13:33 crc kubenswrapper[4612]: I0227 08:13:33.323902 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Feb 27 08:13:33 crc kubenswrapper[4612]: I0227 08:13:33.396376 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Feb 27 08:13:33 crc kubenswrapper[4612]: I0227 08:13:33.404000 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Feb 27 08:13:33 crc kubenswrapper[4612]: I0227 08:13:33.442482 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Feb 27 08:13:33 crc kubenswrapper[4612]: I0227 08:13:33.444370 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Feb 27 08:13:33 crc kubenswrapper[4612]: I0227 08:13:33.447489 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Feb 27 08:13:33 crc kubenswrapper[4612]: I0227 08:13:33.447707 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Feb 27 08:13:33 crc kubenswrapper[4612]: I0227 08:13:33.447769 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Feb 27 08:13:33 crc kubenswrapper[4612]: I0227 08:13:33.447820 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Feb 27 08:13:33 crc kubenswrapper[4612]: I0227 08:13:33.447937 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-lznsf" Feb 27 08:13:33 crc kubenswrapper[4612]: I0227 08:13:33.448094 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Feb 27 08:13:33 crc kubenswrapper[4612]: I0227 08:13:33.449326 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Feb 27 08:13:33 crc kubenswrapper[4612]: I0227 08:13:33.479291 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Feb 27 08:13:33 crc kubenswrapper[4612]: I0227 08:13:33.572094 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/40399d37-c675-448e-8b7e-eb50b5ed5629-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"40399d37-c675-448e-8b7e-eb50b5ed5629\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 08:13:33 crc kubenswrapper[4612]: I0227 08:13:33.572208 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/40399d37-c675-448e-8b7e-eb50b5ed5629-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"40399d37-c675-448e-8b7e-eb50b5ed5629\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 08:13:33 crc kubenswrapper[4612]: I0227 08:13:33.572295 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"40399d37-c675-448e-8b7e-eb50b5ed5629\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 08:13:33 crc kubenswrapper[4612]: I0227 08:13:33.572346 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/40399d37-c675-448e-8b7e-eb50b5ed5629-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"40399d37-c675-448e-8b7e-eb50b5ed5629\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 08:13:33 crc kubenswrapper[4612]: I0227 08:13:33.572456 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/40399d37-c675-448e-8b7e-eb50b5ed5629-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"40399d37-c675-448e-8b7e-eb50b5ed5629\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 08:13:33 crc kubenswrapper[4612]: I0227 08:13:33.572510 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/40399d37-c675-448e-8b7e-eb50b5ed5629-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"40399d37-c675-448e-8b7e-eb50b5ed5629\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 08:13:33 crc kubenswrapper[4612]: I0227 08:13:33.572531 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5564v\" (UniqueName: \"kubernetes.io/projected/40399d37-c675-448e-8b7e-eb50b5ed5629-kube-api-access-5564v\") pod \"rabbitmq-cell1-server-0\" (UID: \"40399d37-c675-448e-8b7e-eb50b5ed5629\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 08:13:33 crc kubenswrapper[4612]: I0227 08:13:33.572555 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/40399d37-c675-448e-8b7e-eb50b5ed5629-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"40399d37-c675-448e-8b7e-eb50b5ed5629\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 08:13:33 crc kubenswrapper[4612]: I0227 08:13:33.572591 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/40399d37-c675-448e-8b7e-eb50b5ed5629-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"40399d37-c675-448e-8b7e-eb50b5ed5629\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 08:13:33 crc kubenswrapper[4612]: I0227 08:13:33.572720 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/40399d37-c675-448e-8b7e-eb50b5ed5629-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"40399d37-c675-448e-8b7e-eb50b5ed5629\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 08:13:33 crc kubenswrapper[4612]: I0227 08:13:33.572805 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/40399d37-c675-448e-8b7e-eb50b5ed5629-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"40399d37-c675-448e-8b7e-eb50b5ed5629\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 08:13:33 crc kubenswrapper[4612]: I0227 08:13:33.675073 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/40399d37-c675-448e-8b7e-eb50b5ed5629-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"40399d37-c675-448e-8b7e-eb50b5ed5629\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 08:13:33 crc kubenswrapper[4612]: I0227 08:13:33.675128 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5564v\" (UniqueName: \"kubernetes.io/projected/40399d37-c675-448e-8b7e-eb50b5ed5629-kube-api-access-5564v\") pod \"rabbitmq-cell1-server-0\" (UID: \"40399d37-c675-448e-8b7e-eb50b5ed5629\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 08:13:33 crc kubenswrapper[4612]: I0227 08:13:33.675157 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/40399d37-c675-448e-8b7e-eb50b5ed5629-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"40399d37-c675-448e-8b7e-eb50b5ed5629\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 08:13:33 crc kubenswrapper[4612]: I0227 08:13:33.675199 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/40399d37-c675-448e-8b7e-eb50b5ed5629-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"40399d37-c675-448e-8b7e-eb50b5ed5629\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 08:13:33 crc kubenswrapper[4612]: I0227 08:13:33.675240 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/40399d37-c675-448e-8b7e-eb50b5ed5629-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"40399d37-c675-448e-8b7e-eb50b5ed5629\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 08:13:33 crc kubenswrapper[4612]: I0227 08:13:33.675311 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/40399d37-c675-448e-8b7e-eb50b5ed5629-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"40399d37-c675-448e-8b7e-eb50b5ed5629\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 08:13:33 crc kubenswrapper[4612]: I0227 08:13:33.675368 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/40399d37-c675-448e-8b7e-eb50b5ed5629-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"40399d37-c675-448e-8b7e-eb50b5ed5629\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 08:13:33 crc kubenswrapper[4612]: I0227 08:13:33.675390 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/40399d37-c675-448e-8b7e-eb50b5ed5629-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"40399d37-c675-448e-8b7e-eb50b5ed5629\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 08:13:33 crc kubenswrapper[4612]: I0227 08:13:33.675430 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"40399d37-c675-448e-8b7e-eb50b5ed5629\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 08:13:33 crc kubenswrapper[4612]: I0227 08:13:33.675453 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/40399d37-c675-448e-8b7e-eb50b5ed5629-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"40399d37-c675-448e-8b7e-eb50b5ed5629\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 08:13:33 crc kubenswrapper[4612]: I0227 08:13:33.675510 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/40399d37-c675-448e-8b7e-eb50b5ed5629-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"40399d37-c675-448e-8b7e-eb50b5ed5629\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 08:13:33 crc kubenswrapper[4612]: I0227 08:13:33.675581 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/40399d37-c675-448e-8b7e-eb50b5ed5629-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"40399d37-c675-448e-8b7e-eb50b5ed5629\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 08:13:33 crc kubenswrapper[4612]: I0227 08:13:33.676172 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/40399d37-c675-448e-8b7e-eb50b5ed5629-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"40399d37-c675-448e-8b7e-eb50b5ed5629\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 08:13:33 crc kubenswrapper[4612]: I0227 08:13:33.676227 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/40399d37-c675-448e-8b7e-eb50b5ed5629-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"40399d37-c675-448e-8b7e-eb50b5ed5629\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 08:13:33 crc kubenswrapper[4612]: I0227 08:13:33.676435 4612 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"40399d37-c675-448e-8b7e-eb50b5ed5629\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/rabbitmq-cell1-server-0" Feb 27 08:13:33 crc kubenswrapper[4612]: I0227 08:13:33.676766 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/40399d37-c675-448e-8b7e-eb50b5ed5629-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"40399d37-c675-448e-8b7e-eb50b5ed5629\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 08:13:33 crc kubenswrapper[4612]: I0227 08:13:33.677088 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/40399d37-c675-448e-8b7e-eb50b5ed5629-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"40399d37-c675-448e-8b7e-eb50b5ed5629\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 08:13:33 crc kubenswrapper[4612]: I0227 08:13:33.681091 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/40399d37-c675-448e-8b7e-eb50b5ed5629-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"40399d37-c675-448e-8b7e-eb50b5ed5629\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 08:13:33 crc kubenswrapper[4612]: I0227 08:13:33.681153 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/40399d37-c675-448e-8b7e-eb50b5ed5629-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"40399d37-c675-448e-8b7e-eb50b5ed5629\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 08:13:33 crc kubenswrapper[4612]: I0227 08:13:33.683652 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/40399d37-c675-448e-8b7e-eb50b5ed5629-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"40399d37-c675-448e-8b7e-eb50b5ed5629\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 08:13:33 crc kubenswrapper[4612]: I0227 08:13:33.688327 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/40399d37-c675-448e-8b7e-eb50b5ed5629-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"40399d37-c675-448e-8b7e-eb50b5ed5629\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 08:13:33 crc kubenswrapper[4612]: I0227 08:13:33.712148 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5564v\" (UniqueName: \"kubernetes.io/projected/40399d37-c675-448e-8b7e-eb50b5ed5629-kube-api-access-5564v\") pod \"rabbitmq-cell1-server-0\" (UID: \"40399d37-c675-448e-8b7e-eb50b5ed5629\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 08:13:33 crc kubenswrapper[4612]: I0227 08:13:33.716124 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"40399d37-c675-448e-8b7e-eb50b5ed5629\") " pod="openstack/rabbitmq-cell1-server-0" Feb 27 08:13:33 crc kubenswrapper[4612]: I0227 08:13:33.780317 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Feb 27 08:13:33 crc kubenswrapper[4612]: I0227 08:13:33.816172 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-d558885bc-fx2zp"] Feb 27 08:13:34 crc kubenswrapper[4612]: W0227 08:13:34.259368 4612 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod40399d37_c675_448e_8b7e_eb50b5ed5629.slice/crio-edd1b935a6e1f51f14a19d7ab91aa501a09ddd43a2742b50999a5fb7d8cec2ce WatchSource:0}: Error finding container edd1b935a6e1f51f14a19d7ab91aa501a09ddd43a2742b50999a5fb7d8cec2ce: Status 404 returned error can't find the container with id edd1b935a6e1f51f14a19d7ab91aa501a09ddd43a2742b50999a5fb7d8cec2ce Feb 27 08:13:34 crc kubenswrapper[4612]: I0227 08:13:34.260723 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Feb 27 08:13:34 crc kubenswrapper[4612]: I0227 08:13:34.340887 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"40399d37-c675-448e-8b7e-eb50b5ed5629","Type":"ContainerStarted","Data":"edd1b935a6e1f51f14a19d7ab91aa501a09ddd43a2742b50999a5fb7d8cec2ce"} Feb 27 08:13:34 crc kubenswrapper[4612]: I0227 08:13:34.344324 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"ba8e827f-294b-4532-8636-c81eeb511ba4","Type":"ContainerStarted","Data":"49b47efcba74cb09e6469c9bf2263f5a2d61146f2e219f58eafb51edf9085442"} Feb 27 08:13:34 crc kubenswrapper[4612]: I0227 08:13:34.350080 4612 generic.go:334] "Generic (PLEG): container finished" podID="86fd845c-88df-4e7e-8442-89a9f3dddac0" containerID="0e5b628de1d22592b9bf603582578f4f9b2a6682b077358f1d2d50c68dcf8e7f" exitCode=0 Feb 27 08:13:34 crc kubenswrapper[4612]: I0227 08:13:34.350126 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-d558885bc-fx2zp" event={"ID":"86fd845c-88df-4e7e-8442-89a9f3dddac0","Type":"ContainerDied","Data":"0e5b628de1d22592b9bf603582578f4f9b2a6682b077358f1d2d50c68dcf8e7f"} Feb 27 08:13:34 crc kubenswrapper[4612]: I0227 08:13:34.350150 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-d558885bc-fx2zp" event={"ID":"86fd845c-88df-4e7e-8442-89a9f3dddac0","Type":"ContainerStarted","Data":"32e35dfe4ad018eac36c5ab1aad178d598ffc30cb3a3fd64dec573882b069800"} Feb 27 08:13:34 crc kubenswrapper[4612]: I0227 08:13:34.482134 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-h2dm7"] Feb 27 08:13:34 crc kubenswrapper[4612]: I0227 08:13:34.483857 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-h2dm7" Feb 27 08:13:34 crc kubenswrapper[4612]: I0227 08:13:34.547459 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-h2dm7"] Feb 27 08:13:34 crc kubenswrapper[4612]: I0227 08:13:34.608839 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hgphz\" (UniqueName: \"kubernetes.io/projected/93719823-fc05-4f44-9402-76ba74230326-kube-api-access-hgphz\") pod \"redhat-marketplace-h2dm7\" (UID: \"93719823-fc05-4f44-9402-76ba74230326\") " pod="openshift-marketplace/redhat-marketplace-h2dm7" Feb 27 08:13:34 crc kubenswrapper[4612]: I0227 08:13:34.608952 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/93719823-fc05-4f44-9402-76ba74230326-utilities\") pod \"redhat-marketplace-h2dm7\" (UID: \"93719823-fc05-4f44-9402-76ba74230326\") " pod="openshift-marketplace/redhat-marketplace-h2dm7" Feb 27 08:13:34 crc kubenswrapper[4612]: I0227 08:13:34.609080 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/93719823-fc05-4f44-9402-76ba74230326-catalog-content\") pod \"redhat-marketplace-h2dm7\" (UID: \"93719823-fc05-4f44-9402-76ba74230326\") " pod="openshift-marketplace/redhat-marketplace-h2dm7" Feb 27 08:13:34 crc kubenswrapper[4612]: I0227 08:13:34.711988 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/93719823-fc05-4f44-9402-76ba74230326-utilities\") pod \"redhat-marketplace-h2dm7\" (UID: \"93719823-fc05-4f44-9402-76ba74230326\") " pod="openshift-marketplace/redhat-marketplace-h2dm7" Feb 27 08:13:34 crc kubenswrapper[4612]: I0227 08:13:34.712437 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/93719823-fc05-4f44-9402-76ba74230326-utilities\") pod \"redhat-marketplace-h2dm7\" (UID: \"93719823-fc05-4f44-9402-76ba74230326\") " pod="openshift-marketplace/redhat-marketplace-h2dm7" Feb 27 08:13:34 crc kubenswrapper[4612]: I0227 08:13:34.712458 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/93719823-fc05-4f44-9402-76ba74230326-catalog-content\") pod \"redhat-marketplace-h2dm7\" (UID: \"93719823-fc05-4f44-9402-76ba74230326\") " pod="openshift-marketplace/redhat-marketplace-h2dm7" Feb 27 08:13:34 crc kubenswrapper[4612]: I0227 08:13:34.712534 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hgphz\" (UniqueName: \"kubernetes.io/projected/93719823-fc05-4f44-9402-76ba74230326-kube-api-access-hgphz\") pod \"redhat-marketplace-h2dm7\" (UID: \"93719823-fc05-4f44-9402-76ba74230326\") " pod="openshift-marketplace/redhat-marketplace-h2dm7" Feb 27 08:13:34 crc kubenswrapper[4612]: I0227 08:13:34.712773 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/93719823-fc05-4f44-9402-76ba74230326-catalog-content\") pod \"redhat-marketplace-h2dm7\" (UID: \"93719823-fc05-4f44-9402-76ba74230326\") " pod="openshift-marketplace/redhat-marketplace-h2dm7" Feb 27 08:13:34 crc kubenswrapper[4612]: I0227 08:13:34.734140 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hgphz\" (UniqueName: \"kubernetes.io/projected/93719823-fc05-4f44-9402-76ba74230326-kube-api-access-hgphz\") pod \"redhat-marketplace-h2dm7\" (UID: \"93719823-fc05-4f44-9402-76ba74230326\") " pod="openshift-marketplace/redhat-marketplace-h2dm7" Feb 27 08:13:34 crc kubenswrapper[4612]: I0227 08:13:34.816307 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-h2dm7" Feb 27 08:13:34 crc kubenswrapper[4612]: I0227 08:13:34.863576 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6d1ba6db-1cae-461e-9716-5d5b36350218" path="/var/lib/kubelet/pods/6d1ba6db-1cae-461e-9716-5d5b36350218/volumes" Feb 27 08:13:35 crc kubenswrapper[4612]: I0227 08:13:35.319943 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-h2dm7"] Feb 27 08:13:35 crc kubenswrapper[4612]: I0227 08:13:35.361161 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h2dm7" event={"ID":"93719823-fc05-4f44-9402-76ba74230326","Type":"ContainerStarted","Data":"1bd193f25b5ec5589f832eef157c7215434995f48a55f7e6206343c4174a3d0f"} Feb 27 08:13:35 crc kubenswrapper[4612]: I0227 08:13:35.364077 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-d558885bc-fx2zp" event={"ID":"86fd845c-88df-4e7e-8442-89a9f3dddac0","Type":"ContainerStarted","Data":"653404ebb73c3674bb987a8aa8b279ec752c75313a3897db84090edd5c198022"} Feb 27 08:13:35 crc kubenswrapper[4612]: I0227 08:13:35.364210 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-d558885bc-fx2zp" Feb 27 08:13:35 crc kubenswrapper[4612]: I0227 08:13:35.393860 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-d558885bc-fx2zp" podStartSLOduration=3.393842917 podStartE2EDuration="3.393842917s" podCreationTimestamp="2026-02-27 08:13:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 08:13:35.389523613 +0000 UTC m=+1473.243453611" watchObservedRunningTime="2026-02-27 08:13:35.393842917 +0000 UTC m=+1473.247772915" Feb 27 08:13:36 crc kubenswrapper[4612]: I0227 08:13:36.171093 4612 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-qvxtt" podUID="1d73fa1d-60eb-4c32-a82d-96f26011b04d" containerName="registry-server" probeResult="failure" output=< Feb 27 08:13:36 crc kubenswrapper[4612]: timeout: failed to connect service ":50051" within 1s Feb 27 08:13:36 crc kubenswrapper[4612]: > Feb 27 08:13:36 crc kubenswrapper[4612]: I0227 08:13:36.373587 4612 generic.go:334] "Generic (PLEG): container finished" podID="93719823-fc05-4f44-9402-76ba74230326" containerID="5e2181baec2bbde628a7b821c108de1a3631c6b8d79797722a1925b040f87500" exitCode=0 Feb 27 08:13:36 crc kubenswrapper[4612]: I0227 08:13:36.373656 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h2dm7" event={"ID":"93719823-fc05-4f44-9402-76ba74230326","Type":"ContainerDied","Data":"5e2181baec2bbde628a7b821c108de1a3631c6b8d79797722a1925b040f87500"} Feb 27 08:13:36 crc kubenswrapper[4612]: I0227 08:13:36.375566 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"40399d37-c675-448e-8b7e-eb50b5ed5629","Type":"ContainerStarted","Data":"c2e61335d72afdc4fcef30f6f230070b4ae671da2043414ddfb970feaf274381"} Feb 27 08:13:37 crc kubenswrapper[4612]: I0227 08:13:37.397863 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h2dm7" event={"ID":"93719823-fc05-4f44-9402-76ba74230326","Type":"ContainerStarted","Data":"718b917b7e67227f8e55dbb62054ab0e615389be2899b0178f6d4d28e8a6f3b1"} Feb 27 08:13:38 crc kubenswrapper[4612]: I0227 08:13:38.407894 4612 generic.go:334] "Generic (PLEG): container finished" podID="93719823-fc05-4f44-9402-76ba74230326" containerID="718b917b7e67227f8e55dbb62054ab0e615389be2899b0178f6d4d28e8a6f3b1" exitCode=0 Feb 27 08:13:38 crc kubenswrapper[4612]: I0227 08:13:38.407974 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h2dm7" event={"ID":"93719823-fc05-4f44-9402-76ba74230326","Type":"ContainerDied","Data":"718b917b7e67227f8e55dbb62054ab0e615389be2899b0178f6d4d28e8a6f3b1"} Feb 27 08:13:39 crc kubenswrapper[4612]: I0227 08:13:39.418696 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h2dm7" event={"ID":"93719823-fc05-4f44-9402-76ba74230326","Type":"ContainerStarted","Data":"09da15bb5bce0e92dfcffd8cb61715ea41da635a37c38e80e388d41b650d32f8"} Feb 27 08:13:39 crc kubenswrapper[4612]: I0227 08:13:39.441755 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-h2dm7" podStartSLOduration=3.01874451 podStartE2EDuration="5.441735949s" podCreationTimestamp="2026-02-27 08:13:34 +0000 UTC" firstStartedPulling="2026-02-27 08:13:36.375648468 +0000 UTC m=+1474.229578456" lastFinishedPulling="2026-02-27 08:13:38.798639877 +0000 UTC m=+1476.652569895" observedRunningTime="2026-02-27 08:13:39.43622332 +0000 UTC m=+1477.290153318" watchObservedRunningTime="2026-02-27 08:13:39.441735949 +0000 UTC m=+1477.295665937" Feb 27 08:13:43 crc kubenswrapper[4612]: I0227 08:13:43.308922 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-d558885bc-fx2zp" Feb 27 08:13:43 crc kubenswrapper[4612]: I0227 08:13:43.411798 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-cd5cbd7b9-dczkk"] Feb 27 08:13:43 crc kubenswrapper[4612]: I0227 08:13:43.412309 4612 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-cd5cbd7b9-dczkk" podUID="63e633cb-a49f-4adb-89f2-dddff70954b6" containerName="dnsmasq-dns" containerID="cri-o://11796b4f0faab1b25a428fd96b25aa78aace1d64498d85d49d882191d4817abe" gracePeriod=10 Feb 27 08:13:43 crc kubenswrapper[4612]: I0227 08:13:43.630104 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-798f84bf5f-g5zjp"] Feb 27 08:13:43 crc kubenswrapper[4612]: I0227 08:13:43.631799 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-798f84bf5f-g5zjp" Feb 27 08:13:43 crc kubenswrapper[4612]: I0227 08:13:43.681269 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-798f84bf5f-g5zjp"] Feb 27 08:13:43 crc kubenswrapper[4612]: I0227 08:13:43.787554 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3c836155-a5e3-4181-9460-4147825f36ab-ovsdbserver-sb\") pod \"dnsmasq-dns-798f84bf5f-g5zjp\" (UID: \"3c836155-a5e3-4181-9460-4147825f36ab\") " pod="openstack/dnsmasq-dns-798f84bf5f-g5zjp" Feb 27 08:13:43 crc kubenswrapper[4612]: I0227 08:13:43.787627 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8qd9n\" (UniqueName: \"kubernetes.io/projected/3c836155-a5e3-4181-9460-4147825f36ab-kube-api-access-8qd9n\") pod \"dnsmasq-dns-798f84bf5f-g5zjp\" (UID: \"3c836155-a5e3-4181-9460-4147825f36ab\") " pod="openstack/dnsmasq-dns-798f84bf5f-g5zjp" Feb 27 08:13:43 crc kubenswrapper[4612]: I0227 08:13:43.787651 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3c836155-a5e3-4181-9460-4147825f36ab-dns-swift-storage-0\") pod \"dnsmasq-dns-798f84bf5f-g5zjp\" (UID: \"3c836155-a5e3-4181-9460-4147825f36ab\") " pod="openstack/dnsmasq-dns-798f84bf5f-g5zjp" Feb 27 08:13:43 crc kubenswrapper[4612]: I0227 08:13:43.787668 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/3c836155-a5e3-4181-9460-4147825f36ab-openstack-edpm-ipam\") pod \"dnsmasq-dns-798f84bf5f-g5zjp\" (UID: \"3c836155-a5e3-4181-9460-4147825f36ab\") " pod="openstack/dnsmasq-dns-798f84bf5f-g5zjp" Feb 27 08:13:43 crc kubenswrapper[4612]: I0227 08:13:43.787732 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3c836155-a5e3-4181-9460-4147825f36ab-ovsdbserver-nb\") pod \"dnsmasq-dns-798f84bf5f-g5zjp\" (UID: \"3c836155-a5e3-4181-9460-4147825f36ab\") " pod="openstack/dnsmasq-dns-798f84bf5f-g5zjp" Feb 27 08:13:43 crc kubenswrapper[4612]: I0227 08:13:43.787763 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3c836155-a5e3-4181-9460-4147825f36ab-dns-svc\") pod \"dnsmasq-dns-798f84bf5f-g5zjp\" (UID: \"3c836155-a5e3-4181-9460-4147825f36ab\") " pod="openstack/dnsmasq-dns-798f84bf5f-g5zjp" Feb 27 08:13:43 crc kubenswrapper[4612]: I0227 08:13:43.787794 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3c836155-a5e3-4181-9460-4147825f36ab-config\") pod \"dnsmasq-dns-798f84bf5f-g5zjp\" (UID: \"3c836155-a5e3-4181-9460-4147825f36ab\") " pod="openstack/dnsmasq-dns-798f84bf5f-g5zjp" Feb 27 08:13:43 crc kubenswrapper[4612]: I0227 08:13:43.891781 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3c836155-a5e3-4181-9460-4147825f36ab-ovsdbserver-sb\") pod \"dnsmasq-dns-798f84bf5f-g5zjp\" (UID: \"3c836155-a5e3-4181-9460-4147825f36ab\") " pod="openstack/dnsmasq-dns-798f84bf5f-g5zjp" Feb 27 08:13:43 crc kubenswrapper[4612]: I0227 08:13:43.891854 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8qd9n\" (UniqueName: \"kubernetes.io/projected/3c836155-a5e3-4181-9460-4147825f36ab-kube-api-access-8qd9n\") pod \"dnsmasq-dns-798f84bf5f-g5zjp\" (UID: \"3c836155-a5e3-4181-9460-4147825f36ab\") " pod="openstack/dnsmasq-dns-798f84bf5f-g5zjp" Feb 27 08:13:43 crc kubenswrapper[4612]: I0227 08:13:43.891878 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3c836155-a5e3-4181-9460-4147825f36ab-dns-swift-storage-0\") pod \"dnsmasq-dns-798f84bf5f-g5zjp\" (UID: \"3c836155-a5e3-4181-9460-4147825f36ab\") " pod="openstack/dnsmasq-dns-798f84bf5f-g5zjp" Feb 27 08:13:43 crc kubenswrapper[4612]: I0227 08:13:43.891895 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/3c836155-a5e3-4181-9460-4147825f36ab-openstack-edpm-ipam\") pod \"dnsmasq-dns-798f84bf5f-g5zjp\" (UID: \"3c836155-a5e3-4181-9460-4147825f36ab\") " pod="openstack/dnsmasq-dns-798f84bf5f-g5zjp" Feb 27 08:13:43 crc kubenswrapper[4612]: I0227 08:13:43.891944 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3c836155-a5e3-4181-9460-4147825f36ab-ovsdbserver-nb\") pod \"dnsmasq-dns-798f84bf5f-g5zjp\" (UID: \"3c836155-a5e3-4181-9460-4147825f36ab\") " pod="openstack/dnsmasq-dns-798f84bf5f-g5zjp" Feb 27 08:13:43 crc kubenswrapper[4612]: I0227 08:13:43.891974 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3c836155-a5e3-4181-9460-4147825f36ab-dns-svc\") pod \"dnsmasq-dns-798f84bf5f-g5zjp\" (UID: \"3c836155-a5e3-4181-9460-4147825f36ab\") " pod="openstack/dnsmasq-dns-798f84bf5f-g5zjp" Feb 27 08:13:43 crc kubenswrapper[4612]: I0227 08:13:43.892008 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3c836155-a5e3-4181-9460-4147825f36ab-config\") pod \"dnsmasq-dns-798f84bf5f-g5zjp\" (UID: \"3c836155-a5e3-4181-9460-4147825f36ab\") " pod="openstack/dnsmasq-dns-798f84bf5f-g5zjp" Feb 27 08:13:43 crc kubenswrapper[4612]: I0227 08:13:43.893325 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3c836155-a5e3-4181-9460-4147825f36ab-dns-swift-storage-0\") pod \"dnsmasq-dns-798f84bf5f-g5zjp\" (UID: \"3c836155-a5e3-4181-9460-4147825f36ab\") " pod="openstack/dnsmasq-dns-798f84bf5f-g5zjp" Feb 27 08:13:43 crc kubenswrapper[4612]: I0227 08:13:43.893343 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3c836155-a5e3-4181-9460-4147825f36ab-config\") pod \"dnsmasq-dns-798f84bf5f-g5zjp\" (UID: \"3c836155-a5e3-4181-9460-4147825f36ab\") " pod="openstack/dnsmasq-dns-798f84bf5f-g5zjp" Feb 27 08:13:43 crc kubenswrapper[4612]: I0227 08:13:43.893910 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3c836155-a5e3-4181-9460-4147825f36ab-ovsdbserver-nb\") pod \"dnsmasq-dns-798f84bf5f-g5zjp\" (UID: \"3c836155-a5e3-4181-9460-4147825f36ab\") " pod="openstack/dnsmasq-dns-798f84bf5f-g5zjp" Feb 27 08:13:43 crc kubenswrapper[4612]: I0227 08:13:43.893960 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3c836155-a5e3-4181-9460-4147825f36ab-ovsdbserver-sb\") pod \"dnsmasq-dns-798f84bf5f-g5zjp\" (UID: \"3c836155-a5e3-4181-9460-4147825f36ab\") " pod="openstack/dnsmasq-dns-798f84bf5f-g5zjp" Feb 27 08:13:43 crc kubenswrapper[4612]: I0227 08:13:43.893987 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3c836155-a5e3-4181-9460-4147825f36ab-dns-svc\") pod \"dnsmasq-dns-798f84bf5f-g5zjp\" (UID: \"3c836155-a5e3-4181-9460-4147825f36ab\") " pod="openstack/dnsmasq-dns-798f84bf5f-g5zjp" Feb 27 08:13:43 crc kubenswrapper[4612]: I0227 08:13:43.894068 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/3c836155-a5e3-4181-9460-4147825f36ab-openstack-edpm-ipam\") pod \"dnsmasq-dns-798f84bf5f-g5zjp\" (UID: \"3c836155-a5e3-4181-9460-4147825f36ab\") " pod="openstack/dnsmasq-dns-798f84bf5f-g5zjp" Feb 27 08:13:43 crc kubenswrapper[4612]: I0227 08:13:43.943684 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8qd9n\" (UniqueName: \"kubernetes.io/projected/3c836155-a5e3-4181-9460-4147825f36ab-kube-api-access-8qd9n\") pod \"dnsmasq-dns-798f84bf5f-g5zjp\" (UID: \"3c836155-a5e3-4181-9460-4147825f36ab\") " pod="openstack/dnsmasq-dns-798f84bf5f-g5zjp" Feb 27 08:13:43 crc kubenswrapper[4612]: I0227 08:13:43.999127 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-798f84bf5f-g5zjp" Feb 27 08:13:44 crc kubenswrapper[4612]: I0227 08:13:44.130968 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cd5cbd7b9-dczkk" Feb 27 08:13:44 crc kubenswrapper[4612]: I0227 08:13:44.304602 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lk7xj\" (UniqueName: \"kubernetes.io/projected/63e633cb-a49f-4adb-89f2-dddff70954b6-kube-api-access-lk7xj\") pod \"63e633cb-a49f-4adb-89f2-dddff70954b6\" (UID: \"63e633cb-a49f-4adb-89f2-dddff70954b6\") " Feb 27 08:13:44 crc kubenswrapper[4612]: I0227 08:13:44.304660 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/63e633cb-a49f-4adb-89f2-dddff70954b6-config\") pod \"63e633cb-a49f-4adb-89f2-dddff70954b6\" (UID: \"63e633cb-a49f-4adb-89f2-dddff70954b6\") " Feb 27 08:13:44 crc kubenswrapper[4612]: I0227 08:13:44.304718 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/63e633cb-a49f-4adb-89f2-dddff70954b6-ovsdbserver-nb\") pod \"63e633cb-a49f-4adb-89f2-dddff70954b6\" (UID: \"63e633cb-a49f-4adb-89f2-dddff70954b6\") " Feb 27 08:13:44 crc kubenswrapper[4612]: I0227 08:13:44.304793 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/63e633cb-a49f-4adb-89f2-dddff70954b6-ovsdbserver-sb\") pod \"63e633cb-a49f-4adb-89f2-dddff70954b6\" (UID: \"63e633cb-a49f-4adb-89f2-dddff70954b6\") " Feb 27 08:13:44 crc kubenswrapper[4612]: I0227 08:13:44.304918 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/63e633cb-a49f-4adb-89f2-dddff70954b6-dns-swift-storage-0\") pod \"63e633cb-a49f-4adb-89f2-dddff70954b6\" (UID: \"63e633cb-a49f-4adb-89f2-dddff70954b6\") " Feb 27 08:13:44 crc kubenswrapper[4612]: I0227 08:13:44.304986 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/63e633cb-a49f-4adb-89f2-dddff70954b6-dns-svc\") pod \"63e633cb-a49f-4adb-89f2-dddff70954b6\" (UID: \"63e633cb-a49f-4adb-89f2-dddff70954b6\") " Feb 27 08:13:44 crc kubenswrapper[4612]: I0227 08:13:44.314140 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/63e633cb-a49f-4adb-89f2-dddff70954b6-kube-api-access-lk7xj" (OuterVolumeSpecName: "kube-api-access-lk7xj") pod "63e633cb-a49f-4adb-89f2-dddff70954b6" (UID: "63e633cb-a49f-4adb-89f2-dddff70954b6"). InnerVolumeSpecName "kube-api-access-lk7xj". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 08:13:44 crc kubenswrapper[4612]: I0227 08:13:44.393627 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/63e633cb-a49f-4adb-89f2-dddff70954b6-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "63e633cb-a49f-4adb-89f2-dddff70954b6" (UID: "63e633cb-a49f-4adb-89f2-dddff70954b6"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 08:13:44 crc kubenswrapper[4612]: I0227 08:13:44.407457 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lk7xj\" (UniqueName: \"kubernetes.io/projected/63e633cb-a49f-4adb-89f2-dddff70954b6-kube-api-access-lk7xj\") on node \"crc\" DevicePath \"\"" Feb 27 08:13:44 crc kubenswrapper[4612]: I0227 08:13:44.407480 4612 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/63e633cb-a49f-4adb-89f2-dddff70954b6-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Feb 27 08:13:44 crc kubenswrapper[4612]: I0227 08:13:44.409875 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/63e633cb-a49f-4adb-89f2-dddff70954b6-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "63e633cb-a49f-4adb-89f2-dddff70954b6" (UID: "63e633cb-a49f-4adb-89f2-dddff70954b6"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 08:13:44 crc kubenswrapper[4612]: I0227 08:13:44.469291 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/63e633cb-a49f-4adb-89f2-dddff70954b6-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "63e633cb-a49f-4adb-89f2-dddff70954b6" (UID: "63e633cb-a49f-4adb-89f2-dddff70954b6"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 08:13:44 crc kubenswrapper[4612]: I0227 08:13:44.484319 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/63e633cb-a49f-4adb-89f2-dddff70954b6-config" (OuterVolumeSpecName: "config") pod "63e633cb-a49f-4adb-89f2-dddff70954b6" (UID: "63e633cb-a49f-4adb-89f2-dddff70954b6"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 08:13:44 crc kubenswrapper[4612]: I0227 08:13:44.491058 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/63e633cb-a49f-4adb-89f2-dddff70954b6-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "63e633cb-a49f-4adb-89f2-dddff70954b6" (UID: "63e633cb-a49f-4adb-89f2-dddff70954b6"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 08:13:44 crc kubenswrapper[4612]: I0227 08:13:44.510666 4612 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/63e633cb-a49f-4adb-89f2-dddff70954b6-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 27 08:13:44 crc kubenswrapper[4612]: I0227 08:13:44.510695 4612 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/63e633cb-a49f-4adb-89f2-dddff70954b6-config\") on node \"crc\" DevicePath \"\"" Feb 27 08:13:44 crc kubenswrapper[4612]: I0227 08:13:44.510716 4612 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/63e633cb-a49f-4adb-89f2-dddff70954b6-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 27 08:13:44 crc kubenswrapper[4612]: I0227 08:13:44.510726 4612 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/63e633cb-a49f-4adb-89f2-dddff70954b6-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 27 08:13:44 crc kubenswrapper[4612]: I0227 08:13:44.511957 4612 generic.go:334] "Generic (PLEG): container finished" podID="63e633cb-a49f-4adb-89f2-dddff70954b6" containerID="11796b4f0faab1b25a428fd96b25aa78aace1d64498d85d49d882191d4817abe" exitCode=0 Feb 27 08:13:44 crc kubenswrapper[4612]: I0227 08:13:44.512036 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cd5cbd7b9-dczkk" event={"ID":"63e633cb-a49f-4adb-89f2-dddff70954b6","Type":"ContainerDied","Data":"11796b4f0faab1b25a428fd96b25aa78aace1d64498d85d49d882191d4817abe"} Feb 27 08:13:44 crc kubenswrapper[4612]: I0227 08:13:44.512115 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cd5cbd7b9-dczkk" event={"ID":"63e633cb-a49f-4adb-89f2-dddff70954b6","Type":"ContainerDied","Data":"f0f5b202d1415e736d16a149f2bb068b1a55fa5bf017d30829b3d46650fed7b7"} Feb 27 08:13:44 crc kubenswrapper[4612]: I0227 08:13:44.512153 4612 scope.go:117] "RemoveContainer" containerID="11796b4f0faab1b25a428fd96b25aa78aace1d64498d85d49d882191d4817abe" Feb 27 08:13:44 crc kubenswrapper[4612]: I0227 08:13:44.512018 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cd5cbd7b9-dczkk" Feb 27 08:13:44 crc kubenswrapper[4612]: I0227 08:13:44.614029 4612 scope.go:117] "RemoveContainer" containerID="6f2784c8b6f934cebc87686e3e4cd3c59b6ab6954f42c1fb0f94b8e1f080e216" Feb 27 08:13:44 crc kubenswrapper[4612]: I0227 08:13:44.638735 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-cd5cbd7b9-dczkk"] Feb 27 08:13:44 crc kubenswrapper[4612]: I0227 08:13:44.650297 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-cd5cbd7b9-dczkk"] Feb 27 08:13:44 crc kubenswrapper[4612]: I0227 08:13:44.673971 4612 scope.go:117] "RemoveContainer" containerID="11796b4f0faab1b25a428fd96b25aa78aace1d64498d85d49d882191d4817abe" Feb 27 08:13:44 crc kubenswrapper[4612]: E0227 08:13:44.675821 4612 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"11796b4f0faab1b25a428fd96b25aa78aace1d64498d85d49d882191d4817abe\": container with ID starting with 11796b4f0faab1b25a428fd96b25aa78aace1d64498d85d49d882191d4817abe not found: ID does not exist" containerID="11796b4f0faab1b25a428fd96b25aa78aace1d64498d85d49d882191d4817abe" Feb 27 08:13:44 crc kubenswrapper[4612]: I0227 08:13:44.675868 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"11796b4f0faab1b25a428fd96b25aa78aace1d64498d85d49d882191d4817abe"} err="failed to get container status \"11796b4f0faab1b25a428fd96b25aa78aace1d64498d85d49d882191d4817abe\": rpc error: code = NotFound desc = could not find container \"11796b4f0faab1b25a428fd96b25aa78aace1d64498d85d49d882191d4817abe\": container with ID starting with 11796b4f0faab1b25a428fd96b25aa78aace1d64498d85d49d882191d4817abe not found: ID does not exist" Feb 27 08:13:44 crc kubenswrapper[4612]: I0227 08:13:44.675894 4612 scope.go:117] "RemoveContainer" containerID="6f2784c8b6f934cebc87686e3e4cd3c59b6ab6954f42c1fb0f94b8e1f080e216" Feb 27 08:13:44 crc kubenswrapper[4612]: E0227 08:13:44.680091 4612 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6f2784c8b6f934cebc87686e3e4cd3c59b6ab6954f42c1fb0f94b8e1f080e216\": container with ID starting with 6f2784c8b6f934cebc87686e3e4cd3c59b6ab6954f42c1fb0f94b8e1f080e216 not found: ID does not exist" containerID="6f2784c8b6f934cebc87686e3e4cd3c59b6ab6954f42c1fb0f94b8e1f080e216" Feb 27 08:13:44 crc kubenswrapper[4612]: I0227 08:13:44.680163 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6f2784c8b6f934cebc87686e3e4cd3c59b6ab6954f42c1fb0f94b8e1f080e216"} err="failed to get container status \"6f2784c8b6f934cebc87686e3e4cd3c59b6ab6954f42c1fb0f94b8e1f080e216\": rpc error: code = NotFound desc = could not find container \"6f2784c8b6f934cebc87686e3e4cd3c59b6ab6954f42c1fb0f94b8e1f080e216\": container with ID starting with 6f2784c8b6f934cebc87686e3e4cd3c59b6ab6954f42c1fb0f94b8e1f080e216 not found: ID does not exist" Feb 27 08:13:44 crc kubenswrapper[4612]: I0227 08:13:44.739064 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-798f84bf5f-g5zjp"] Feb 27 08:13:44 crc kubenswrapper[4612]: I0227 08:13:44.817134 4612 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-h2dm7" Feb 27 08:13:44 crc kubenswrapper[4612]: I0227 08:13:44.818179 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-h2dm7" Feb 27 08:13:44 crc kubenswrapper[4612]: I0227 08:13:44.876897 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="63e633cb-a49f-4adb-89f2-dddff70954b6" path="/var/lib/kubelet/pods/63e633cb-a49f-4adb-89f2-dddff70954b6/volumes" Feb 27 08:13:44 crc kubenswrapper[4612]: I0227 08:13:44.887642 4612 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-h2dm7" Feb 27 08:13:45 crc kubenswrapper[4612]: I0227 08:13:45.176210 4612 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-qvxtt" Feb 27 08:13:45 crc kubenswrapper[4612]: I0227 08:13:45.247999 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-qvxtt" Feb 27 08:13:45 crc kubenswrapper[4612]: I0227 08:13:45.522915 4612 generic.go:334] "Generic (PLEG): container finished" podID="3c836155-a5e3-4181-9460-4147825f36ab" containerID="b0966ffa4f2b0548856ed33231df6393ed6ff7311221170312ad71725e930b77" exitCode=0 Feb 27 08:13:45 crc kubenswrapper[4612]: I0227 08:13:45.523025 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-798f84bf5f-g5zjp" event={"ID":"3c836155-a5e3-4181-9460-4147825f36ab","Type":"ContainerDied","Data":"b0966ffa4f2b0548856ed33231df6393ed6ff7311221170312ad71725e930b77"} Feb 27 08:13:45 crc kubenswrapper[4612]: I0227 08:13:45.523321 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-798f84bf5f-g5zjp" event={"ID":"3c836155-a5e3-4181-9460-4147825f36ab","Type":"ContainerStarted","Data":"df6c00216b4ff5658cbc7b722850b3f1ebca5147e7817866424dcb096cca5443"} Feb 27 08:13:45 crc kubenswrapper[4612]: I0227 08:13:45.623043 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-h2dm7" Feb 27 08:13:45 crc kubenswrapper[4612]: I0227 08:13:45.681216 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-qvxtt"] Feb 27 08:13:46 crc kubenswrapper[4612]: I0227 08:13:46.538627 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-798f84bf5f-g5zjp" event={"ID":"3c836155-a5e3-4181-9460-4147825f36ab","Type":"ContainerStarted","Data":"5f2b4cced331f3906f027c807f2a026e2f267117ce1ac57b91896704408673d3"} Feb 27 08:13:46 crc kubenswrapper[4612]: I0227 08:13:46.538865 4612 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-qvxtt" podUID="1d73fa1d-60eb-4c32-a82d-96f26011b04d" containerName="registry-server" containerID="cri-o://4df1e0965c1632a4513b96f78bb286899b2034df53742ac3953ed35fe0e59ed8" gracePeriod=2 Feb 27 08:13:46 crc kubenswrapper[4612]: I0227 08:13:46.539149 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-798f84bf5f-g5zjp" Feb 27 08:13:46 crc kubenswrapper[4612]: I0227 08:13:46.569548 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-798f84bf5f-g5zjp" podStartSLOduration=3.569526958 podStartE2EDuration="3.569526958s" podCreationTimestamp="2026-02-27 08:13:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 08:13:46.564288188 +0000 UTC m=+1484.418218196" watchObservedRunningTime="2026-02-27 08:13:46.569526958 +0000 UTC m=+1484.423456956" Feb 27 08:13:47 crc kubenswrapper[4612]: I0227 08:13:47.035110 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qvxtt" Feb 27 08:13:47 crc kubenswrapper[4612]: I0227 08:13:47.163427 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d73fa1d-60eb-4c32-a82d-96f26011b04d-catalog-content\") pod \"1d73fa1d-60eb-4c32-a82d-96f26011b04d\" (UID: \"1d73fa1d-60eb-4c32-a82d-96f26011b04d\") " Feb 27 08:13:47 crc kubenswrapper[4612]: I0227 08:13:47.163578 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d73fa1d-60eb-4c32-a82d-96f26011b04d-utilities\") pod \"1d73fa1d-60eb-4c32-a82d-96f26011b04d\" (UID: \"1d73fa1d-60eb-4c32-a82d-96f26011b04d\") " Feb 27 08:13:47 crc kubenswrapper[4612]: I0227 08:13:47.163770 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2n9dl\" (UniqueName: \"kubernetes.io/projected/1d73fa1d-60eb-4c32-a82d-96f26011b04d-kube-api-access-2n9dl\") pod \"1d73fa1d-60eb-4c32-a82d-96f26011b04d\" (UID: \"1d73fa1d-60eb-4c32-a82d-96f26011b04d\") " Feb 27 08:13:47 crc kubenswrapper[4612]: I0227 08:13:47.166292 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d73fa1d-60eb-4c32-a82d-96f26011b04d-utilities" (OuterVolumeSpecName: "utilities") pod "1d73fa1d-60eb-4c32-a82d-96f26011b04d" (UID: "1d73fa1d-60eb-4c32-a82d-96f26011b04d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 08:13:47 crc kubenswrapper[4612]: I0227 08:13:47.184291 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d73fa1d-60eb-4c32-a82d-96f26011b04d-kube-api-access-2n9dl" (OuterVolumeSpecName: "kube-api-access-2n9dl") pod "1d73fa1d-60eb-4c32-a82d-96f26011b04d" (UID: "1d73fa1d-60eb-4c32-a82d-96f26011b04d"). InnerVolumeSpecName "kube-api-access-2n9dl". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 08:13:47 crc kubenswrapper[4612]: I0227 08:13:47.265766 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2n9dl\" (UniqueName: \"kubernetes.io/projected/1d73fa1d-60eb-4c32-a82d-96f26011b04d-kube-api-access-2n9dl\") on node \"crc\" DevicePath \"\"" Feb 27 08:13:47 crc kubenswrapper[4612]: I0227 08:13:47.265792 4612 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d73fa1d-60eb-4c32-a82d-96f26011b04d-utilities\") on node \"crc\" DevicePath \"\"" Feb 27 08:13:47 crc kubenswrapper[4612]: I0227 08:13:47.283668 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d73fa1d-60eb-4c32-a82d-96f26011b04d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d73fa1d-60eb-4c32-a82d-96f26011b04d" (UID: "1d73fa1d-60eb-4c32-a82d-96f26011b04d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 08:13:47 crc kubenswrapper[4612]: I0227 08:13:47.367608 4612 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d73fa1d-60eb-4c32-a82d-96f26011b04d-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 27 08:13:47 crc kubenswrapper[4612]: I0227 08:13:47.548306 4612 generic.go:334] "Generic (PLEG): container finished" podID="1d73fa1d-60eb-4c32-a82d-96f26011b04d" containerID="4df1e0965c1632a4513b96f78bb286899b2034df53742ac3953ed35fe0e59ed8" exitCode=0 Feb 27 08:13:47 crc kubenswrapper[4612]: I0227 08:13:47.548373 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qvxtt" Feb 27 08:13:47 crc kubenswrapper[4612]: I0227 08:13:47.548368 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qvxtt" event={"ID":"1d73fa1d-60eb-4c32-a82d-96f26011b04d","Type":"ContainerDied","Data":"4df1e0965c1632a4513b96f78bb286899b2034df53742ac3953ed35fe0e59ed8"} Feb 27 08:13:47 crc kubenswrapper[4612]: I0227 08:13:47.548833 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qvxtt" event={"ID":"1d73fa1d-60eb-4c32-a82d-96f26011b04d","Type":"ContainerDied","Data":"5dbdb6991e292e05257ebb41b982169110d884f2736c6e613a1921cf15789e83"} Feb 27 08:13:47 crc kubenswrapper[4612]: I0227 08:13:47.548862 4612 scope.go:117] "RemoveContainer" containerID="4df1e0965c1632a4513b96f78bb286899b2034df53742ac3953ed35fe0e59ed8" Feb 27 08:13:47 crc kubenswrapper[4612]: I0227 08:13:47.575825 4612 scope.go:117] "RemoveContainer" containerID="f79143078a8cbf26ec7fdff20b8ad5880be4facea8dc60f5969d90525543ffc4" Feb 27 08:13:47 crc kubenswrapper[4612]: I0227 08:13:47.608330 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-qvxtt"] Feb 27 08:13:47 crc kubenswrapper[4612]: I0227 08:13:47.616582 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-qvxtt"] Feb 27 08:13:47 crc kubenswrapper[4612]: I0227 08:13:47.617990 4612 scope.go:117] "RemoveContainer" containerID="ae32025fde9b0cf665bb5327e5651f0c68488f2101ddc8578c862c314de37807" Feb 27 08:13:47 crc kubenswrapper[4612]: I0227 08:13:47.650325 4612 scope.go:117] "RemoveContainer" containerID="4df1e0965c1632a4513b96f78bb286899b2034df53742ac3953ed35fe0e59ed8" Feb 27 08:13:47 crc kubenswrapper[4612]: E0227 08:13:47.650715 4612 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4df1e0965c1632a4513b96f78bb286899b2034df53742ac3953ed35fe0e59ed8\": container with ID starting with 4df1e0965c1632a4513b96f78bb286899b2034df53742ac3953ed35fe0e59ed8 not found: ID does not exist" containerID="4df1e0965c1632a4513b96f78bb286899b2034df53742ac3953ed35fe0e59ed8" Feb 27 08:13:47 crc kubenswrapper[4612]: I0227 08:13:47.650750 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4df1e0965c1632a4513b96f78bb286899b2034df53742ac3953ed35fe0e59ed8"} err="failed to get container status \"4df1e0965c1632a4513b96f78bb286899b2034df53742ac3953ed35fe0e59ed8\": rpc error: code = NotFound desc = could not find container \"4df1e0965c1632a4513b96f78bb286899b2034df53742ac3953ed35fe0e59ed8\": container with ID starting with 4df1e0965c1632a4513b96f78bb286899b2034df53742ac3953ed35fe0e59ed8 not found: ID does not exist" Feb 27 08:13:47 crc kubenswrapper[4612]: I0227 08:13:47.650778 4612 scope.go:117] "RemoveContainer" containerID="f79143078a8cbf26ec7fdff20b8ad5880be4facea8dc60f5969d90525543ffc4" Feb 27 08:13:47 crc kubenswrapper[4612]: E0227 08:13:47.651315 4612 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f79143078a8cbf26ec7fdff20b8ad5880be4facea8dc60f5969d90525543ffc4\": container with ID starting with f79143078a8cbf26ec7fdff20b8ad5880be4facea8dc60f5969d90525543ffc4 not found: ID does not exist" containerID="f79143078a8cbf26ec7fdff20b8ad5880be4facea8dc60f5969d90525543ffc4" Feb 27 08:13:47 crc kubenswrapper[4612]: I0227 08:13:47.651357 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f79143078a8cbf26ec7fdff20b8ad5880be4facea8dc60f5969d90525543ffc4"} err="failed to get container status \"f79143078a8cbf26ec7fdff20b8ad5880be4facea8dc60f5969d90525543ffc4\": rpc error: code = NotFound desc = could not find container \"f79143078a8cbf26ec7fdff20b8ad5880be4facea8dc60f5969d90525543ffc4\": container with ID starting with f79143078a8cbf26ec7fdff20b8ad5880be4facea8dc60f5969d90525543ffc4 not found: ID does not exist" Feb 27 08:13:47 crc kubenswrapper[4612]: I0227 08:13:47.651385 4612 scope.go:117] "RemoveContainer" containerID="ae32025fde9b0cf665bb5327e5651f0c68488f2101ddc8578c862c314de37807" Feb 27 08:13:47 crc kubenswrapper[4612]: E0227 08:13:47.651765 4612 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ae32025fde9b0cf665bb5327e5651f0c68488f2101ddc8578c862c314de37807\": container with ID starting with ae32025fde9b0cf665bb5327e5651f0c68488f2101ddc8578c862c314de37807 not found: ID does not exist" containerID="ae32025fde9b0cf665bb5327e5651f0c68488f2101ddc8578c862c314de37807" Feb 27 08:13:47 crc kubenswrapper[4612]: I0227 08:13:47.651816 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ae32025fde9b0cf665bb5327e5651f0c68488f2101ddc8578c862c314de37807"} err="failed to get container status \"ae32025fde9b0cf665bb5327e5651f0c68488f2101ddc8578c862c314de37807\": rpc error: code = NotFound desc = could not find container \"ae32025fde9b0cf665bb5327e5651f0c68488f2101ddc8578c862c314de37807\": container with ID starting with ae32025fde9b0cf665bb5327e5651f0c68488f2101ddc8578c862c314de37807 not found: ID does not exist" Feb 27 08:13:47 crc kubenswrapper[4612]: I0227 08:13:47.869861 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-h2dm7"] Feb 27 08:13:48 crc kubenswrapper[4612]: I0227 08:13:48.565067 4612 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-h2dm7" podUID="93719823-fc05-4f44-9402-76ba74230326" containerName="registry-server" containerID="cri-o://09da15bb5bce0e92dfcffd8cb61715ea41da635a37c38e80e388d41b650d32f8" gracePeriod=2 Feb 27 08:13:48 crc kubenswrapper[4612]: I0227 08:13:48.866666 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d73fa1d-60eb-4c32-a82d-96f26011b04d" path="/var/lib/kubelet/pods/1d73fa1d-60eb-4c32-a82d-96f26011b04d/volumes" Feb 27 08:13:49 crc kubenswrapper[4612]: I0227 08:13:49.130932 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-h2dm7" Feb 27 08:13:49 crc kubenswrapper[4612]: I0227 08:13:49.206415 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hgphz\" (UniqueName: \"kubernetes.io/projected/93719823-fc05-4f44-9402-76ba74230326-kube-api-access-hgphz\") pod \"93719823-fc05-4f44-9402-76ba74230326\" (UID: \"93719823-fc05-4f44-9402-76ba74230326\") " Feb 27 08:13:49 crc kubenswrapper[4612]: I0227 08:13:49.206559 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/93719823-fc05-4f44-9402-76ba74230326-utilities\") pod \"93719823-fc05-4f44-9402-76ba74230326\" (UID: \"93719823-fc05-4f44-9402-76ba74230326\") " Feb 27 08:13:49 crc kubenswrapper[4612]: I0227 08:13:49.206666 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/93719823-fc05-4f44-9402-76ba74230326-catalog-content\") pod \"93719823-fc05-4f44-9402-76ba74230326\" (UID: \"93719823-fc05-4f44-9402-76ba74230326\") " Feb 27 08:13:49 crc kubenswrapper[4612]: I0227 08:13:49.208246 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/93719823-fc05-4f44-9402-76ba74230326-utilities" (OuterVolumeSpecName: "utilities") pod "93719823-fc05-4f44-9402-76ba74230326" (UID: "93719823-fc05-4f44-9402-76ba74230326"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 08:13:49 crc kubenswrapper[4612]: I0227 08:13:49.228812 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/93719823-fc05-4f44-9402-76ba74230326-kube-api-access-hgphz" (OuterVolumeSpecName: "kube-api-access-hgphz") pod "93719823-fc05-4f44-9402-76ba74230326" (UID: "93719823-fc05-4f44-9402-76ba74230326"). InnerVolumeSpecName "kube-api-access-hgphz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 08:13:49 crc kubenswrapper[4612]: I0227 08:13:49.231502 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/93719823-fc05-4f44-9402-76ba74230326-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "93719823-fc05-4f44-9402-76ba74230326" (UID: "93719823-fc05-4f44-9402-76ba74230326"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 08:13:49 crc kubenswrapper[4612]: I0227 08:13:49.309653 4612 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/93719823-fc05-4f44-9402-76ba74230326-utilities\") on node \"crc\" DevicePath \"\"" Feb 27 08:13:49 crc kubenswrapper[4612]: I0227 08:13:49.309710 4612 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/93719823-fc05-4f44-9402-76ba74230326-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 27 08:13:49 crc kubenswrapper[4612]: I0227 08:13:49.309728 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hgphz\" (UniqueName: \"kubernetes.io/projected/93719823-fc05-4f44-9402-76ba74230326-kube-api-access-hgphz\") on node \"crc\" DevicePath \"\"" Feb 27 08:13:49 crc kubenswrapper[4612]: I0227 08:13:49.578792 4612 generic.go:334] "Generic (PLEG): container finished" podID="93719823-fc05-4f44-9402-76ba74230326" containerID="09da15bb5bce0e92dfcffd8cb61715ea41da635a37c38e80e388d41b650d32f8" exitCode=0 Feb 27 08:13:49 crc kubenswrapper[4612]: I0227 08:13:49.578846 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h2dm7" event={"ID":"93719823-fc05-4f44-9402-76ba74230326","Type":"ContainerDied","Data":"09da15bb5bce0e92dfcffd8cb61715ea41da635a37c38e80e388d41b650d32f8"} Feb 27 08:13:49 crc kubenswrapper[4612]: I0227 08:13:49.578858 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-h2dm7" Feb 27 08:13:49 crc kubenswrapper[4612]: I0227 08:13:49.578884 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h2dm7" event={"ID":"93719823-fc05-4f44-9402-76ba74230326","Type":"ContainerDied","Data":"1bd193f25b5ec5589f832eef157c7215434995f48a55f7e6206343c4174a3d0f"} Feb 27 08:13:49 crc kubenswrapper[4612]: I0227 08:13:49.578905 4612 scope.go:117] "RemoveContainer" containerID="09da15bb5bce0e92dfcffd8cb61715ea41da635a37c38e80e388d41b650d32f8" Feb 27 08:13:49 crc kubenswrapper[4612]: I0227 08:13:49.605857 4612 scope.go:117] "RemoveContainer" containerID="718b917b7e67227f8e55dbb62054ab0e615389be2899b0178f6d4d28e8a6f3b1" Feb 27 08:13:49 crc kubenswrapper[4612]: I0227 08:13:49.637211 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-h2dm7"] Feb 27 08:13:49 crc kubenswrapper[4612]: I0227 08:13:49.646810 4612 scope.go:117] "RemoveContainer" containerID="5e2181baec2bbde628a7b821c108de1a3631c6b8d79797722a1925b040f87500" Feb 27 08:13:49 crc kubenswrapper[4612]: I0227 08:13:49.648013 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-h2dm7"] Feb 27 08:13:49 crc kubenswrapper[4612]: I0227 08:13:49.692831 4612 scope.go:117] "RemoveContainer" containerID="09da15bb5bce0e92dfcffd8cb61715ea41da635a37c38e80e388d41b650d32f8" Feb 27 08:13:49 crc kubenswrapper[4612]: E0227 08:13:49.693851 4612 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"09da15bb5bce0e92dfcffd8cb61715ea41da635a37c38e80e388d41b650d32f8\": container with ID starting with 09da15bb5bce0e92dfcffd8cb61715ea41da635a37c38e80e388d41b650d32f8 not found: ID does not exist" containerID="09da15bb5bce0e92dfcffd8cb61715ea41da635a37c38e80e388d41b650d32f8" Feb 27 08:13:49 crc kubenswrapper[4612]: I0227 08:13:49.693890 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"09da15bb5bce0e92dfcffd8cb61715ea41da635a37c38e80e388d41b650d32f8"} err="failed to get container status \"09da15bb5bce0e92dfcffd8cb61715ea41da635a37c38e80e388d41b650d32f8\": rpc error: code = NotFound desc = could not find container \"09da15bb5bce0e92dfcffd8cb61715ea41da635a37c38e80e388d41b650d32f8\": container with ID starting with 09da15bb5bce0e92dfcffd8cb61715ea41da635a37c38e80e388d41b650d32f8 not found: ID does not exist" Feb 27 08:13:49 crc kubenswrapper[4612]: I0227 08:13:49.693917 4612 scope.go:117] "RemoveContainer" containerID="718b917b7e67227f8e55dbb62054ab0e615389be2899b0178f6d4d28e8a6f3b1" Feb 27 08:13:49 crc kubenswrapper[4612]: E0227 08:13:49.694388 4612 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"718b917b7e67227f8e55dbb62054ab0e615389be2899b0178f6d4d28e8a6f3b1\": container with ID starting with 718b917b7e67227f8e55dbb62054ab0e615389be2899b0178f6d4d28e8a6f3b1 not found: ID does not exist" containerID="718b917b7e67227f8e55dbb62054ab0e615389be2899b0178f6d4d28e8a6f3b1" Feb 27 08:13:49 crc kubenswrapper[4612]: I0227 08:13:49.694418 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"718b917b7e67227f8e55dbb62054ab0e615389be2899b0178f6d4d28e8a6f3b1"} err="failed to get container status \"718b917b7e67227f8e55dbb62054ab0e615389be2899b0178f6d4d28e8a6f3b1\": rpc error: code = NotFound desc = could not find container \"718b917b7e67227f8e55dbb62054ab0e615389be2899b0178f6d4d28e8a6f3b1\": container with ID starting with 718b917b7e67227f8e55dbb62054ab0e615389be2899b0178f6d4d28e8a6f3b1 not found: ID does not exist" Feb 27 08:13:49 crc kubenswrapper[4612]: I0227 08:13:49.694437 4612 scope.go:117] "RemoveContainer" containerID="5e2181baec2bbde628a7b821c108de1a3631c6b8d79797722a1925b040f87500" Feb 27 08:13:49 crc kubenswrapper[4612]: E0227 08:13:49.694691 4612 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5e2181baec2bbde628a7b821c108de1a3631c6b8d79797722a1925b040f87500\": container with ID starting with 5e2181baec2bbde628a7b821c108de1a3631c6b8d79797722a1925b040f87500 not found: ID does not exist" containerID="5e2181baec2bbde628a7b821c108de1a3631c6b8d79797722a1925b040f87500" Feb 27 08:13:49 crc kubenswrapper[4612]: I0227 08:13:49.694794 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5e2181baec2bbde628a7b821c108de1a3631c6b8d79797722a1925b040f87500"} err="failed to get container status \"5e2181baec2bbde628a7b821c108de1a3631c6b8d79797722a1925b040f87500\": rpc error: code = NotFound desc = could not find container \"5e2181baec2bbde628a7b821c108de1a3631c6b8d79797722a1925b040f87500\": container with ID starting with 5e2181baec2bbde628a7b821c108de1a3631c6b8d79797722a1925b040f87500 not found: ID does not exist" Feb 27 08:13:50 crc kubenswrapper[4612]: I0227 08:13:50.868895 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="93719823-fc05-4f44-9402-76ba74230326" path="/var/lib/kubelet/pods/93719823-fc05-4f44-9402-76ba74230326/volumes" Feb 27 08:13:54 crc kubenswrapper[4612]: I0227 08:13:54.001478 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-798f84bf5f-g5zjp" Feb 27 08:13:54 crc kubenswrapper[4612]: I0227 08:13:54.096377 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-d558885bc-fx2zp"] Feb 27 08:13:54 crc kubenswrapper[4612]: I0227 08:13:54.096945 4612 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-d558885bc-fx2zp" podUID="86fd845c-88df-4e7e-8442-89a9f3dddac0" containerName="dnsmasq-dns" containerID="cri-o://653404ebb73c3674bb987a8aa8b279ec752c75313a3897db84090edd5c198022" gracePeriod=10 Feb 27 08:13:54 crc kubenswrapper[4612]: I0227 08:13:54.628246 4612 generic.go:334] "Generic (PLEG): container finished" podID="86fd845c-88df-4e7e-8442-89a9f3dddac0" containerID="653404ebb73c3674bb987a8aa8b279ec752c75313a3897db84090edd5c198022" exitCode=0 Feb 27 08:13:54 crc kubenswrapper[4612]: I0227 08:13:54.628296 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-d558885bc-fx2zp" event={"ID":"86fd845c-88df-4e7e-8442-89a9f3dddac0","Type":"ContainerDied","Data":"653404ebb73c3674bb987a8aa8b279ec752c75313a3897db84090edd5c198022"} Feb 27 08:13:55 crc kubenswrapper[4612]: I0227 08:13:55.176918 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-d558885bc-fx2zp" Feb 27 08:13:55 crc kubenswrapper[4612]: I0227 08:13:55.349233 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/86fd845c-88df-4e7e-8442-89a9f3dddac0-ovsdbserver-sb\") pod \"86fd845c-88df-4e7e-8442-89a9f3dddac0\" (UID: \"86fd845c-88df-4e7e-8442-89a9f3dddac0\") " Feb 27 08:13:55 crc kubenswrapper[4612]: I0227 08:13:55.349334 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/86fd845c-88df-4e7e-8442-89a9f3dddac0-dns-swift-storage-0\") pod \"86fd845c-88df-4e7e-8442-89a9f3dddac0\" (UID: \"86fd845c-88df-4e7e-8442-89a9f3dddac0\") " Feb 27 08:13:55 crc kubenswrapper[4612]: I0227 08:13:55.349407 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/86fd845c-88df-4e7e-8442-89a9f3dddac0-dns-svc\") pod \"86fd845c-88df-4e7e-8442-89a9f3dddac0\" (UID: \"86fd845c-88df-4e7e-8442-89a9f3dddac0\") " Feb 27 08:13:55 crc kubenswrapper[4612]: I0227 08:13:55.349478 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/86fd845c-88df-4e7e-8442-89a9f3dddac0-config\") pod \"86fd845c-88df-4e7e-8442-89a9f3dddac0\" (UID: \"86fd845c-88df-4e7e-8442-89a9f3dddac0\") " Feb 27 08:13:55 crc kubenswrapper[4612]: I0227 08:13:55.349504 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/86fd845c-88df-4e7e-8442-89a9f3dddac0-openstack-edpm-ipam\") pod \"86fd845c-88df-4e7e-8442-89a9f3dddac0\" (UID: \"86fd845c-88df-4e7e-8442-89a9f3dddac0\") " Feb 27 08:13:55 crc kubenswrapper[4612]: I0227 08:13:55.349524 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/86fd845c-88df-4e7e-8442-89a9f3dddac0-ovsdbserver-nb\") pod \"86fd845c-88df-4e7e-8442-89a9f3dddac0\" (UID: \"86fd845c-88df-4e7e-8442-89a9f3dddac0\") " Feb 27 08:13:55 crc kubenswrapper[4612]: I0227 08:13:55.349624 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qrgpd\" (UniqueName: \"kubernetes.io/projected/86fd845c-88df-4e7e-8442-89a9f3dddac0-kube-api-access-qrgpd\") pod \"86fd845c-88df-4e7e-8442-89a9f3dddac0\" (UID: \"86fd845c-88df-4e7e-8442-89a9f3dddac0\") " Feb 27 08:13:55 crc kubenswrapper[4612]: I0227 08:13:55.354107 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/86fd845c-88df-4e7e-8442-89a9f3dddac0-kube-api-access-qrgpd" (OuterVolumeSpecName: "kube-api-access-qrgpd") pod "86fd845c-88df-4e7e-8442-89a9f3dddac0" (UID: "86fd845c-88df-4e7e-8442-89a9f3dddac0"). InnerVolumeSpecName "kube-api-access-qrgpd". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 08:13:55 crc kubenswrapper[4612]: I0227 08:13:55.395467 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/86fd845c-88df-4e7e-8442-89a9f3dddac0-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "86fd845c-88df-4e7e-8442-89a9f3dddac0" (UID: "86fd845c-88df-4e7e-8442-89a9f3dddac0"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 08:13:55 crc kubenswrapper[4612]: I0227 08:13:55.406263 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/86fd845c-88df-4e7e-8442-89a9f3dddac0-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "86fd845c-88df-4e7e-8442-89a9f3dddac0" (UID: "86fd845c-88df-4e7e-8442-89a9f3dddac0"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 08:13:55 crc kubenswrapper[4612]: I0227 08:13:55.413086 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/86fd845c-88df-4e7e-8442-89a9f3dddac0-openstack-edpm-ipam" (OuterVolumeSpecName: "openstack-edpm-ipam") pod "86fd845c-88df-4e7e-8442-89a9f3dddac0" (UID: "86fd845c-88df-4e7e-8442-89a9f3dddac0"). InnerVolumeSpecName "openstack-edpm-ipam". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 08:13:55 crc kubenswrapper[4612]: I0227 08:13:55.413418 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/86fd845c-88df-4e7e-8442-89a9f3dddac0-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "86fd845c-88df-4e7e-8442-89a9f3dddac0" (UID: "86fd845c-88df-4e7e-8442-89a9f3dddac0"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 08:13:55 crc kubenswrapper[4612]: I0227 08:13:55.413666 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/86fd845c-88df-4e7e-8442-89a9f3dddac0-config" (OuterVolumeSpecName: "config") pod "86fd845c-88df-4e7e-8442-89a9f3dddac0" (UID: "86fd845c-88df-4e7e-8442-89a9f3dddac0"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 08:13:55 crc kubenswrapper[4612]: I0227 08:13:55.424185 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/86fd845c-88df-4e7e-8442-89a9f3dddac0-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "86fd845c-88df-4e7e-8442-89a9f3dddac0" (UID: "86fd845c-88df-4e7e-8442-89a9f3dddac0"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 08:13:55 crc kubenswrapper[4612]: I0227 08:13:55.451589 4612 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/86fd845c-88df-4e7e-8442-89a9f3dddac0-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Feb 27 08:13:55 crc kubenswrapper[4612]: I0227 08:13:55.451624 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qrgpd\" (UniqueName: \"kubernetes.io/projected/86fd845c-88df-4e7e-8442-89a9f3dddac0-kube-api-access-qrgpd\") on node \"crc\" DevicePath \"\"" Feb 27 08:13:55 crc kubenswrapper[4612]: I0227 08:13:55.451635 4612 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/86fd845c-88df-4e7e-8442-89a9f3dddac0-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Feb 27 08:13:55 crc kubenswrapper[4612]: I0227 08:13:55.451644 4612 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/86fd845c-88df-4e7e-8442-89a9f3dddac0-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Feb 27 08:13:55 crc kubenswrapper[4612]: I0227 08:13:55.451654 4612 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/86fd845c-88df-4e7e-8442-89a9f3dddac0-dns-svc\") on node \"crc\" DevicePath \"\"" Feb 27 08:13:55 crc kubenswrapper[4612]: I0227 08:13:55.451665 4612 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/86fd845c-88df-4e7e-8442-89a9f3dddac0-config\") on node \"crc\" DevicePath \"\"" Feb 27 08:13:55 crc kubenswrapper[4612]: I0227 08:13:55.451675 4612 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/86fd845c-88df-4e7e-8442-89a9f3dddac0-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 27 08:13:55 crc kubenswrapper[4612]: I0227 08:13:55.639319 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-d558885bc-fx2zp" event={"ID":"86fd845c-88df-4e7e-8442-89a9f3dddac0","Type":"ContainerDied","Data":"32e35dfe4ad018eac36c5ab1aad178d598ffc30cb3a3fd64dec573882b069800"} Feb 27 08:13:55 crc kubenswrapper[4612]: I0227 08:13:55.639377 4612 scope.go:117] "RemoveContainer" containerID="653404ebb73c3674bb987a8aa8b279ec752c75313a3897db84090edd5c198022" Feb 27 08:13:55 crc kubenswrapper[4612]: I0227 08:13:55.639915 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-d558885bc-fx2zp" Feb 27 08:13:55 crc kubenswrapper[4612]: I0227 08:13:55.669679 4612 scope.go:117] "RemoveContainer" containerID="0e5b628de1d22592b9bf603582578f4f9b2a6682b077358f1d2d50c68dcf8e7f" Feb 27 08:13:55 crc kubenswrapper[4612]: I0227 08:13:55.707731 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-d558885bc-fx2zp"] Feb 27 08:13:55 crc kubenswrapper[4612]: I0227 08:13:55.717544 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-d558885bc-fx2zp"] Feb 27 08:13:56 crc kubenswrapper[4612]: I0227 08:13:56.863646 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="86fd845c-88df-4e7e-8442-89a9f3dddac0" path="/var/lib/kubelet/pods/86fd845c-88df-4e7e-8442-89a9f3dddac0/volumes" Feb 27 08:14:00 crc kubenswrapper[4612]: I0227 08:14:00.143138 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29536334-bzlzh"] Feb 27 08:14:00 crc kubenswrapper[4612]: E0227 08:14:00.144199 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63e633cb-a49f-4adb-89f2-dddff70954b6" containerName="dnsmasq-dns" Feb 27 08:14:00 crc kubenswrapper[4612]: I0227 08:14:00.144216 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="63e633cb-a49f-4adb-89f2-dddff70954b6" containerName="dnsmasq-dns" Feb 27 08:14:00 crc kubenswrapper[4612]: E0227 08:14:00.144231 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93719823-fc05-4f44-9402-76ba74230326" containerName="extract-utilities" Feb 27 08:14:00 crc kubenswrapper[4612]: I0227 08:14:00.144241 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="93719823-fc05-4f44-9402-76ba74230326" containerName="extract-utilities" Feb 27 08:14:00 crc kubenswrapper[4612]: E0227 08:14:00.144255 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93719823-fc05-4f44-9402-76ba74230326" containerName="registry-server" Feb 27 08:14:00 crc kubenswrapper[4612]: I0227 08:14:00.144266 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="93719823-fc05-4f44-9402-76ba74230326" containerName="registry-server" Feb 27 08:14:00 crc kubenswrapper[4612]: E0227 08:14:00.144290 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1d73fa1d-60eb-4c32-a82d-96f26011b04d" containerName="extract-content" Feb 27 08:14:00 crc kubenswrapper[4612]: I0227 08:14:00.144297 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="1d73fa1d-60eb-4c32-a82d-96f26011b04d" containerName="extract-content" Feb 27 08:14:00 crc kubenswrapper[4612]: E0227 08:14:00.144315 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1d73fa1d-60eb-4c32-a82d-96f26011b04d" containerName="registry-server" Feb 27 08:14:00 crc kubenswrapper[4612]: I0227 08:14:00.144325 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="1d73fa1d-60eb-4c32-a82d-96f26011b04d" containerName="registry-server" Feb 27 08:14:00 crc kubenswrapper[4612]: E0227 08:14:00.144344 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93719823-fc05-4f44-9402-76ba74230326" containerName="extract-content" Feb 27 08:14:00 crc kubenswrapper[4612]: I0227 08:14:00.144354 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="93719823-fc05-4f44-9402-76ba74230326" containerName="extract-content" Feb 27 08:14:00 crc kubenswrapper[4612]: E0227 08:14:00.144373 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1d73fa1d-60eb-4c32-a82d-96f26011b04d" containerName="extract-utilities" Feb 27 08:14:00 crc kubenswrapper[4612]: I0227 08:14:00.144384 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="1d73fa1d-60eb-4c32-a82d-96f26011b04d" containerName="extract-utilities" Feb 27 08:14:00 crc kubenswrapper[4612]: E0227 08:14:00.144399 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="86fd845c-88df-4e7e-8442-89a9f3dddac0" containerName="dnsmasq-dns" Feb 27 08:14:00 crc kubenswrapper[4612]: I0227 08:14:00.144407 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="86fd845c-88df-4e7e-8442-89a9f3dddac0" containerName="dnsmasq-dns" Feb 27 08:14:00 crc kubenswrapper[4612]: E0227 08:14:00.144424 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="86fd845c-88df-4e7e-8442-89a9f3dddac0" containerName="init" Feb 27 08:14:00 crc kubenswrapper[4612]: I0227 08:14:00.144434 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="86fd845c-88df-4e7e-8442-89a9f3dddac0" containerName="init" Feb 27 08:14:00 crc kubenswrapper[4612]: E0227 08:14:00.144453 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63e633cb-a49f-4adb-89f2-dddff70954b6" containerName="init" Feb 27 08:14:00 crc kubenswrapper[4612]: I0227 08:14:00.144462 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="63e633cb-a49f-4adb-89f2-dddff70954b6" containerName="init" Feb 27 08:14:00 crc kubenswrapper[4612]: I0227 08:14:00.144998 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="63e633cb-a49f-4adb-89f2-dddff70954b6" containerName="dnsmasq-dns" Feb 27 08:14:00 crc kubenswrapper[4612]: I0227 08:14:00.145024 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="86fd845c-88df-4e7e-8442-89a9f3dddac0" containerName="dnsmasq-dns" Feb 27 08:14:00 crc kubenswrapper[4612]: I0227 08:14:00.145035 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="1d73fa1d-60eb-4c32-a82d-96f26011b04d" containerName="registry-server" Feb 27 08:14:00 crc kubenswrapper[4612]: I0227 08:14:00.145045 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="93719823-fc05-4f44-9402-76ba74230326" containerName="registry-server" Feb 27 08:14:00 crc kubenswrapper[4612]: I0227 08:14:00.145891 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536334-bzlzh" Feb 27 08:14:00 crc kubenswrapper[4612]: I0227 08:14:00.148729 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 27 08:14:00 crc kubenswrapper[4612]: I0227 08:14:00.148855 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 27 08:14:00 crc kubenswrapper[4612]: I0227 08:14:00.162797 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-5zwfx" Feb 27 08:14:00 crc kubenswrapper[4612]: I0227 08:14:00.175667 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536334-bzlzh"] Feb 27 08:14:00 crc kubenswrapper[4612]: I0227 08:14:00.266277 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5kpj4\" (UniqueName: \"kubernetes.io/projected/92a7b490-f9bd-4150-a650-56a3881ec95a-kube-api-access-5kpj4\") pod \"auto-csr-approver-29536334-bzlzh\" (UID: \"92a7b490-f9bd-4150-a650-56a3881ec95a\") " pod="openshift-infra/auto-csr-approver-29536334-bzlzh" Feb 27 08:14:00 crc kubenswrapper[4612]: I0227 08:14:00.368349 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5kpj4\" (UniqueName: \"kubernetes.io/projected/92a7b490-f9bd-4150-a650-56a3881ec95a-kube-api-access-5kpj4\") pod \"auto-csr-approver-29536334-bzlzh\" (UID: \"92a7b490-f9bd-4150-a650-56a3881ec95a\") " pod="openshift-infra/auto-csr-approver-29536334-bzlzh" Feb 27 08:14:00 crc kubenswrapper[4612]: I0227 08:14:00.390536 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5kpj4\" (UniqueName: \"kubernetes.io/projected/92a7b490-f9bd-4150-a650-56a3881ec95a-kube-api-access-5kpj4\") pod \"auto-csr-approver-29536334-bzlzh\" (UID: \"92a7b490-f9bd-4150-a650-56a3881ec95a\") " pod="openshift-infra/auto-csr-approver-29536334-bzlzh" Feb 27 08:14:00 crc kubenswrapper[4612]: I0227 08:14:00.464134 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536334-bzlzh" Feb 27 08:14:01 crc kubenswrapper[4612]: I0227 08:14:01.005913 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536334-bzlzh"] Feb 27 08:14:01 crc kubenswrapper[4612]: I0227 08:14:01.716533 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536334-bzlzh" event={"ID":"92a7b490-f9bd-4150-a650-56a3881ec95a","Type":"ContainerStarted","Data":"b9dad5d46f763ff2c2b6cbf7a80413f2ddb401d19d0f91f553d0372a89704085"} Feb 27 08:14:03 crc kubenswrapper[4612]: I0227 08:14:03.742403 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536334-bzlzh" event={"ID":"92a7b490-f9bd-4150-a650-56a3881ec95a","Type":"ContainerStarted","Data":"450cb4092f6cbf1bfa5907a300e2f111a1e01887f00e70a6fad596acd3495bb6"} Feb 27 08:14:03 crc kubenswrapper[4612]: I0227 08:14:03.766432 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29536334-bzlzh" podStartSLOduration=2.240257336 podStartE2EDuration="3.766411974s" podCreationTimestamp="2026-02-27 08:14:00 +0000 UTC" firstStartedPulling="2026-02-27 08:14:01.012892601 +0000 UTC m=+1498.866822589" lastFinishedPulling="2026-02-27 08:14:02.539047229 +0000 UTC m=+1500.392977227" observedRunningTime="2026-02-27 08:14:03.757887989 +0000 UTC m=+1501.611817987" watchObservedRunningTime="2026-02-27 08:14:03.766411974 +0000 UTC m=+1501.620341982" Feb 27 08:14:04 crc kubenswrapper[4612]: I0227 08:14:04.754827 4612 generic.go:334] "Generic (PLEG): container finished" podID="92a7b490-f9bd-4150-a650-56a3881ec95a" containerID="450cb4092f6cbf1bfa5907a300e2f111a1e01887f00e70a6fad596acd3495bb6" exitCode=0 Feb 27 08:14:04 crc kubenswrapper[4612]: I0227 08:14:04.755082 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536334-bzlzh" event={"ID":"92a7b490-f9bd-4150-a650-56a3881ec95a","Type":"ContainerDied","Data":"450cb4092f6cbf1bfa5907a300e2f111a1e01887f00e70a6fad596acd3495bb6"} Feb 27 08:14:05 crc kubenswrapper[4612]: I0227 08:14:05.769833 4612 generic.go:334] "Generic (PLEG): container finished" podID="ba8e827f-294b-4532-8636-c81eeb511ba4" containerID="49b47efcba74cb09e6469c9bf2263f5a2d61146f2e219f58eafb51edf9085442" exitCode=0 Feb 27 08:14:05 crc kubenswrapper[4612]: I0227 08:14:05.769957 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"ba8e827f-294b-4532-8636-c81eeb511ba4","Type":"ContainerDied","Data":"49b47efcba74cb09e6469c9bf2263f5a2d61146f2e219f58eafb51edf9085442"} Feb 27 08:14:06 crc kubenswrapper[4612]: I0227 08:14:06.237937 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536334-bzlzh" Feb 27 08:14:06 crc kubenswrapper[4612]: I0227 08:14:06.324167 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5kpj4\" (UniqueName: \"kubernetes.io/projected/92a7b490-f9bd-4150-a650-56a3881ec95a-kube-api-access-5kpj4\") pod \"92a7b490-f9bd-4150-a650-56a3881ec95a\" (UID: \"92a7b490-f9bd-4150-a650-56a3881ec95a\") " Feb 27 08:14:06 crc kubenswrapper[4612]: I0227 08:14:06.331999 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/92a7b490-f9bd-4150-a650-56a3881ec95a-kube-api-access-5kpj4" (OuterVolumeSpecName: "kube-api-access-5kpj4") pod "92a7b490-f9bd-4150-a650-56a3881ec95a" (UID: "92a7b490-f9bd-4150-a650-56a3881ec95a"). InnerVolumeSpecName "kube-api-access-5kpj4". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 08:14:06 crc kubenswrapper[4612]: I0227 08:14:06.426060 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5kpj4\" (UniqueName: \"kubernetes.io/projected/92a7b490-f9bd-4150-a650-56a3881ec95a-kube-api-access-5kpj4\") on node \"crc\" DevicePath \"\"" Feb 27 08:14:06 crc kubenswrapper[4612]: I0227 08:14:06.810295 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"ba8e827f-294b-4532-8636-c81eeb511ba4","Type":"ContainerStarted","Data":"a9ac416556c9cabb5213af4527338efdce10a09313042c539890195ee0880bc0"} Feb 27 08:14:06 crc kubenswrapper[4612]: I0227 08:14:06.811519 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Feb 27 08:14:06 crc kubenswrapper[4612]: I0227 08:14:06.817279 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536334-bzlzh" event={"ID":"92a7b490-f9bd-4150-a650-56a3881ec95a","Type":"ContainerDied","Data":"b9dad5d46f763ff2c2b6cbf7a80413f2ddb401d19d0f91f553d0372a89704085"} Feb 27 08:14:06 crc kubenswrapper[4612]: I0227 08:14:06.817301 4612 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b9dad5d46f763ff2c2b6cbf7a80413f2ddb401d19d0f91f553d0372a89704085" Feb 27 08:14:06 crc kubenswrapper[4612]: I0227 08:14:06.817339 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536334-bzlzh" Feb 27 08:14:06 crc kubenswrapper[4612]: I0227 08:14:06.849321 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=35.849299437 podStartE2EDuration="35.849299437s" podCreationTimestamp="2026-02-27 08:13:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 08:14:06.831739733 +0000 UTC m=+1504.685669761" watchObservedRunningTime="2026-02-27 08:14:06.849299437 +0000 UTC m=+1504.703229445" Feb 27 08:14:07 crc kubenswrapper[4612]: I0227 08:14:07.375199 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29536328-gbk7h"] Feb 27 08:14:07 crc kubenswrapper[4612]: I0227 08:14:07.382509 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29536328-gbk7h"] Feb 27 08:14:08 crc kubenswrapper[4612]: I0227 08:14:08.834650 4612 generic.go:334] "Generic (PLEG): container finished" podID="40399d37-c675-448e-8b7e-eb50b5ed5629" containerID="c2e61335d72afdc4fcef30f6f230070b4ae671da2043414ddfb970feaf274381" exitCode=0 Feb 27 08:14:08 crc kubenswrapper[4612]: I0227 08:14:08.834705 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"40399d37-c675-448e-8b7e-eb50b5ed5629","Type":"ContainerDied","Data":"c2e61335d72afdc4fcef30f6f230070b4ae671da2043414ddfb970feaf274381"} Feb 27 08:14:08 crc kubenswrapper[4612]: I0227 08:14:08.872482 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ff8b1f18-8c91-41d1-b157-ea9c37c2a4ff" path="/var/lib/kubelet/pods/ff8b1f18-8c91-41d1-b157-ea9c37c2a4ff/volumes" Feb 27 08:14:09 crc kubenswrapper[4612]: I0227 08:14:09.846051 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"40399d37-c675-448e-8b7e-eb50b5ed5629","Type":"ContainerStarted","Data":"ef3073bf23dba531996dc36d2d67c475efe8c046b33081a0a4d8d4eb94efca5f"} Feb 27 08:14:09 crc kubenswrapper[4612]: I0227 08:14:09.846662 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Feb 27 08:14:13 crc kubenswrapper[4612]: I0227 08:14:13.239806 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=40.239782328 podStartE2EDuration="40.239782328s" podCreationTimestamp="2026-02-27 08:13:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 08:14:09.906177133 +0000 UTC m=+1507.760107131" watchObservedRunningTime="2026-02-27 08:14:13.239782328 +0000 UTC m=+1511.093712336" Feb 27 08:14:13 crc kubenswrapper[4612]: I0227 08:14:13.248017 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-g99jn"] Feb 27 08:14:13 crc kubenswrapper[4612]: E0227 08:14:13.248538 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92a7b490-f9bd-4150-a650-56a3881ec95a" containerName="oc" Feb 27 08:14:13 crc kubenswrapper[4612]: I0227 08:14:13.248585 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="92a7b490-f9bd-4150-a650-56a3881ec95a" containerName="oc" Feb 27 08:14:13 crc kubenswrapper[4612]: I0227 08:14:13.248965 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="92a7b490-f9bd-4150-a650-56a3881ec95a" containerName="oc" Feb 27 08:14:13 crc kubenswrapper[4612]: I0227 08:14:13.249981 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-g99jn" Feb 27 08:14:13 crc kubenswrapper[4612]: I0227 08:14:13.254426 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 27 08:14:13 crc kubenswrapper[4612]: I0227 08:14:13.254460 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 27 08:14:13 crc kubenswrapper[4612]: I0227 08:14:13.254756 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-2pwrb" Feb 27 08:14:13 crc kubenswrapper[4612]: I0227 08:14:13.254959 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 27 08:14:13 crc kubenswrapper[4612]: I0227 08:14:13.274303 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-g99jn"] Feb 27 08:14:13 crc kubenswrapper[4612]: I0227 08:14:13.382104 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/ab3a1ccc-dae6-40e3-859e-e4c0997b9071-ssh-key-openstack-edpm-ipam\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-g99jn\" (UID: \"ab3a1ccc-dae6-40e3-859e-e4c0997b9071\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-g99jn" Feb 27 08:14:13 crc kubenswrapper[4612]: I0227 08:14:13.382213 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-46vjs\" (UniqueName: \"kubernetes.io/projected/ab3a1ccc-dae6-40e3-859e-e4c0997b9071-kube-api-access-46vjs\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-g99jn\" (UID: \"ab3a1ccc-dae6-40e3-859e-e4c0997b9071\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-g99jn" Feb 27 08:14:13 crc kubenswrapper[4612]: I0227 08:14:13.382263 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab3a1ccc-dae6-40e3-859e-e4c0997b9071-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-g99jn\" (UID: \"ab3a1ccc-dae6-40e3-859e-e4c0997b9071\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-g99jn" Feb 27 08:14:13 crc kubenswrapper[4612]: I0227 08:14:13.382314 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ab3a1ccc-dae6-40e3-859e-e4c0997b9071-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-g99jn\" (UID: \"ab3a1ccc-dae6-40e3-859e-e4c0997b9071\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-g99jn" Feb 27 08:14:13 crc kubenswrapper[4612]: I0227 08:14:13.483742 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ab3a1ccc-dae6-40e3-859e-e4c0997b9071-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-g99jn\" (UID: \"ab3a1ccc-dae6-40e3-859e-e4c0997b9071\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-g99jn" Feb 27 08:14:13 crc kubenswrapper[4612]: I0227 08:14:13.483926 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/ab3a1ccc-dae6-40e3-859e-e4c0997b9071-ssh-key-openstack-edpm-ipam\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-g99jn\" (UID: \"ab3a1ccc-dae6-40e3-859e-e4c0997b9071\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-g99jn" Feb 27 08:14:13 crc kubenswrapper[4612]: I0227 08:14:13.483980 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-46vjs\" (UniqueName: \"kubernetes.io/projected/ab3a1ccc-dae6-40e3-859e-e4c0997b9071-kube-api-access-46vjs\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-g99jn\" (UID: \"ab3a1ccc-dae6-40e3-859e-e4c0997b9071\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-g99jn" Feb 27 08:14:13 crc kubenswrapper[4612]: I0227 08:14:13.484049 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab3a1ccc-dae6-40e3-859e-e4c0997b9071-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-g99jn\" (UID: \"ab3a1ccc-dae6-40e3-859e-e4c0997b9071\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-g99jn" Feb 27 08:14:13 crc kubenswrapper[4612]: I0227 08:14:13.489792 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ab3a1ccc-dae6-40e3-859e-e4c0997b9071-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-g99jn\" (UID: \"ab3a1ccc-dae6-40e3-859e-e4c0997b9071\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-g99jn" Feb 27 08:14:13 crc kubenswrapper[4612]: I0227 08:14:13.490637 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab3a1ccc-dae6-40e3-859e-e4c0997b9071-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-g99jn\" (UID: \"ab3a1ccc-dae6-40e3-859e-e4c0997b9071\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-g99jn" Feb 27 08:14:13 crc kubenswrapper[4612]: I0227 08:14:13.506400 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/ab3a1ccc-dae6-40e3-859e-e4c0997b9071-ssh-key-openstack-edpm-ipam\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-g99jn\" (UID: \"ab3a1ccc-dae6-40e3-859e-e4c0997b9071\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-g99jn" Feb 27 08:14:13 crc kubenswrapper[4612]: I0227 08:14:13.509151 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-46vjs\" (UniqueName: \"kubernetes.io/projected/ab3a1ccc-dae6-40e3-859e-e4c0997b9071-kube-api-access-46vjs\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-g99jn\" (UID: \"ab3a1ccc-dae6-40e3-859e-e4c0997b9071\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-g99jn" Feb 27 08:14:13 crc kubenswrapper[4612]: I0227 08:14:13.575039 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-g99jn" Feb 27 08:14:14 crc kubenswrapper[4612]: I0227 08:14:14.394884 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-g99jn"] Feb 27 08:14:14 crc kubenswrapper[4612]: W0227 08:14:14.401173 4612 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podab3a1ccc_dae6_40e3_859e_e4c0997b9071.slice/crio-df7c223821a71db939033be30e78e427e935f45719a564fb7c9229ff589ea5af WatchSource:0}: Error finding container df7c223821a71db939033be30e78e427e935f45719a564fb7c9229ff589ea5af: Status 404 returned error can't find the container with id df7c223821a71db939033be30e78e427e935f45719a564fb7c9229ff589ea5af Feb 27 08:14:14 crc kubenswrapper[4612]: I0227 08:14:14.540129 4612 scope.go:117] "RemoveContainer" containerID="0992a9fec848e35a9a3164334231e441c0a8eb1353824fed5c37ccddfc2eed38" Feb 27 08:14:14 crc kubenswrapper[4612]: I0227 08:14:14.602938 4612 scope.go:117] "RemoveContainer" containerID="eb1bb5464e5938959380380e0adb9e0444116a9729409966f1310218a5de0a2a" Feb 27 08:14:14 crc kubenswrapper[4612]: I0227 08:14:14.631630 4612 scope.go:117] "RemoveContainer" containerID="839c0ba56d6f1ff42f16427d18510198097a735f0b3fdf5b535bc49979e9110b" Feb 27 08:14:14 crc kubenswrapper[4612]: I0227 08:14:14.686757 4612 scope.go:117] "RemoveContainer" containerID="d673d33a8032b5a980f2b7df29867154100118156686660391827c5296f434f8" Feb 27 08:14:14 crc kubenswrapper[4612]: I0227 08:14:14.916611 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-g99jn" event={"ID":"ab3a1ccc-dae6-40e3-859e-e4c0997b9071","Type":"ContainerStarted","Data":"df7c223821a71db939033be30e78e427e935f45719a564fb7c9229ff589ea5af"} Feb 27 08:14:16 crc kubenswrapper[4612]: I0227 08:14:16.026818 4612 patch_prober.go:28] interesting pod/machine-config-daemon-924vb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 27 08:14:16 crc kubenswrapper[4612]: I0227 08:14:16.026883 4612 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 27 08:14:21 crc kubenswrapper[4612]: I0227 08:14:21.695818 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Feb 27 08:14:23 crc kubenswrapper[4612]: I0227 08:14:23.783856 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Feb 27 08:14:38 crc kubenswrapper[4612]: I0227 08:14:38.197297 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-g99jn" event={"ID":"ab3a1ccc-dae6-40e3-859e-e4c0997b9071","Type":"ContainerStarted","Data":"ccdb6ba4f619f1a75c091a0e024c6f2deddc5556ea45894585a1c698ff393941"} Feb 27 08:14:38 crc kubenswrapper[4612]: I0227 08:14:38.230114 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-g99jn" podStartSLOduration=2.181555774 podStartE2EDuration="25.230085582s" podCreationTimestamp="2026-02-27 08:14:13 +0000 UTC" firstStartedPulling="2026-02-27 08:14:14.406898083 +0000 UTC m=+1512.260828081" lastFinishedPulling="2026-02-27 08:14:37.455427891 +0000 UTC m=+1535.309357889" observedRunningTime="2026-02-27 08:14:38.224479131 +0000 UTC m=+1536.078409169" watchObservedRunningTime="2026-02-27 08:14:38.230085582 +0000 UTC m=+1536.084015610" Feb 27 08:14:42 crc kubenswrapper[4612]: I0227 08:14:42.374768 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-4jr57"] Feb 27 08:14:42 crc kubenswrapper[4612]: I0227 08:14:42.379108 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4jr57" Feb 27 08:14:42 crc kubenswrapper[4612]: I0227 08:14:42.392533 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-4jr57"] Feb 27 08:14:42 crc kubenswrapper[4612]: I0227 08:14:42.493761 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t4t88\" (UniqueName: \"kubernetes.io/projected/6480a60b-9c50-4712-a5d1-fd2fbce66d76-kube-api-access-t4t88\") pod \"certified-operators-4jr57\" (UID: \"6480a60b-9c50-4712-a5d1-fd2fbce66d76\") " pod="openshift-marketplace/certified-operators-4jr57" Feb 27 08:14:42 crc kubenswrapper[4612]: I0227 08:14:42.494016 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6480a60b-9c50-4712-a5d1-fd2fbce66d76-utilities\") pod \"certified-operators-4jr57\" (UID: \"6480a60b-9c50-4712-a5d1-fd2fbce66d76\") " pod="openshift-marketplace/certified-operators-4jr57" Feb 27 08:14:42 crc kubenswrapper[4612]: I0227 08:14:42.494297 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6480a60b-9c50-4712-a5d1-fd2fbce66d76-catalog-content\") pod \"certified-operators-4jr57\" (UID: \"6480a60b-9c50-4712-a5d1-fd2fbce66d76\") " pod="openshift-marketplace/certified-operators-4jr57" Feb 27 08:14:42 crc kubenswrapper[4612]: I0227 08:14:42.596302 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6480a60b-9c50-4712-a5d1-fd2fbce66d76-catalog-content\") pod \"certified-operators-4jr57\" (UID: \"6480a60b-9c50-4712-a5d1-fd2fbce66d76\") " pod="openshift-marketplace/certified-operators-4jr57" Feb 27 08:14:42 crc kubenswrapper[4612]: I0227 08:14:42.596355 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t4t88\" (UniqueName: \"kubernetes.io/projected/6480a60b-9c50-4712-a5d1-fd2fbce66d76-kube-api-access-t4t88\") pod \"certified-operators-4jr57\" (UID: \"6480a60b-9c50-4712-a5d1-fd2fbce66d76\") " pod="openshift-marketplace/certified-operators-4jr57" Feb 27 08:14:42 crc kubenswrapper[4612]: I0227 08:14:42.596379 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6480a60b-9c50-4712-a5d1-fd2fbce66d76-utilities\") pod \"certified-operators-4jr57\" (UID: \"6480a60b-9c50-4712-a5d1-fd2fbce66d76\") " pod="openshift-marketplace/certified-operators-4jr57" Feb 27 08:14:42 crc kubenswrapper[4612]: I0227 08:14:42.596917 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6480a60b-9c50-4712-a5d1-fd2fbce66d76-catalog-content\") pod \"certified-operators-4jr57\" (UID: \"6480a60b-9c50-4712-a5d1-fd2fbce66d76\") " pod="openshift-marketplace/certified-operators-4jr57" Feb 27 08:14:42 crc kubenswrapper[4612]: I0227 08:14:42.596939 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6480a60b-9c50-4712-a5d1-fd2fbce66d76-utilities\") pod \"certified-operators-4jr57\" (UID: \"6480a60b-9c50-4712-a5d1-fd2fbce66d76\") " pod="openshift-marketplace/certified-operators-4jr57" Feb 27 08:14:42 crc kubenswrapper[4612]: I0227 08:14:42.616306 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t4t88\" (UniqueName: \"kubernetes.io/projected/6480a60b-9c50-4712-a5d1-fd2fbce66d76-kube-api-access-t4t88\") pod \"certified-operators-4jr57\" (UID: \"6480a60b-9c50-4712-a5d1-fd2fbce66d76\") " pod="openshift-marketplace/certified-operators-4jr57" Feb 27 08:14:42 crc kubenswrapper[4612]: I0227 08:14:42.753388 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4jr57" Feb 27 08:14:43 crc kubenswrapper[4612]: I0227 08:14:43.309852 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-4jr57"] Feb 27 08:14:44 crc kubenswrapper[4612]: I0227 08:14:44.253296 4612 generic.go:334] "Generic (PLEG): container finished" podID="6480a60b-9c50-4712-a5d1-fd2fbce66d76" containerID="4f7e6836df4a2bc335aa1e9c23777b23ddd1b3c2991e88ded6af8a900cc63692" exitCode=0 Feb 27 08:14:44 crc kubenswrapper[4612]: I0227 08:14:44.253405 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4jr57" event={"ID":"6480a60b-9c50-4712-a5d1-fd2fbce66d76","Type":"ContainerDied","Data":"4f7e6836df4a2bc335aa1e9c23777b23ddd1b3c2991e88ded6af8a900cc63692"} Feb 27 08:14:44 crc kubenswrapper[4612]: I0227 08:14:44.253578 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4jr57" event={"ID":"6480a60b-9c50-4712-a5d1-fd2fbce66d76","Type":"ContainerStarted","Data":"65be91a33b6af385983643cc0eeab71edfbe2b9764480698351db64086625ea4"} Feb 27 08:14:45 crc kubenswrapper[4612]: I0227 08:14:45.263828 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4jr57" event={"ID":"6480a60b-9c50-4712-a5d1-fd2fbce66d76","Type":"ContainerStarted","Data":"50d34184851535939b113195699f46c3ec15b6caaf5b1989188f231883e83e55"} Feb 27 08:14:46 crc kubenswrapper[4612]: I0227 08:14:46.027544 4612 patch_prober.go:28] interesting pod/machine-config-daemon-924vb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 27 08:14:46 crc kubenswrapper[4612]: I0227 08:14:46.027622 4612 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 27 08:14:47 crc kubenswrapper[4612]: I0227 08:14:47.298643 4612 generic.go:334] "Generic (PLEG): container finished" podID="6480a60b-9c50-4712-a5d1-fd2fbce66d76" containerID="50d34184851535939b113195699f46c3ec15b6caaf5b1989188f231883e83e55" exitCode=0 Feb 27 08:14:47 crc kubenswrapper[4612]: I0227 08:14:47.299066 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4jr57" event={"ID":"6480a60b-9c50-4712-a5d1-fd2fbce66d76","Type":"ContainerDied","Data":"50d34184851535939b113195699f46c3ec15b6caaf5b1989188f231883e83e55"} Feb 27 08:14:48 crc kubenswrapper[4612]: I0227 08:14:48.310625 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4jr57" event={"ID":"6480a60b-9c50-4712-a5d1-fd2fbce66d76","Type":"ContainerStarted","Data":"3b21934250f871b4d50db07138b3cf64d39d8db01c10cb56254fc6c8b3ae0cdf"} Feb 27 08:14:48 crc kubenswrapper[4612]: I0227 08:14:48.335877 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-4jr57" podStartSLOduration=2.900889825 podStartE2EDuration="6.335859891s" podCreationTimestamp="2026-02-27 08:14:42 +0000 UTC" firstStartedPulling="2026-02-27 08:14:44.255026803 +0000 UTC m=+1542.108956801" lastFinishedPulling="2026-02-27 08:14:47.689996829 +0000 UTC m=+1545.543926867" observedRunningTime="2026-02-27 08:14:48.327748418 +0000 UTC m=+1546.181678426" watchObservedRunningTime="2026-02-27 08:14:48.335859891 +0000 UTC m=+1546.189789889" Feb 27 08:14:49 crc kubenswrapper[4612]: I0227 08:14:49.320947 4612 generic.go:334] "Generic (PLEG): container finished" podID="ab3a1ccc-dae6-40e3-859e-e4c0997b9071" containerID="ccdb6ba4f619f1a75c091a0e024c6f2deddc5556ea45894585a1c698ff393941" exitCode=0 Feb 27 08:14:49 crc kubenswrapper[4612]: I0227 08:14:49.320984 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-g99jn" event={"ID":"ab3a1ccc-dae6-40e3-859e-e4c0997b9071","Type":"ContainerDied","Data":"ccdb6ba4f619f1a75c091a0e024c6f2deddc5556ea45894585a1c698ff393941"} Feb 27 08:14:50 crc kubenswrapper[4612]: I0227 08:14:50.848969 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-g99jn" Feb 27 08:14:50 crc kubenswrapper[4612]: I0227 08:14:50.962438 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/ab3a1ccc-dae6-40e3-859e-e4c0997b9071-ssh-key-openstack-edpm-ipam\") pod \"ab3a1ccc-dae6-40e3-859e-e4c0997b9071\" (UID: \"ab3a1ccc-dae6-40e3-859e-e4c0997b9071\") " Feb 27 08:14:50 crc kubenswrapper[4612]: I0227 08:14:50.962555 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-46vjs\" (UniqueName: \"kubernetes.io/projected/ab3a1ccc-dae6-40e3-859e-e4c0997b9071-kube-api-access-46vjs\") pod \"ab3a1ccc-dae6-40e3-859e-e4c0997b9071\" (UID: \"ab3a1ccc-dae6-40e3-859e-e4c0997b9071\") " Feb 27 08:14:50 crc kubenswrapper[4612]: I0227 08:14:50.962756 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab3a1ccc-dae6-40e3-859e-e4c0997b9071-repo-setup-combined-ca-bundle\") pod \"ab3a1ccc-dae6-40e3-859e-e4c0997b9071\" (UID: \"ab3a1ccc-dae6-40e3-859e-e4c0997b9071\") " Feb 27 08:14:50 crc kubenswrapper[4612]: I0227 08:14:50.962792 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ab3a1ccc-dae6-40e3-859e-e4c0997b9071-inventory\") pod \"ab3a1ccc-dae6-40e3-859e-e4c0997b9071\" (UID: \"ab3a1ccc-dae6-40e3-859e-e4c0997b9071\") " Feb 27 08:14:50 crc kubenswrapper[4612]: I0227 08:14:50.968689 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ab3a1ccc-dae6-40e3-859e-e4c0997b9071-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "ab3a1ccc-dae6-40e3-859e-e4c0997b9071" (UID: "ab3a1ccc-dae6-40e3-859e-e4c0997b9071"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:14:50 crc kubenswrapper[4612]: I0227 08:14:50.973680 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ab3a1ccc-dae6-40e3-859e-e4c0997b9071-kube-api-access-46vjs" (OuterVolumeSpecName: "kube-api-access-46vjs") pod "ab3a1ccc-dae6-40e3-859e-e4c0997b9071" (UID: "ab3a1ccc-dae6-40e3-859e-e4c0997b9071"). InnerVolumeSpecName "kube-api-access-46vjs". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 08:14:50 crc kubenswrapper[4612]: I0227 08:14:50.989358 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ab3a1ccc-dae6-40e3-859e-e4c0997b9071-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "ab3a1ccc-dae6-40e3-859e-e4c0997b9071" (UID: "ab3a1ccc-dae6-40e3-859e-e4c0997b9071"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:14:50 crc kubenswrapper[4612]: I0227 08:14:50.993314 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ab3a1ccc-dae6-40e3-859e-e4c0997b9071-inventory" (OuterVolumeSpecName: "inventory") pod "ab3a1ccc-dae6-40e3-859e-e4c0997b9071" (UID: "ab3a1ccc-dae6-40e3-859e-e4c0997b9071"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:14:51 crc kubenswrapper[4612]: I0227 08:14:51.067162 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-46vjs\" (UniqueName: \"kubernetes.io/projected/ab3a1ccc-dae6-40e3-859e-e4c0997b9071-kube-api-access-46vjs\") on node \"crc\" DevicePath \"\"" Feb 27 08:14:51 crc kubenswrapper[4612]: I0227 08:14:51.067253 4612 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ab3a1ccc-dae6-40e3-859e-e4c0997b9071-inventory\") on node \"crc\" DevicePath \"\"" Feb 27 08:14:51 crc kubenswrapper[4612]: I0227 08:14:51.067399 4612 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab3a1ccc-dae6-40e3-859e-e4c0997b9071-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 08:14:51 crc kubenswrapper[4612]: I0227 08:14:51.067428 4612 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/ab3a1ccc-dae6-40e3-859e-e4c0997b9071-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 27 08:14:51 crc kubenswrapper[4612]: I0227 08:14:51.361022 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-g99jn" event={"ID":"ab3a1ccc-dae6-40e3-859e-e4c0997b9071","Type":"ContainerDied","Data":"df7c223821a71db939033be30e78e427e935f45719a564fb7c9229ff589ea5af"} Feb 27 08:14:51 crc kubenswrapper[4612]: I0227 08:14:51.361071 4612 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="df7c223821a71db939033be30e78e427e935f45719a564fb7c9229ff589ea5af" Feb 27 08:14:51 crc kubenswrapper[4612]: I0227 08:14:51.361139 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-g99jn" Feb 27 08:14:51 crc kubenswrapper[4612]: I0227 08:14:51.449525 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-vz4qv"] Feb 27 08:14:51 crc kubenswrapper[4612]: E0227 08:14:51.449910 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ab3a1ccc-dae6-40e3-859e-e4c0997b9071" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Feb 27 08:14:51 crc kubenswrapper[4612]: I0227 08:14:51.449928 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="ab3a1ccc-dae6-40e3-859e-e4c0997b9071" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Feb 27 08:14:51 crc kubenswrapper[4612]: I0227 08:14:51.450115 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="ab3a1ccc-dae6-40e3-859e-e4c0997b9071" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Feb 27 08:14:51 crc kubenswrapper[4612]: I0227 08:14:51.450659 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-vz4qv" Feb 27 08:14:51 crc kubenswrapper[4612]: I0227 08:14:51.452744 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 27 08:14:51 crc kubenswrapper[4612]: I0227 08:14:51.452848 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 27 08:14:51 crc kubenswrapper[4612]: I0227 08:14:51.453343 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-2pwrb" Feb 27 08:14:51 crc kubenswrapper[4612]: I0227 08:14:51.453361 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 27 08:14:51 crc kubenswrapper[4612]: I0227 08:14:51.466310 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-vz4qv"] Feb 27 08:14:51 crc kubenswrapper[4612]: I0227 08:14:51.482920 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9995bc24-fa2b-4916-8e71-627b11a39de7-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-vz4qv\" (UID: \"9995bc24-fa2b-4916-8e71-627b11a39de7\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-vz4qv" Feb 27 08:14:51 crc kubenswrapper[4612]: I0227 08:14:51.483055 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/9995bc24-fa2b-4916-8e71-627b11a39de7-ssh-key-openstack-edpm-ipam\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-vz4qv\" (UID: \"9995bc24-fa2b-4916-8e71-627b11a39de7\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-vz4qv" Feb 27 08:14:51 crc kubenswrapper[4612]: I0227 08:14:51.483124 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zjhvw\" (UniqueName: \"kubernetes.io/projected/9995bc24-fa2b-4916-8e71-627b11a39de7-kube-api-access-zjhvw\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-vz4qv\" (UID: \"9995bc24-fa2b-4916-8e71-627b11a39de7\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-vz4qv" Feb 27 08:14:51 crc kubenswrapper[4612]: I0227 08:14:51.585007 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/9995bc24-fa2b-4916-8e71-627b11a39de7-ssh-key-openstack-edpm-ipam\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-vz4qv\" (UID: \"9995bc24-fa2b-4916-8e71-627b11a39de7\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-vz4qv" Feb 27 08:14:51 crc kubenswrapper[4612]: I0227 08:14:51.585333 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zjhvw\" (UniqueName: \"kubernetes.io/projected/9995bc24-fa2b-4916-8e71-627b11a39de7-kube-api-access-zjhvw\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-vz4qv\" (UID: \"9995bc24-fa2b-4916-8e71-627b11a39de7\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-vz4qv" Feb 27 08:14:51 crc kubenswrapper[4612]: I0227 08:14:51.585544 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9995bc24-fa2b-4916-8e71-627b11a39de7-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-vz4qv\" (UID: \"9995bc24-fa2b-4916-8e71-627b11a39de7\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-vz4qv" Feb 27 08:14:51 crc kubenswrapper[4612]: I0227 08:14:51.594668 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/9995bc24-fa2b-4916-8e71-627b11a39de7-ssh-key-openstack-edpm-ipam\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-vz4qv\" (UID: \"9995bc24-fa2b-4916-8e71-627b11a39de7\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-vz4qv" Feb 27 08:14:51 crc kubenswrapper[4612]: I0227 08:14:51.598348 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9995bc24-fa2b-4916-8e71-627b11a39de7-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-vz4qv\" (UID: \"9995bc24-fa2b-4916-8e71-627b11a39de7\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-vz4qv" Feb 27 08:14:51 crc kubenswrapper[4612]: I0227 08:14:51.601178 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zjhvw\" (UniqueName: \"kubernetes.io/projected/9995bc24-fa2b-4916-8e71-627b11a39de7-kube-api-access-zjhvw\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-vz4qv\" (UID: \"9995bc24-fa2b-4916-8e71-627b11a39de7\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-vz4qv" Feb 27 08:14:51 crc kubenswrapper[4612]: I0227 08:14:51.807421 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-vz4qv" Feb 27 08:14:52 crc kubenswrapper[4612]: W0227 08:14:52.360211 4612 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9995bc24_fa2b_4916_8e71_627b11a39de7.slice/crio-8dd61a08a166fa66e7aef71fa8ecea4c247e21855cc24f8a217c104f58f1dea4 WatchSource:0}: Error finding container 8dd61a08a166fa66e7aef71fa8ecea4c247e21855cc24f8a217c104f58f1dea4: Status 404 returned error can't find the container with id 8dd61a08a166fa66e7aef71fa8ecea4c247e21855cc24f8a217c104f58f1dea4 Feb 27 08:14:52 crc kubenswrapper[4612]: I0227 08:14:52.367173 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-vz4qv"] Feb 27 08:14:52 crc kubenswrapper[4612]: I0227 08:14:52.372100 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-vz4qv" event={"ID":"9995bc24-fa2b-4916-8e71-627b11a39de7","Type":"ContainerStarted","Data":"8dd61a08a166fa66e7aef71fa8ecea4c247e21855cc24f8a217c104f58f1dea4"} Feb 27 08:14:52 crc kubenswrapper[4612]: I0227 08:14:52.753590 4612 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-4jr57" Feb 27 08:14:52 crc kubenswrapper[4612]: I0227 08:14:52.753656 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-4jr57" Feb 27 08:14:52 crc kubenswrapper[4612]: I0227 08:14:52.808005 4612 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-4jr57" Feb 27 08:14:53 crc kubenswrapper[4612]: I0227 08:14:53.429138 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-4jr57" Feb 27 08:14:53 crc kubenswrapper[4612]: I0227 08:14:53.483020 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-4jr57"] Feb 27 08:14:55 crc kubenswrapper[4612]: I0227 08:14:55.398049 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-vz4qv" event={"ID":"9995bc24-fa2b-4916-8e71-627b11a39de7","Type":"ContainerStarted","Data":"480f953375b6159c9d93766fa580287e708c9b765189fbdd73dcdc4a7c5bd4b1"} Feb 27 08:14:55 crc kubenswrapper[4612]: I0227 08:14:55.398287 4612 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-4jr57" podUID="6480a60b-9c50-4712-a5d1-fd2fbce66d76" containerName="registry-server" containerID="cri-o://3b21934250f871b4d50db07138b3cf64d39d8db01c10cb56254fc6c8b3ae0cdf" gracePeriod=2 Feb 27 08:14:55 crc kubenswrapper[4612]: I0227 08:14:55.425711 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-vz4qv" podStartSLOduration=2.358486848 podStartE2EDuration="4.425676787s" podCreationTimestamp="2026-02-27 08:14:51 +0000 UTC" firstStartedPulling="2026-02-27 08:14:52.363005828 +0000 UTC m=+1550.216935826" lastFinishedPulling="2026-02-27 08:14:54.430195777 +0000 UTC m=+1552.284125765" observedRunningTime="2026-02-27 08:14:55.420095597 +0000 UTC m=+1553.274025645" watchObservedRunningTime="2026-02-27 08:14:55.425676787 +0000 UTC m=+1553.279606785" Feb 27 08:14:56 crc kubenswrapper[4612]: I0227 08:14:56.411344 4612 generic.go:334] "Generic (PLEG): container finished" podID="6480a60b-9c50-4712-a5d1-fd2fbce66d76" containerID="3b21934250f871b4d50db07138b3cf64d39d8db01c10cb56254fc6c8b3ae0cdf" exitCode=0 Feb 27 08:14:56 crc kubenswrapper[4612]: I0227 08:14:56.411407 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4jr57" event={"ID":"6480a60b-9c50-4712-a5d1-fd2fbce66d76","Type":"ContainerDied","Data":"3b21934250f871b4d50db07138b3cf64d39d8db01c10cb56254fc6c8b3ae0cdf"} Feb 27 08:14:56 crc kubenswrapper[4612]: I0227 08:14:56.963033 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4jr57" Feb 27 08:14:57 crc kubenswrapper[4612]: I0227 08:14:57.029889 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t4t88\" (UniqueName: \"kubernetes.io/projected/6480a60b-9c50-4712-a5d1-fd2fbce66d76-kube-api-access-t4t88\") pod \"6480a60b-9c50-4712-a5d1-fd2fbce66d76\" (UID: \"6480a60b-9c50-4712-a5d1-fd2fbce66d76\") " Feb 27 08:14:57 crc kubenswrapper[4612]: I0227 08:14:57.029975 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6480a60b-9c50-4712-a5d1-fd2fbce66d76-catalog-content\") pod \"6480a60b-9c50-4712-a5d1-fd2fbce66d76\" (UID: \"6480a60b-9c50-4712-a5d1-fd2fbce66d76\") " Feb 27 08:14:57 crc kubenswrapper[4612]: I0227 08:14:57.030036 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6480a60b-9c50-4712-a5d1-fd2fbce66d76-utilities\") pod \"6480a60b-9c50-4712-a5d1-fd2fbce66d76\" (UID: \"6480a60b-9c50-4712-a5d1-fd2fbce66d76\") " Feb 27 08:14:57 crc kubenswrapper[4612]: I0227 08:14:57.031141 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6480a60b-9c50-4712-a5d1-fd2fbce66d76-utilities" (OuterVolumeSpecName: "utilities") pod "6480a60b-9c50-4712-a5d1-fd2fbce66d76" (UID: "6480a60b-9c50-4712-a5d1-fd2fbce66d76"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 08:14:57 crc kubenswrapper[4612]: I0227 08:14:57.035792 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6480a60b-9c50-4712-a5d1-fd2fbce66d76-kube-api-access-t4t88" (OuterVolumeSpecName: "kube-api-access-t4t88") pod "6480a60b-9c50-4712-a5d1-fd2fbce66d76" (UID: "6480a60b-9c50-4712-a5d1-fd2fbce66d76"). InnerVolumeSpecName "kube-api-access-t4t88". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 08:14:57 crc kubenswrapper[4612]: I0227 08:14:57.084002 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6480a60b-9c50-4712-a5d1-fd2fbce66d76-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6480a60b-9c50-4712-a5d1-fd2fbce66d76" (UID: "6480a60b-9c50-4712-a5d1-fd2fbce66d76"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 08:14:57 crc kubenswrapper[4612]: I0227 08:14:57.131463 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t4t88\" (UniqueName: \"kubernetes.io/projected/6480a60b-9c50-4712-a5d1-fd2fbce66d76-kube-api-access-t4t88\") on node \"crc\" DevicePath \"\"" Feb 27 08:14:57 crc kubenswrapper[4612]: I0227 08:14:57.131754 4612 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6480a60b-9c50-4712-a5d1-fd2fbce66d76-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 27 08:14:57 crc kubenswrapper[4612]: I0227 08:14:57.131763 4612 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6480a60b-9c50-4712-a5d1-fd2fbce66d76-utilities\") on node \"crc\" DevicePath \"\"" Feb 27 08:14:57 crc kubenswrapper[4612]: I0227 08:14:57.423103 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4jr57" event={"ID":"6480a60b-9c50-4712-a5d1-fd2fbce66d76","Type":"ContainerDied","Data":"65be91a33b6af385983643cc0eeab71edfbe2b9764480698351db64086625ea4"} Feb 27 08:14:57 crc kubenswrapper[4612]: I0227 08:14:57.423164 4612 scope.go:117] "RemoveContainer" containerID="3b21934250f871b4d50db07138b3cf64d39d8db01c10cb56254fc6c8b3ae0cdf" Feb 27 08:14:57 crc kubenswrapper[4612]: I0227 08:14:57.423307 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4jr57" Feb 27 08:14:57 crc kubenswrapper[4612]: I0227 08:14:57.447901 4612 scope.go:117] "RemoveContainer" containerID="50d34184851535939b113195699f46c3ec15b6caaf5b1989188f231883e83e55" Feb 27 08:14:57 crc kubenswrapper[4612]: I0227 08:14:57.475720 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-4jr57"] Feb 27 08:14:57 crc kubenswrapper[4612]: I0227 08:14:57.482930 4612 scope.go:117] "RemoveContainer" containerID="4f7e6836df4a2bc335aa1e9c23777b23ddd1b3c2991e88ded6af8a900cc63692" Feb 27 08:14:57 crc kubenswrapper[4612]: I0227 08:14:57.485767 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-4jr57"] Feb 27 08:14:58 crc kubenswrapper[4612]: I0227 08:14:58.442942 4612 generic.go:334] "Generic (PLEG): container finished" podID="9995bc24-fa2b-4916-8e71-627b11a39de7" containerID="480f953375b6159c9d93766fa580287e708c9b765189fbdd73dcdc4a7c5bd4b1" exitCode=0 Feb 27 08:14:58 crc kubenswrapper[4612]: I0227 08:14:58.443046 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-vz4qv" event={"ID":"9995bc24-fa2b-4916-8e71-627b11a39de7","Type":"ContainerDied","Data":"480f953375b6159c9d93766fa580287e708c9b765189fbdd73dcdc4a7c5bd4b1"} Feb 27 08:14:58 crc kubenswrapper[4612]: I0227 08:14:58.876987 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6480a60b-9c50-4712-a5d1-fd2fbce66d76" path="/var/lib/kubelet/pods/6480a60b-9c50-4712-a5d1-fd2fbce66d76/volumes" Feb 27 08:14:59 crc kubenswrapper[4612]: I0227 08:14:59.916545 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-vz4qv" Feb 27 08:14:59 crc kubenswrapper[4612]: I0227 08:14:59.991825 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zjhvw\" (UniqueName: \"kubernetes.io/projected/9995bc24-fa2b-4916-8e71-627b11a39de7-kube-api-access-zjhvw\") pod \"9995bc24-fa2b-4916-8e71-627b11a39de7\" (UID: \"9995bc24-fa2b-4916-8e71-627b11a39de7\") " Feb 27 08:14:59 crc kubenswrapper[4612]: I0227 08:14:59.991914 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9995bc24-fa2b-4916-8e71-627b11a39de7-inventory\") pod \"9995bc24-fa2b-4916-8e71-627b11a39de7\" (UID: \"9995bc24-fa2b-4916-8e71-627b11a39de7\") " Feb 27 08:14:59 crc kubenswrapper[4612]: I0227 08:14:59.992047 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/9995bc24-fa2b-4916-8e71-627b11a39de7-ssh-key-openstack-edpm-ipam\") pod \"9995bc24-fa2b-4916-8e71-627b11a39de7\" (UID: \"9995bc24-fa2b-4916-8e71-627b11a39de7\") " Feb 27 08:15:00 crc kubenswrapper[4612]: I0227 08:15:00.002249 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9995bc24-fa2b-4916-8e71-627b11a39de7-kube-api-access-zjhvw" (OuterVolumeSpecName: "kube-api-access-zjhvw") pod "9995bc24-fa2b-4916-8e71-627b11a39de7" (UID: "9995bc24-fa2b-4916-8e71-627b11a39de7"). InnerVolumeSpecName "kube-api-access-zjhvw". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 08:15:00 crc kubenswrapper[4612]: I0227 08:15:00.021195 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9995bc24-fa2b-4916-8e71-627b11a39de7-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "9995bc24-fa2b-4916-8e71-627b11a39de7" (UID: "9995bc24-fa2b-4916-8e71-627b11a39de7"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:15:00 crc kubenswrapper[4612]: I0227 08:15:00.022783 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9995bc24-fa2b-4916-8e71-627b11a39de7-inventory" (OuterVolumeSpecName: "inventory") pod "9995bc24-fa2b-4916-8e71-627b11a39de7" (UID: "9995bc24-fa2b-4916-8e71-627b11a39de7"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:15:00 crc kubenswrapper[4612]: I0227 08:15:00.095426 4612 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9995bc24-fa2b-4916-8e71-627b11a39de7-inventory\") on node \"crc\" DevicePath \"\"" Feb 27 08:15:00 crc kubenswrapper[4612]: I0227 08:15:00.095459 4612 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/9995bc24-fa2b-4916-8e71-627b11a39de7-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 27 08:15:00 crc kubenswrapper[4612]: I0227 08:15:00.095470 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zjhvw\" (UniqueName: \"kubernetes.io/projected/9995bc24-fa2b-4916-8e71-627b11a39de7-kube-api-access-zjhvw\") on node \"crc\" DevicePath \"\"" Feb 27 08:15:00 crc kubenswrapper[4612]: I0227 08:15:00.141321 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29536335-7bn5c"] Feb 27 08:15:00 crc kubenswrapper[4612]: E0227 08:15:00.141844 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6480a60b-9c50-4712-a5d1-fd2fbce66d76" containerName="extract-content" Feb 27 08:15:00 crc kubenswrapper[4612]: I0227 08:15:00.141871 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="6480a60b-9c50-4712-a5d1-fd2fbce66d76" containerName="extract-content" Feb 27 08:15:00 crc kubenswrapper[4612]: E0227 08:15:00.141887 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6480a60b-9c50-4712-a5d1-fd2fbce66d76" containerName="extract-utilities" Feb 27 08:15:00 crc kubenswrapper[4612]: I0227 08:15:00.141896 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="6480a60b-9c50-4712-a5d1-fd2fbce66d76" containerName="extract-utilities" Feb 27 08:15:00 crc kubenswrapper[4612]: E0227 08:15:00.141912 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9995bc24-fa2b-4916-8e71-627b11a39de7" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Feb 27 08:15:00 crc kubenswrapper[4612]: I0227 08:15:00.141922 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="9995bc24-fa2b-4916-8e71-627b11a39de7" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Feb 27 08:15:00 crc kubenswrapper[4612]: E0227 08:15:00.141956 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6480a60b-9c50-4712-a5d1-fd2fbce66d76" containerName="registry-server" Feb 27 08:15:00 crc kubenswrapper[4612]: I0227 08:15:00.141964 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="6480a60b-9c50-4712-a5d1-fd2fbce66d76" containerName="registry-server" Feb 27 08:15:00 crc kubenswrapper[4612]: I0227 08:15:00.142232 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="6480a60b-9c50-4712-a5d1-fd2fbce66d76" containerName="registry-server" Feb 27 08:15:00 crc kubenswrapper[4612]: I0227 08:15:00.142264 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="9995bc24-fa2b-4916-8e71-627b11a39de7" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Feb 27 08:15:00 crc kubenswrapper[4612]: I0227 08:15:00.143257 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29536335-7bn5c" Feb 27 08:15:00 crc kubenswrapper[4612]: I0227 08:15:00.145992 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Feb 27 08:15:00 crc kubenswrapper[4612]: I0227 08:15:00.148016 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Feb 27 08:15:00 crc kubenswrapper[4612]: I0227 08:15:00.152386 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29536335-7bn5c"] Feb 27 08:15:00 crc kubenswrapper[4612]: I0227 08:15:00.197357 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4ce49ed5-459c-42b9-bc5b-55c3f9ec5b45-secret-volume\") pod \"collect-profiles-29536335-7bn5c\" (UID: \"4ce49ed5-459c-42b9-bc5b-55c3f9ec5b45\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29536335-7bn5c" Feb 27 08:15:00 crc kubenswrapper[4612]: I0227 08:15:00.197500 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7pj86\" (UniqueName: \"kubernetes.io/projected/4ce49ed5-459c-42b9-bc5b-55c3f9ec5b45-kube-api-access-7pj86\") pod \"collect-profiles-29536335-7bn5c\" (UID: \"4ce49ed5-459c-42b9-bc5b-55c3f9ec5b45\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29536335-7bn5c" Feb 27 08:15:00 crc kubenswrapper[4612]: I0227 08:15:00.197847 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4ce49ed5-459c-42b9-bc5b-55c3f9ec5b45-config-volume\") pod \"collect-profiles-29536335-7bn5c\" (UID: \"4ce49ed5-459c-42b9-bc5b-55c3f9ec5b45\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29536335-7bn5c" Feb 27 08:15:00 crc kubenswrapper[4612]: I0227 08:15:00.300252 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4ce49ed5-459c-42b9-bc5b-55c3f9ec5b45-config-volume\") pod \"collect-profiles-29536335-7bn5c\" (UID: \"4ce49ed5-459c-42b9-bc5b-55c3f9ec5b45\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29536335-7bn5c" Feb 27 08:15:00 crc kubenswrapper[4612]: I0227 08:15:00.300309 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4ce49ed5-459c-42b9-bc5b-55c3f9ec5b45-secret-volume\") pod \"collect-profiles-29536335-7bn5c\" (UID: \"4ce49ed5-459c-42b9-bc5b-55c3f9ec5b45\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29536335-7bn5c" Feb 27 08:15:00 crc kubenswrapper[4612]: I0227 08:15:00.300385 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7pj86\" (UniqueName: \"kubernetes.io/projected/4ce49ed5-459c-42b9-bc5b-55c3f9ec5b45-kube-api-access-7pj86\") pod \"collect-profiles-29536335-7bn5c\" (UID: \"4ce49ed5-459c-42b9-bc5b-55c3f9ec5b45\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29536335-7bn5c" Feb 27 08:15:00 crc kubenswrapper[4612]: I0227 08:15:00.301394 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4ce49ed5-459c-42b9-bc5b-55c3f9ec5b45-config-volume\") pod \"collect-profiles-29536335-7bn5c\" (UID: \"4ce49ed5-459c-42b9-bc5b-55c3f9ec5b45\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29536335-7bn5c" Feb 27 08:15:00 crc kubenswrapper[4612]: I0227 08:15:00.314674 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4ce49ed5-459c-42b9-bc5b-55c3f9ec5b45-secret-volume\") pod \"collect-profiles-29536335-7bn5c\" (UID: \"4ce49ed5-459c-42b9-bc5b-55c3f9ec5b45\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29536335-7bn5c" Feb 27 08:15:00 crc kubenswrapper[4612]: I0227 08:15:00.322372 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7pj86\" (UniqueName: \"kubernetes.io/projected/4ce49ed5-459c-42b9-bc5b-55c3f9ec5b45-kube-api-access-7pj86\") pod \"collect-profiles-29536335-7bn5c\" (UID: \"4ce49ed5-459c-42b9-bc5b-55c3f9ec5b45\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29536335-7bn5c" Feb 27 08:15:00 crc kubenswrapper[4612]: I0227 08:15:00.472974 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29536335-7bn5c" Feb 27 08:15:00 crc kubenswrapper[4612]: I0227 08:15:00.473852 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-vz4qv" event={"ID":"9995bc24-fa2b-4916-8e71-627b11a39de7","Type":"ContainerDied","Data":"8dd61a08a166fa66e7aef71fa8ecea4c247e21855cc24f8a217c104f58f1dea4"} Feb 27 08:15:00 crc kubenswrapper[4612]: I0227 08:15:00.473885 4612 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8dd61a08a166fa66e7aef71fa8ecea4c247e21855cc24f8a217c104f58f1dea4" Feb 27 08:15:00 crc kubenswrapper[4612]: I0227 08:15:00.473939 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-vz4qv" Feb 27 08:15:00 crc kubenswrapper[4612]: I0227 08:15:00.560733 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k4jmw"] Feb 27 08:15:00 crc kubenswrapper[4612]: I0227 08:15:00.562298 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k4jmw" Feb 27 08:15:00 crc kubenswrapper[4612]: I0227 08:15:00.564855 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 27 08:15:00 crc kubenswrapper[4612]: I0227 08:15:00.565350 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-2pwrb" Feb 27 08:15:00 crc kubenswrapper[4612]: I0227 08:15:00.565251 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 27 08:15:00 crc kubenswrapper[4612]: I0227 08:15:00.565942 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 27 08:15:00 crc kubenswrapper[4612]: I0227 08:15:00.582119 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k4jmw"] Feb 27 08:15:00 crc kubenswrapper[4612]: I0227 08:15:00.608945 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9caaee00-687f-47de-8336-859a249751f7-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-k4jmw\" (UID: \"9caaee00-687f-47de-8336-859a249751f7\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k4jmw" Feb 27 08:15:00 crc kubenswrapper[4612]: I0227 08:15:00.609031 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/9caaee00-687f-47de-8336-859a249751f7-ssh-key-openstack-edpm-ipam\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-k4jmw\" (UID: \"9caaee00-687f-47de-8336-859a249751f7\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k4jmw" Feb 27 08:15:00 crc kubenswrapper[4612]: I0227 08:15:00.609106 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9caaee00-687f-47de-8336-859a249751f7-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-k4jmw\" (UID: \"9caaee00-687f-47de-8336-859a249751f7\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k4jmw" Feb 27 08:15:00 crc kubenswrapper[4612]: I0227 08:15:00.609123 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rmbqf\" (UniqueName: \"kubernetes.io/projected/9caaee00-687f-47de-8336-859a249751f7-kube-api-access-rmbqf\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-k4jmw\" (UID: \"9caaee00-687f-47de-8336-859a249751f7\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k4jmw" Feb 27 08:15:00 crc kubenswrapper[4612]: I0227 08:15:00.709821 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/9caaee00-687f-47de-8336-859a249751f7-ssh-key-openstack-edpm-ipam\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-k4jmw\" (UID: \"9caaee00-687f-47de-8336-859a249751f7\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k4jmw" Feb 27 08:15:00 crc kubenswrapper[4612]: I0227 08:15:00.710116 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9caaee00-687f-47de-8336-859a249751f7-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-k4jmw\" (UID: \"9caaee00-687f-47de-8336-859a249751f7\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k4jmw" Feb 27 08:15:00 crc kubenswrapper[4612]: I0227 08:15:00.710139 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rmbqf\" (UniqueName: \"kubernetes.io/projected/9caaee00-687f-47de-8336-859a249751f7-kube-api-access-rmbqf\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-k4jmw\" (UID: \"9caaee00-687f-47de-8336-859a249751f7\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k4jmw" Feb 27 08:15:00 crc kubenswrapper[4612]: I0227 08:15:00.710218 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9caaee00-687f-47de-8336-859a249751f7-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-k4jmw\" (UID: \"9caaee00-687f-47de-8336-859a249751f7\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k4jmw" Feb 27 08:15:00 crc kubenswrapper[4612]: I0227 08:15:00.715566 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/9caaee00-687f-47de-8336-859a249751f7-ssh-key-openstack-edpm-ipam\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-k4jmw\" (UID: \"9caaee00-687f-47de-8336-859a249751f7\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k4jmw" Feb 27 08:15:00 crc kubenswrapper[4612]: I0227 08:15:00.716157 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9caaee00-687f-47de-8336-859a249751f7-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-k4jmw\" (UID: \"9caaee00-687f-47de-8336-859a249751f7\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k4jmw" Feb 27 08:15:00 crc kubenswrapper[4612]: I0227 08:15:00.727883 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9caaee00-687f-47de-8336-859a249751f7-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-k4jmw\" (UID: \"9caaee00-687f-47de-8336-859a249751f7\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k4jmw" Feb 27 08:15:00 crc kubenswrapper[4612]: I0227 08:15:00.731152 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rmbqf\" (UniqueName: \"kubernetes.io/projected/9caaee00-687f-47de-8336-859a249751f7-kube-api-access-rmbqf\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-k4jmw\" (UID: \"9caaee00-687f-47de-8336-859a249751f7\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k4jmw" Feb 27 08:15:00 crc kubenswrapper[4612]: I0227 08:15:00.892940 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k4jmw" Feb 27 08:15:01 crc kubenswrapper[4612]: I0227 08:15:00.998415 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29536335-7bn5c"] Feb 27 08:15:01 crc kubenswrapper[4612]: W0227 08:15:01.029528 4612 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4ce49ed5_459c_42b9_bc5b_55c3f9ec5b45.slice/crio-cdc634c45c629da9bcd55429dd521f9168618d4d318675e47a9c725e35a9fc3c WatchSource:0}: Error finding container cdc634c45c629da9bcd55429dd521f9168618d4d318675e47a9c725e35a9fc3c: Status 404 returned error can't find the container with id cdc634c45c629da9bcd55429dd521f9168618d4d318675e47a9c725e35a9fc3c Feb 27 08:15:01 crc kubenswrapper[4612]: I0227 08:15:01.453965 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k4jmw"] Feb 27 08:15:01 crc kubenswrapper[4612]: W0227 08:15:01.460834 4612 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9caaee00_687f_47de_8336_859a249751f7.slice/crio-dbdcbfa8f96224d25f81c22813a82307508a22670ff9d133f683b20a3c6943bc WatchSource:0}: Error finding container dbdcbfa8f96224d25f81c22813a82307508a22670ff9d133f683b20a3c6943bc: Status 404 returned error can't find the container with id dbdcbfa8f96224d25f81c22813a82307508a22670ff9d133f683b20a3c6943bc Feb 27 08:15:01 crc kubenswrapper[4612]: I0227 08:15:01.486381 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k4jmw" event={"ID":"9caaee00-687f-47de-8336-859a249751f7","Type":"ContainerStarted","Data":"dbdcbfa8f96224d25f81c22813a82307508a22670ff9d133f683b20a3c6943bc"} Feb 27 08:15:01 crc kubenswrapper[4612]: I0227 08:15:01.488619 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29536335-7bn5c" event={"ID":"4ce49ed5-459c-42b9-bc5b-55c3f9ec5b45","Type":"ContainerStarted","Data":"c749d475d5e71926c5a54a3210fb4522047e7ef2624c67d3201f8856bfcb8b67"} Feb 27 08:15:01 crc kubenswrapper[4612]: I0227 08:15:01.488647 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29536335-7bn5c" event={"ID":"4ce49ed5-459c-42b9-bc5b-55c3f9ec5b45","Type":"ContainerStarted","Data":"cdc634c45c629da9bcd55429dd521f9168618d4d318675e47a9c725e35a9fc3c"} Feb 27 08:15:01 crc kubenswrapper[4612]: I0227 08:15:01.521204 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29536335-7bn5c" podStartSLOduration=1.5211823679999998 podStartE2EDuration="1.521182368s" podCreationTimestamp="2026-02-27 08:15:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 08:15:01.50627067 +0000 UTC m=+1559.360200678" watchObservedRunningTime="2026-02-27 08:15:01.521182368 +0000 UTC m=+1559.375112366" Feb 27 08:15:02 crc kubenswrapper[4612]: I0227 08:15:02.499273 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k4jmw" event={"ID":"9caaee00-687f-47de-8336-859a249751f7","Type":"ContainerStarted","Data":"1f5ff8e7d901ac9bf87a06c056cdb03c0f9fcbde59a3140e29a56905eae9fbe8"} Feb 27 08:15:02 crc kubenswrapper[4612]: I0227 08:15:02.503228 4612 generic.go:334] "Generic (PLEG): container finished" podID="4ce49ed5-459c-42b9-bc5b-55c3f9ec5b45" containerID="c749d475d5e71926c5a54a3210fb4522047e7ef2624c67d3201f8856bfcb8b67" exitCode=0 Feb 27 08:15:02 crc kubenswrapper[4612]: I0227 08:15:02.503347 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29536335-7bn5c" event={"ID":"4ce49ed5-459c-42b9-bc5b-55c3f9ec5b45","Type":"ContainerDied","Data":"c749d475d5e71926c5a54a3210fb4522047e7ef2624c67d3201f8856bfcb8b67"} Feb 27 08:15:02 crc kubenswrapper[4612]: I0227 08:15:02.567713 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k4jmw" podStartSLOduration=2.431531803 podStartE2EDuration="2.567671113s" podCreationTimestamp="2026-02-27 08:15:00 +0000 UTC" firstStartedPulling="2026-02-27 08:15:01.46414634 +0000 UTC m=+1559.318076328" lastFinishedPulling="2026-02-27 08:15:01.60028565 +0000 UTC m=+1559.454215638" observedRunningTime="2026-02-27 08:15:02.532036459 +0000 UTC m=+1560.385966547" watchObservedRunningTime="2026-02-27 08:15:02.567671113 +0000 UTC m=+1560.421601121" Feb 27 08:15:03 crc kubenswrapper[4612]: I0227 08:15:03.886959 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29536335-7bn5c" Feb 27 08:15:04 crc kubenswrapper[4612]: I0227 08:15:04.069238 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4ce49ed5-459c-42b9-bc5b-55c3f9ec5b45-secret-volume\") pod \"4ce49ed5-459c-42b9-bc5b-55c3f9ec5b45\" (UID: \"4ce49ed5-459c-42b9-bc5b-55c3f9ec5b45\") " Feb 27 08:15:04 crc kubenswrapper[4612]: I0227 08:15:04.069662 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7pj86\" (UniqueName: \"kubernetes.io/projected/4ce49ed5-459c-42b9-bc5b-55c3f9ec5b45-kube-api-access-7pj86\") pod \"4ce49ed5-459c-42b9-bc5b-55c3f9ec5b45\" (UID: \"4ce49ed5-459c-42b9-bc5b-55c3f9ec5b45\") " Feb 27 08:15:04 crc kubenswrapper[4612]: I0227 08:15:04.070060 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4ce49ed5-459c-42b9-bc5b-55c3f9ec5b45-config-volume\") pod \"4ce49ed5-459c-42b9-bc5b-55c3f9ec5b45\" (UID: \"4ce49ed5-459c-42b9-bc5b-55c3f9ec5b45\") " Feb 27 08:15:04 crc kubenswrapper[4612]: I0227 08:15:04.070988 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4ce49ed5-459c-42b9-bc5b-55c3f9ec5b45-config-volume" (OuterVolumeSpecName: "config-volume") pod "4ce49ed5-459c-42b9-bc5b-55c3f9ec5b45" (UID: "4ce49ed5-459c-42b9-bc5b-55c3f9ec5b45"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 08:15:04 crc kubenswrapper[4612]: I0227 08:15:04.076974 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4ce49ed5-459c-42b9-bc5b-55c3f9ec5b45-kube-api-access-7pj86" (OuterVolumeSpecName: "kube-api-access-7pj86") pod "4ce49ed5-459c-42b9-bc5b-55c3f9ec5b45" (UID: "4ce49ed5-459c-42b9-bc5b-55c3f9ec5b45"). InnerVolumeSpecName "kube-api-access-7pj86". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 08:15:04 crc kubenswrapper[4612]: I0227 08:15:04.078320 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4ce49ed5-459c-42b9-bc5b-55c3f9ec5b45-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "4ce49ed5-459c-42b9-bc5b-55c3f9ec5b45" (UID: "4ce49ed5-459c-42b9-bc5b-55c3f9ec5b45"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:15:04 crc kubenswrapper[4612]: I0227 08:15:04.171378 4612 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4ce49ed5-459c-42b9-bc5b-55c3f9ec5b45-secret-volume\") on node \"crc\" DevicePath \"\"" Feb 27 08:15:04 crc kubenswrapper[4612]: I0227 08:15:04.171423 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7pj86\" (UniqueName: \"kubernetes.io/projected/4ce49ed5-459c-42b9-bc5b-55c3f9ec5b45-kube-api-access-7pj86\") on node \"crc\" DevicePath \"\"" Feb 27 08:15:04 crc kubenswrapper[4612]: I0227 08:15:04.171435 4612 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4ce49ed5-459c-42b9-bc5b-55c3f9ec5b45-config-volume\") on node \"crc\" DevicePath \"\"" Feb 27 08:15:04 crc kubenswrapper[4612]: I0227 08:15:04.527675 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29536335-7bn5c" event={"ID":"4ce49ed5-459c-42b9-bc5b-55c3f9ec5b45","Type":"ContainerDied","Data":"cdc634c45c629da9bcd55429dd521f9168618d4d318675e47a9c725e35a9fc3c"} Feb 27 08:15:04 crc kubenswrapper[4612]: I0227 08:15:04.527737 4612 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cdc634c45c629da9bcd55429dd521f9168618d4d318675e47a9c725e35a9fc3c" Feb 27 08:15:04 crc kubenswrapper[4612]: I0227 08:15:04.527751 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29536335-7bn5c" Feb 27 08:15:14 crc kubenswrapper[4612]: I0227 08:15:14.943714 4612 scope.go:117] "RemoveContainer" containerID="69d83402d6290299d70278d5befc9ca664f942eed422c9ff4db495ca44b8976c" Feb 27 08:15:14 crc kubenswrapper[4612]: I0227 08:15:14.985206 4612 scope.go:117] "RemoveContainer" containerID="edd43cdd132a714448da88e522b53d539c8bad75e13f40e8d8ad5319c1936c48" Feb 27 08:15:15 crc kubenswrapper[4612]: I0227 08:15:15.019110 4612 scope.go:117] "RemoveContainer" containerID="d4bbac865331908abcaa583aff895c0dfe51bae301a0ff6fa4b2062f92b46c45" Feb 27 08:15:15 crc kubenswrapper[4612]: I0227 08:15:15.044611 4612 scope.go:117] "RemoveContainer" containerID="9a40f17b5d0d9218b64aab79491862655274f46636610ed2f3fac9e5e60ebdf4" Feb 27 08:15:16 crc kubenswrapper[4612]: I0227 08:15:16.026592 4612 patch_prober.go:28] interesting pod/machine-config-daemon-924vb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 27 08:15:16 crc kubenswrapper[4612]: I0227 08:15:16.026645 4612 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 27 08:15:16 crc kubenswrapper[4612]: I0227 08:15:16.026684 4612 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-924vb" Feb 27 08:15:16 crc kubenswrapper[4612]: I0227 08:15:16.027422 4612 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"7c62ebf970b4c78c5d31a0ce6596e9b398afbc8a6935511ce1b20bcd4f436b08"} pod="openshift-machine-config-operator/machine-config-daemon-924vb" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 27 08:15:16 crc kubenswrapper[4612]: I0227 08:15:16.027479 4612 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" containerName="machine-config-daemon" containerID="cri-o://7c62ebf970b4c78c5d31a0ce6596e9b398afbc8a6935511ce1b20bcd4f436b08" gracePeriod=600 Feb 27 08:15:16 crc kubenswrapper[4612]: E0227 08:15:16.152345 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-924vb_openshift-machine-config-operator(24599373-7adc-4a1b-8bb4-797bf726f43d)\"" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" Feb 27 08:15:16 crc kubenswrapper[4612]: I0227 08:15:16.736245 4612 generic.go:334] "Generic (PLEG): container finished" podID="24599373-7adc-4a1b-8bb4-797bf726f43d" containerID="7c62ebf970b4c78c5d31a0ce6596e9b398afbc8a6935511ce1b20bcd4f436b08" exitCode=0 Feb 27 08:15:16 crc kubenswrapper[4612]: I0227 08:15:16.736313 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-924vb" event={"ID":"24599373-7adc-4a1b-8bb4-797bf726f43d","Type":"ContainerDied","Data":"7c62ebf970b4c78c5d31a0ce6596e9b398afbc8a6935511ce1b20bcd4f436b08"} Feb 27 08:15:16 crc kubenswrapper[4612]: I0227 08:15:16.736356 4612 scope.go:117] "RemoveContainer" containerID="650e97edc5e74c75f461feb02d9fe0c4cbdce4ff0887e7ec2d2f50e6d2e7c100" Feb 27 08:15:16 crc kubenswrapper[4612]: I0227 08:15:16.737604 4612 scope.go:117] "RemoveContainer" containerID="7c62ebf970b4c78c5d31a0ce6596e9b398afbc8a6935511ce1b20bcd4f436b08" Feb 27 08:15:16 crc kubenswrapper[4612]: E0227 08:15:16.738213 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-924vb_openshift-machine-config-operator(24599373-7adc-4a1b-8bb4-797bf726f43d)\"" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" Feb 27 08:15:28 crc kubenswrapper[4612]: I0227 08:15:28.853364 4612 scope.go:117] "RemoveContainer" containerID="7c62ebf970b4c78c5d31a0ce6596e9b398afbc8a6935511ce1b20bcd4f436b08" Feb 27 08:15:28 crc kubenswrapper[4612]: E0227 08:15:28.854392 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-924vb_openshift-machine-config-operator(24599373-7adc-4a1b-8bb4-797bf726f43d)\"" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" Feb 27 08:15:42 crc kubenswrapper[4612]: I0227 08:15:42.865188 4612 scope.go:117] "RemoveContainer" containerID="7c62ebf970b4c78c5d31a0ce6596e9b398afbc8a6935511ce1b20bcd4f436b08" Feb 27 08:15:42 crc kubenswrapper[4612]: E0227 08:15:42.866059 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-924vb_openshift-machine-config-operator(24599373-7adc-4a1b-8bb4-797bf726f43d)\"" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" Feb 27 08:15:56 crc kubenswrapper[4612]: I0227 08:15:56.854182 4612 scope.go:117] "RemoveContainer" containerID="7c62ebf970b4c78c5d31a0ce6596e9b398afbc8a6935511ce1b20bcd4f436b08" Feb 27 08:15:56 crc kubenswrapper[4612]: E0227 08:15:56.855262 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-924vb_openshift-machine-config-operator(24599373-7adc-4a1b-8bb4-797bf726f43d)\"" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" Feb 27 08:16:00 crc kubenswrapper[4612]: I0227 08:16:00.178758 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29536336-49sn8"] Feb 27 08:16:00 crc kubenswrapper[4612]: E0227 08:16:00.179915 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ce49ed5-459c-42b9-bc5b-55c3f9ec5b45" containerName="collect-profiles" Feb 27 08:16:00 crc kubenswrapper[4612]: I0227 08:16:00.179938 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ce49ed5-459c-42b9-bc5b-55c3f9ec5b45" containerName="collect-profiles" Feb 27 08:16:00 crc kubenswrapper[4612]: I0227 08:16:00.180192 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="4ce49ed5-459c-42b9-bc5b-55c3f9ec5b45" containerName="collect-profiles" Feb 27 08:16:00 crc kubenswrapper[4612]: I0227 08:16:00.181035 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536336-49sn8" Feb 27 08:16:00 crc kubenswrapper[4612]: I0227 08:16:00.183487 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-5zwfx" Feb 27 08:16:00 crc kubenswrapper[4612]: I0227 08:16:00.183805 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 27 08:16:00 crc kubenswrapper[4612]: I0227 08:16:00.189126 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 27 08:16:00 crc kubenswrapper[4612]: I0227 08:16:00.196739 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536336-49sn8"] Feb 27 08:16:00 crc kubenswrapper[4612]: I0227 08:16:00.334955 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q4x58\" (UniqueName: \"kubernetes.io/projected/abd98e63-1fe3-4643-8df8-7512d3afb557-kube-api-access-q4x58\") pod \"auto-csr-approver-29536336-49sn8\" (UID: \"abd98e63-1fe3-4643-8df8-7512d3afb557\") " pod="openshift-infra/auto-csr-approver-29536336-49sn8" Feb 27 08:16:00 crc kubenswrapper[4612]: I0227 08:16:00.436644 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q4x58\" (UniqueName: \"kubernetes.io/projected/abd98e63-1fe3-4643-8df8-7512d3afb557-kube-api-access-q4x58\") pod \"auto-csr-approver-29536336-49sn8\" (UID: \"abd98e63-1fe3-4643-8df8-7512d3afb557\") " pod="openshift-infra/auto-csr-approver-29536336-49sn8" Feb 27 08:16:00 crc kubenswrapper[4612]: I0227 08:16:00.457312 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q4x58\" (UniqueName: \"kubernetes.io/projected/abd98e63-1fe3-4643-8df8-7512d3afb557-kube-api-access-q4x58\") pod \"auto-csr-approver-29536336-49sn8\" (UID: \"abd98e63-1fe3-4643-8df8-7512d3afb557\") " pod="openshift-infra/auto-csr-approver-29536336-49sn8" Feb 27 08:16:00 crc kubenswrapper[4612]: I0227 08:16:00.509014 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536336-49sn8" Feb 27 08:16:00 crc kubenswrapper[4612]: I0227 08:16:00.958957 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536336-49sn8"] Feb 27 08:16:01 crc kubenswrapper[4612]: I0227 08:16:01.157106 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536336-49sn8" event={"ID":"abd98e63-1fe3-4643-8df8-7512d3afb557","Type":"ContainerStarted","Data":"e3ce8fa25cd64f4ff6a6104d0fcbf86f769834cd6328c1c6ba37523f6050a85e"} Feb 27 08:16:02 crc kubenswrapper[4612]: I0227 08:16:02.169391 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536336-49sn8" event={"ID":"abd98e63-1fe3-4643-8df8-7512d3afb557","Type":"ContainerStarted","Data":"c58a73db1308d49e03d771c746d8a101c60cc387a0f267816a563635eb23794f"} Feb 27 08:16:02 crc kubenswrapper[4612]: I0227 08:16:02.192112 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29536336-49sn8" podStartSLOduration=1.32903114 podStartE2EDuration="2.192089527s" podCreationTimestamp="2026-02-27 08:16:00 +0000 UTC" firstStartedPulling="2026-02-27 08:16:00.960945369 +0000 UTC m=+1618.814875367" lastFinishedPulling="2026-02-27 08:16:01.824003756 +0000 UTC m=+1619.677933754" observedRunningTime="2026-02-27 08:16:02.182286976 +0000 UTC m=+1620.036216994" watchObservedRunningTime="2026-02-27 08:16:02.192089527 +0000 UTC m=+1620.046019535" Feb 27 08:16:03 crc kubenswrapper[4612]: I0227 08:16:03.180431 4612 generic.go:334] "Generic (PLEG): container finished" podID="abd98e63-1fe3-4643-8df8-7512d3afb557" containerID="c58a73db1308d49e03d771c746d8a101c60cc387a0f267816a563635eb23794f" exitCode=0 Feb 27 08:16:03 crc kubenswrapper[4612]: I0227 08:16:03.180506 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536336-49sn8" event={"ID":"abd98e63-1fe3-4643-8df8-7512d3afb557","Type":"ContainerDied","Data":"c58a73db1308d49e03d771c746d8a101c60cc387a0f267816a563635eb23794f"} Feb 27 08:16:04 crc kubenswrapper[4612]: I0227 08:16:04.533139 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536336-49sn8" Feb 27 08:16:04 crc kubenswrapper[4612]: I0227 08:16:04.624403 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q4x58\" (UniqueName: \"kubernetes.io/projected/abd98e63-1fe3-4643-8df8-7512d3afb557-kube-api-access-q4x58\") pod \"abd98e63-1fe3-4643-8df8-7512d3afb557\" (UID: \"abd98e63-1fe3-4643-8df8-7512d3afb557\") " Feb 27 08:16:04 crc kubenswrapper[4612]: I0227 08:16:04.642871 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/abd98e63-1fe3-4643-8df8-7512d3afb557-kube-api-access-q4x58" (OuterVolumeSpecName: "kube-api-access-q4x58") pod "abd98e63-1fe3-4643-8df8-7512d3afb557" (UID: "abd98e63-1fe3-4643-8df8-7512d3afb557"). InnerVolumeSpecName "kube-api-access-q4x58". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 08:16:04 crc kubenswrapper[4612]: I0227 08:16:04.727197 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q4x58\" (UniqueName: \"kubernetes.io/projected/abd98e63-1fe3-4643-8df8-7512d3afb557-kube-api-access-q4x58\") on node \"crc\" DevicePath \"\"" Feb 27 08:16:05 crc kubenswrapper[4612]: I0227 08:16:05.203492 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536336-49sn8" event={"ID":"abd98e63-1fe3-4643-8df8-7512d3afb557","Type":"ContainerDied","Data":"e3ce8fa25cd64f4ff6a6104d0fcbf86f769834cd6328c1c6ba37523f6050a85e"} Feb 27 08:16:05 crc kubenswrapper[4612]: I0227 08:16:05.203545 4612 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e3ce8fa25cd64f4ff6a6104d0fcbf86f769834cd6328c1c6ba37523f6050a85e" Feb 27 08:16:05 crc kubenswrapper[4612]: I0227 08:16:05.203548 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536336-49sn8" Feb 27 08:16:05 crc kubenswrapper[4612]: I0227 08:16:05.251975 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29536330-nck2q"] Feb 27 08:16:05 crc kubenswrapper[4612]: I0227 08:16:05.260148 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29536330-nck2q"] Feb 27 08:16:06 crc kubenswrapper[4612]: I0227 08:16:06.865001 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="073232e9-6f4e-409e-bf29-989b354fcc39" path="/var/lib/kubelet/pods/073232e9-6f4e-409e-bf29-989b354fcc39/volumes" Feb 27 08:16:10 crc kubenswrapper[4612]: I0227 08:16:10.853380 4612 scope.go:117] "RemoveContainer" containerID="7c62ebf970b4c78c5d31a0ce6596e9b398afbc8a6935511ce1b20bcd4f436b08" Feb 27 08:16:10 crc kubenswrapper[4612]: E0227 08:16:10.856648 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-924vb_openshift-machine-config-operator(24599373-7adc-4a1b-8bb4-797bf726f43d)\"" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" Feb 27 08:16:15 crc kubenswrapper[4612]: I0227 08:16:15.176661 4612 scope.go:117] "RemoveContainer" containerID="dd37a24177c3f58809092c5a7984ecc67bec7440f6050f16ab9c5a4322a5fe8a" Feb 27 08:16:21 crc kubenswrapper[4612]: I0227 08:16:21.853463 4612 scope.go:117] "RemoveContainer" containerID="7c62ebf970b4c78c5d31a0ce6596e9b398afbc8a6935511ce1b20bcd4f436b08" Feb 27 08:16:21 crc kubenswrapper[4612]: E0227 08:16:21.855911 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-924vb_openshift-machine-config-operator(24599373-7adc-4a1b-8bb4-797bf726f43d)\"" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" Feb 27 08:16:32 crc kubenswrapper[4612]: I0227 08:16:32.866543 4612 scope.go:117] "RemoveContainer" containerID="7c62ebf970b4c78c5d31a0ce6596e9b398afbc8a6935511ce1b20bcd4f436b08" Feb 27 08:16:32 crc kubenswrapper[4612]: E0227 08:16:32.869596 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-924vb_openshift-machine-config-operator(24599373-7adc-4a1b-8bb4-797bf726f43d)\"" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" Feb 27 08:16:44 crc kubenswrapper[4612]: I0227 08:16:44.853579 4612 scope.go:117] "RemoveContainer" containerID="7c62ebf970b4c78c5d31a0ce6596e9b398afbc8a6935511ce1b20bcd4f436b08" Feb 27 08:16:44 crc kubenswrapper[4612]: E0227 08:16:44.854554 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-924vb_openshift-machine-config-operator(24599373-7adc-4a1b-8bb4-797bf726f43d)\"" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" Feb 27 08:16:56 crc kubenswrapper[4612]: I0227 08:16:56.853322 4612 scope.go:117] "RemoveContainer" containerID="7c62ebf970b4c78c5d31a0ce6596e9b398afbc8a6935511ce1b20bcd4f436b08" Feb 27 08:16:56 crc kubenswrapper[4612]: E0227 08:16:56.855287 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-924vb_openshift-machine-config-operator(24599373-7adc-4a1b-8bb4-797bf726f43d)\"" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" Feb 27 08:17:11 crc kubenswrapper[4612]: I0227 08:17:11.854429 4612 scope.go:117] "RemoveContainer" containerID="7c62ebf970b4c78c5d31a0ce6596e9b398afbc8a6935511ce1b20bcd4f436b08" Feb 27 08:17:11 crc kubenswrapper[4612]: E0227 08:17:11.857592 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-924vb_openshift-machine-config-operator(24599373-7adc-4a1b-8bb4-797bf726f43d)\"" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" Feb 27 08:17:15 crc kubenswrapper[4612]: I0227 08:17:15.271894 4612 scope.go:117] "RemoveContainer" containerID="370921685f3a020371fa0b5634f3b819b4505e8822283b03a1a586a24f9229e8" Feb 27 08:17:15 crc kubenswrapper[4612]: I0227 08:17:15.298493 4612 scope.go:117] "RemoveContainer" containerID="2b830865135697d89e47f2d5e4890e88bf23bb1fe36455912bcdc5ea5f0f65c1" Feb 27 08:17:26 crc kubenswrapper[4612]: I0227 08:17:26.854861 4612 scope.go:117] "RemoveContainer" containerID="7c62ebf970b4c78c5d31a0ce6596e9b398afbc8a6935511ce1b20bcd4f436b08" Feb 27 08:17:26 crc kubenswrapper[4612]: E0227 08:17:26.855862 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-924vb_openshift-machine-config-operator(24599373-7adc-4a1b-8bb4-797bf726f43d)\"" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" Feb 27 08:17:39 crc kubenswrapper[4612]: I0227 08:17:39.853728 4612 scope.go:117] "RemoveContainer" containerID="7c62ebf970b4c78c5d31a0ce6596e9b398afbc8a6935511ce1b20bcd4f436b08" Feb 27 08:17:39 crc kubenswrapper[4612]: E0227 08:17:39.854593 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-924vb_openshift-machine-config-operator(24599373-7adc-4a1b-8bb4-797bf726f43d)\"" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" Feb 27 08:17:53 crc kubenswrapper[4612]: I0227 08:17:53.852809 4612 scope.go:117] "RemoveContainer" containerID="7c62ebf970b4c78c5d31a0ce6596e9b398afbc8a6935511ce1b20bcd4f436b08" Feb 27 08:17:53 crc kubenswrapper[4612]: E0227 08:17:53.853713 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-924vb_openshift-machine-config-operator(24599373-7adc-4a1b-8bb4-797bf726f43d)\"" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" Feb 27 08:18:00 crc kubenswrapper[4612]: I0227 08:18:00.155071 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29536338-smcsj"] Feb 27 08:18:00 crc kubenswrapper[4612]: E0227 08:18:00.156144 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="abd98e63-1fe3-4643-8df8-7512d3afb557" containerName="oc" Feb 27 08:18:00 crc kubenswrapper[4612]: I0227 08:18:00.156163 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="abd98e63-1fe3-4643-8df8-7512d3afb557" containerName="oc" Feb 27 08:18:00 crc kubenswrapper[4612]: I0227 08:18:00.156411 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="abd98e63-1fe3-4643-8df8-7512d3afb557" containerName="oc" Feb 27 08:18:00 crc kubenswrapper[4612]: I0227 08:18:00.157142 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536338-smcsj" Feb 27 08:18:00 crc kubenswrapper[4612]: I0227 08:18:00.159860 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 27 08:18:00 crc kubenswrapper[4612]: I0227 08:18:00.159979 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-5zwfx" Feb 27 08:18:00 crc kubenswrapper[4612]: I0227 08:18:00.162539 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 27 08:18:00 crc kubenswrapper[4612]: I0227 08:18:00.177198 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536338-smcsj"] Feb 27 08:18:00 crc kubenswrapper[4612]: I0227 08:18:00.274767 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gf82l\" (UniqueName: \"kubernetes.io/projected/452831e0-7617-4e8b-bb21-5740a2129fcf-kube-api-access-gf82l\") pod \"auto-csr-approver-29536338-smcsj\" (UID: \"452831e0-7617-4e8b-bb21-5740a2129fcf\") " pod="openshift-infra/auto-csr-approver-29536338-smcsj" Feb 27 08:18:00 crc kubenswrapper[4612]: I0227 08:18:00.377161 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gf82l\" (UniqueName: \"kubernetes.io/projected/452831e0-7617-4e8b-bb21-5740a2129fcf-kube-api-access-gf82l\") pod \"auto-csr-approver-29536338-smcsj\" (UID: \"452831e0-7617-4e8b-bb21-5740a2129fcf\") " pod="openshift-infra/auto-csr-approver-29536338-smcsj" Feb 27 08:18:00 crc kubenswrapper[4612]: I0227 08:18:00.397960 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gf82l\" (UniqueName: \"kubernetes.io/projected/452831e0-7617-4e8b-bb21-5740a2129fcf-kube-api-access-gf82l\") pod \"auto-csr-approver-29536338-smcsj\" (UID: \"452831e0-7617-4e8b-bb21-5740a2129fcf\") " pod="openshift-infra/auto-csr-approver-29536338-smcsj" Feb 27 08:18:00 crc kubenswrapper[4612]: I0227 08:18:00.479518 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536338-smcsj" Feb 27 08:18:01 crc kubenswrapper[4612]: I0227 08:18:01.026778 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536338-smcsj"] Feb 27 08:18:01 crc kubenswrapper[4612]: I0227 08:18:01.538199 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536338-smcsj" event={"ID":"452831e0-7617-4e8b-bb21-5740a2129fcf","Type":"ContainerStarted","Data":"941d7f2edba814a1f0d21e5797966c46e4909f5f00fb209310e56b13c1a8d27c"} Feb 27 08:18:02 crc kubenswrapper[4612]: I0227 08:18:02.552259 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536338-smcsj" event={"ID":"452831e0-7617-4e8b-bb21-5740a2129fcf","Type":"ContainerStarted","Data":"85cacf47d720e93ee6405b6fe2f8eccf305b7741069ec86301c754f80389434a"} Feb 27 08:18:02 crc kubenswrapper[4612]: I0227 08:18:02.580813 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29536338-smcsj" podStartSLOduration=1.626337179 podStartE2EDuration="2.580795441s" podCreationTimestamp="2026-02-27 08:18:00 +0000 UTC" firstStartedPulling="2026-02-27 08:18:01.03376646 +0000 UTC m=+1738.887696468" lastFinishedPulling="2026-02-27 08:18:01.988224692 +0000 UTC m=+1739.842154730" observedRunningTime="2026-02-27 08:18:02.578173645 +0000 UTC m=+1740.432103643" watchObservedRunningTime="2026-02-27 08:18:02.580795441 +0000 UTC m=+1740.434725439" Feb 27 08:18:03 crc kubenswrapper[4612]: I0227 08:18:03.569066 4612 generic.go:334] "Generic (PLEG): container finished" podID="452831e0-7617-4e8b-bb21-5740a2129fcf" containerID="85cacf47d720e93ee6405b6fe2f8eccf305b7741069ec86301c754f80389434a" exitCode=0 Feb 27 08:18:03 crc kubenswrapper[4612]: I0227 08:18:03.569139 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536338-smcsj" event={"ID":"452831e0-7617-4e8b-bb21-5740a2129fcf","Type":"ContainerDied","Data":"85cacf47d720e93ee6405b6fe2f8eccf305b7741069ec86301c754f80389434a"} Feb 27 08:18:04 crc kubenswrapper[4612]: I0227 08:18:04.972832 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536338-smcsj" Feb 27 08:18:04 crc kubenswrapper[4612]: I0227 08:18:04.975758 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf82l\" (UniqueName: \"kubernetes.io/projected/452831e0-7617-4e8b-bb21-5740a2129fcf-kube-api-access-gf82l\") pod \"452831e0-7617-4e8b-bb21-5740a2129fcf\" (UID: \"452831e0-7617-4e8b-bb21-5740a2129fcf\") " Feb 27 08:18:04 crc kubenswrapper[4612]: I0227 08:18:04.981635 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/452831e0-7617-4e8b-bb21-5740a2129fcf-kube-api-access-gf82l" (OuterVolumeSpecName: "kube-api-access-gf82l") pod "452831e0-7617-4e8b-bb21-5740a2129fcf" (UID: "452831e0-7617-4e8b-bb21-5740a2129fcf"). InnerVolumeSpecName "kube-api-access-gf82l". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 08:18:05 crc kubenswrapper[4612]: I0227 08:18:05.078303 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf82l\" (UniqueName: \"kubernetes.io/projected/452831e0-7617-4e8b-bb21-5740a2129fcf-kube-api-access-gf82l\") on node \"crc\" DevicePath \"\"" Feb 27 08:18:05 crc kubenswrapper[4612]: I0227 08:18:05.589055 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536338-smcsj" event={"ID":"452831e0-7617-4e8b-bb21-5740a2129fcf","Type":"ContainerDied","Data":"941d7f2edba814a1f0d21e5797966c46e4909f5f00fb209310e56b13c1a8d27c"} Feb 27 08:18:05 crc kubenswrapper[4612]: I0227 08:18:05.589407 4612 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="941d7f2edba814a1f0d21e5797966c46e4909f5f00fb209310e56b13c1a8d27c" Feb 27 08:18:05 crc kubenswrapper[4612]: I0227 08:18:05.589106 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536338-smcsj" Feb 27 08:18:05 crc kubenswrapper[4612]: I0227 08:18:05.657605 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29536332-8qjw9"] Feb 27 08:18:05 crc kubenswrapper[4612]: I0227 08:18:05.665831 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29536332-8qjw9"] Feb 27 08:18:06 crc kubenswrapper[4612]: I0227 08:18:06.853943 4612 scope.go:117] "RemoveContainer" containerID="7c62ebf970b4c78c5d31a0ce6596e9b398afbc8a6935511ce1b20bcd4f436b08" Feb 27 08:18:06 crc kubenswrapper[4612]: E0227 08:18:06.855026 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-924vb_openshift-machine-config-operator(24599373-7adc-4a1b-8bb4-797bf726f43d)\"" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" Feb 27 08:18:06 crc kubenswrapper[4612]: I0227 08:18:06.874262 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cf3edace-2d20-49a4-95fc-ff3e04fafb77" path="/var/lib/kubelet/pods/cf3edace-2d20-49a4-95fc-ff3e04fafb77/volumes" Feb 27 08:18:08 crc kubenswrapper[4612]: I0227 08:18:08.032394 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-a904-account-create-update-hsnsq"] Feb 27 08:18:08 crc kubenswrapper[4612]: I0227 08:18:08.047261 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-t9l25"] Feb 27 08:18:08 crc kubenswrapper[4612]: I0227 08:18:08.057104 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-a904-account-create-update-hsnsq"] Feb 27 08:18:08 crc kubenswrapper[4612]: I0227 08:18:08.065478 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-t9l25"] Feb 27 08:18:08 crc kubenswrapper[4612]: I0227 08:18:08.866881 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="395f27e6-3047-48b3-a63f-7931bd177ccb" path="/var/lib/kubelet/pods/395f27e6-3047-48b3-a63f-7931bd177ccb/volumes" Feb 27 08:18:08 crc kubenswrapper[4612]: I0227 08:18:08.868404 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="48835d1f-acd8-41db-a37f-7a82a08561bc" path="/var/lib/kubelet/pods/48835d1f-acd8-41db-a37f-7a82a08561bc/volumes" Feb 27 08:18:13 crc kubenswrapper[4612]: I0227 08:18:13.053235 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-r47j8"] Feb 27 08:18:13 crc kubenswrapper[4612]: I0227 08:18:13.075183 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-lkbpz"] Feb 27 08:18:13 crc kubenswrapper[4612]: I0227 08:18:13.085816 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-lkbpz"] Feb 27 08:18:13 crc kubenswrapper[4612]: I0227 08:18:13.094813 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-r47j8"] Feb 27 08:18:14 crc kubenswrapper[4612]: I0227 08:18:14.862938 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1551171a-913f-4d99-b4da-148ea454f7f6" path="/var/lib/kubelet/pods/1551171a-913f-4d99-b4da-148ea454f7f6/volumes" Feb 27 08:18:14 crc kubenswrapper[4612]: I0227 08:18:14.864122 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22a1a3d2-6c5d-4102-8fd4-07f953e402f7" path="/var/lib/kubelet/pods/22a1a3d2-6c5d-4102-8fd4-07f953e402f7/volumes" Feb 27 08:18:15 crc kubenswrapper[4612]: I0227 08:18:15.039791 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-7819-account-create-update-lppdh"] Feb 27 08:18:15 crc kubenswrapper[4612]: I0227 08:18:15.070799 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-3e89-account-create-update-crnl5"] Feb 27 08:18:15 crc kubenswrapper[4612]: I0227 08:18:15.088333 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-7819-account-create-update-lppdh"] Feb 27 08:18:15 crc kubenswrapper[4612]: I0227 08:18:15.100602 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-3e89-account-create-update-crnl5"] Feb 27 08:18:15 crc kubenswrapper[4612]: I0227 08:18:15.363752 4612 scope.go:117] "RemoveContainer" containerID="37544462aa650601b257f91f9bb8064b10303d0824e3aecc659444fad5ec504e" Feb 27 08:18:15 crc kubenswrapper[4612]: I0227 08:18:15.403323 4612 scope.go:117] "RemoveContainer" containerID="22a8e7a514287d732f249a859cb167853f917dc5c1a0c1423aea971bfb518228" Feb 27 08:18:15 crc kubenswrapper[4612]: I0227 08:18:15.466346 4612 scope.go:117] "RemoveContainer" containerID="ccaff13d789c8a216b77ec37c3f21f72a4ebf4890273ea56836de2332d1e0623" Feb 27 08:18:15 crc kubenswrapper[4612]: I0227 08:18:15.493871 4612 scope.go:117] "RemoveContainer" containerID="97ff21863c2120601b515d9bf34ce20c3286d9f973a71672a8251e430cf070dd" Feb 27 08:18:15 crc kubenswrapper[4612]: I0227 08:18:15.534473 4612 scope.go:117] "RemoveContainer" containerID="1355afc2e630e6ea544c146adeee1caff5bf48d29a747aab4df9750c639e1d92" Feb 27 08:18:16 crc kubenswrapper[4612]: I0227 08:18:16.874199 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="03d154b1-4114-4c56-82b0-5bf148511e5f" path="/var/lib/kubelet/pods/03d154b1-4114-4c56-82b0-5bf148511e5f/volumes" Feb 27 08:18:16 crc kubenswrapper[4612]: I0227 08:18:16.877482 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ad4524a-e07e-4a6e-93b1-5e5e2d7b251c" path="/var/lib/kubelet/pods/6ad4524a-e07e-4a6e-93b1-5e5e2d7b251c/volumes" Feb 27 08:18:18 crc kubenswrapper[4612]: I0227 08:18:18.853145 4612 scope.go:117] "RemoveContainer" containerID="7c62ebf970b4c78c5d31a0ce6596e9b398afbc8a6935511ce1b20bcd4f436b08" Feb 27 08:18:18 crc kubenswrapper[4612]: E0227 08:18:18.853648 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-924vb_openshift-machine-config-operator(24599373-7adc-4a1b-8bb4-797bf726f43d)\"" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" Feb 27 08:18:24 crc kubenswrapper[4612]: I0227 08:18:24.795532 4612 generic.go:334] "Generic (PLEG): container finished" podID="9caaee00-687f-47de-8336-859a249751f7" containerID="1f5ff8e7d901ac9bf87a06c056cdb03c0f9fcbde59a3140e29a56905eae9fbe8" exitCode=0 Feb 27 08:18:24 crc kubenswrapper[4612]: I0227 08:18:24.795611 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k4jmw" event={"ID":"9caaee00-687f-47de-8336-859a249751f7","Type":"ContainerDied","Data":"1f5ff8e7d901ac9bf87a06c056cdb03c0f9fcbde59a3140e29a56905eae9fbe8"} Feb 27 08:18:26 crc kubenswrapper[4612]: I0227 08:18:26.275104 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k4jmw" Feb 27 08:18:26 crc kubenswrapper[4612]: I0227 08:18:26.356547 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9caaee00-687f-47de-8336-859a249751f7-inventory\") pod \"9caaee00-687f-47de-8336-859a249751f7\" (UID: \"9caaee00-687f-47de-8336-859a249751f7\") " Feb 27 08:18:26 crc kubenswrapper[4612]: I0227 08:18:26.356707 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/9caaee00-687f-47de-8336-859a249751f7-ssh-key-openstack-edpm-ipam\") pod \"9caaee00-687f-47de-8336-859a249751f7\" (UID: \"9caaee00-687f-47de-8336-859a249751f7\") " Feb 27 08:18:26 crc kubenswrapper[4612]: I0227 08:18:26.356797 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9caaee00-687f-47de-8336-859a249751f7-bootstrap-combined-ca-bundle\") pod \"9caaee00-687f-47de-8336-859a249751f7\" (UID: \"9caaee00-687f-47de-8336-859a249751f7\") " Feb 27 08:18:26 crc kubenswrapper[4612]: I0227 08:18:26.356838 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rmbqf\" (UniqueName: \"kubernetes.io/projected/9caaee00-687f-47de-8336-859a249751f7-kube-api-access-rmbqf\") pod \"9caaee00-687f-47de-8336-859a249751f7\" (UID: \"9caaee00-687f-47de-8336-859a249751f7\") " Feb 27 08:18:26 crc kubenswrapper[4612]: I0227 08:18:26.362192 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9caaee00-687f-47de-8336-859a249751f7-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "9caaee00-687f-47de-8336-859a249751f7" (UID: "9caaee00-687f-47de-8336-859a249751f7"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:18:26 crc kubenswrapper[4612]: I0227 08:18:26.368899 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9caaee00-687f-47de-8336-859a249751f7-kube-api-access-rmbqf" (OuterVolumeSpecName: "kube-api-access-rmbqf") pod "9caaee00-687f-47de-8336-859a249751f7" (UID: "9caaee00-687f-47de-8336-859a249751f7"). InnerVolumeSpecName "kube-api-access-rmbqf". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 08:18:26 crc kubenswrapper[4612]: I0227 08:18:26.395088 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9caaee00-687f-47de-8336-859a249751f7-inventory" (OuterVolumeSpecName: "inventory") pod "9caaee00-687f-47de-8336-859a249751f7" (UID: "9caaee00-687f-47de-8336-859a249751f7"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:18:26 crc kubenswrapper[4612]: I0227 08:18:26.396255 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9caaee00-687f-47de-8336-859a249751f7-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "9caaee00-687f-47de-8336-859a249751f7" (UID: "9caaee00-687f-47de-8336-859a249751f7"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:18:26 crc kubenswrapper[4612]: I0227 08:18:26.459149 4612 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/9caaee00-687f-47de-8336-859a249751f7-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 27 08:18:26 crc kubenswrapper[4612]: I0227 08:18:26.459197 4612 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9caaee00-687f-47de-8336-859a249751f7-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 08:18:26 crc kubenswrapper[4612]: I0227 08:18:26.459216 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rmbqf\" (UniqueName: \"kubernetes.io/projected/9caaee00-687f-47de-8336-859a249751f7-kube-api-access-rmbqf\") on node \"crc\" DevicePath \"\"" Feb 27 08:18:26 crc kubenswrapper[4612]: I0227 08:18:26.459236 4612 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9caaee00-687f-47de-8336-859a249751f7-inventory\") on node \"crc\" DevicePath \"\"" Feb 27 08:18:26 crc kubenswrapper[4612]: I0227 08:18:26.820336 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k4jmw" event={"ID":"9caaee00-687f-47de-8336-859a249751f7","Type":"ContainerDied","Data":"dbdcbfa8f96224d25f81c22813a82307508a22670ff9d133f683b20a3c6943bc"} Feb 27 08:18:26 crc kubenswrapper[4612]: I0227 08:18:26.820398 4612 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dbdcbfa8f96224d25f81c22813a82307508a22670ff9d133f683b20a3c6943bc" Feb 27 08:18:26 crc kubenswrapper[4612]: I0227 08:18:26.820480 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k4jmw" Feb 27 08:18:26 crc kubenswrapper[4612]: I0227 08:18:26.957861 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-t4m6d"] Feb 27 08:18:26 crc kubenswrapper[4612]: E0227 08:18:26.958317 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9caaee00-687f-47de-8336-859a249751f7" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Feb 27 08:18:26 crc kubenswrapper[4612]: I0227 08:18:26.958344 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="9caaee00-687f-47de-8336-859a249751f7" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Feb 27 08:18:26 crc kubenswrapper[4612]: E0227 08:18:26.958362 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="452831e0-7617-4e8b-bb21-5740a2129fcf" containerName="oc" Feb 27 08:18:26 crc kubenswrapper[4612]: I0227 08:18:26.958371 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="452831e0-7617-4e8b-bb21-5740a2129fcf" containerName="oc" Feb 27 08:18:26 crc kubenswrapper[4612]: I0227 08:18:26.958614 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="452831e0-7617-4e8b-bb21-5740a2129fcf" containerName="oc" Feb 27 08:18:26 crc kubenswrapper[4612]: I0227 08:18:26.958637 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="9caaee00-687f-47de-8336-859a249751f7" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Feb 27 08:18:26 crc kubenswrapper[4612]: I0227 08:18:26.959663 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-t4m6d" Feb 27 08:18:26 crc kubenswrapper[4612]: I0227 08:18:26.964161 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 27 08:18:26 crc kubenswrapper[4612]: I0227 08:18:26.964474 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-2pwrb" Feb 27 08:18:26 crc kubenswrapper[4612]: I0227 08:18:26.966608 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 27 08:18:26 crc kubenswrapper[4612]: I0227 08:18:26.966851 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 27 08:18:26 crc kubenswrapper[4612]: I0227 08:18:26.972270 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-t4m6d"] Feb 27 08:18:27 crc kubenswrapper[4612]: I0227 08:18:27.073804 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-44sqh\" (UniqueName: \"kubernetes.io/projected/8439abb9-0c80-44f9-9ec4-ef0c99a50182-kube-api-access-44sqh\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-t4m6d\" (UID: \"8439abb9-0c80-44f9-9ec4-ef0c99a50182\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-t4m6d" Feb 27 08:18:27 crc kubenswrapper[4612]: I0227 08:18:27.074427 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/8439abb9-0c80-44f9-9ec4-ef0c99a50182-ssh-key-openstack-edpm-ipam\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-t4m6d\" (UID: \"8439abb9-0c80-44f9-9ec4-ef0c99a50182\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-t4m6d" Feb 27 08:18:27 crc kubenswrapper[4612]: I0227 08:18:27.074740 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8439abb9-0c80-44f9-9ec4-ef0c99a50182-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-t4m6d\" (UID: \"8439abb9-0c80-44f9-9ec4-ef0c99a50182\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-t4m6d" Feb 27 08:18:27 crc kubenswrapper[4612]: I0227 08:18:27.176891 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-44sqh\" (UniqueName: \"kubernetes.io/projected/8439abb9-0c80-44f9-9ec4-ef0c99a50182-kube-api-access-44sqh\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-t4m6d\" (UID: \"8439abb9-0c80-44f9-9ec4-ef0c99a50182\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-t4m6d" Feb 27 08:18:27 crc kubenswrapper[4612]: I0227 08:18:27.177213 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/8439abb9-0c80-44f9-9ec4-ef0c99a50182-ssh-key-openstack-edpm-ipam\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-t4m6d\" (UID: \"8439abb9-0c80-44f9-9ec4-ef0c99a50182\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-t4m6d" Feb 27 08:18:27 crc kubenswrapper[4612]: I0227 08:18:27.177409 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8439abb9-0c80-44f9-9ec4-ef0c99a50182-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-t4m6d\" (UID: \"8439abb9-0c80-44f9-9ec4-ef0c99a50182\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-t4m6d" Feb 27 08:18:27 crc kubenswrapper[4612]: I0227 08:18:27.182752 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/8439abb9-0c80-44f9-9ec4-ef0c99a50182-ssh-key-openstack-edpm-ipam\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-t4m6d\" (UID: \"8439abb9-0c80-44f9-9ec4-ef0c99a50182\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-t4m6d" Feb 27 08:18:27 crc kubenswrapper[4612]: I0227 08:18:27.184580 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8439abb9-0c80-44f9-9ec4-ef0c99a50182-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-t4m6d\" (UID: \"8439abb9-0c80-44f9-9ec4-ef0c99a50182\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-t4m6d" Feb 27 08:18:27 crc kubenswrapper[4612]: I0227 08:18:27.199197 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-44sqh\" (UniqueName: \"kubernetes.io/projected/8439abb9-0c80-44f9-9ec4-ef0c99a50182-kube-api-access-44sqh\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-t4m6d\" (UID: \"8439abb9-0c80-44f9-9ec4-ef0c99a50182\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-t4m6d" Feb 27 08:18:27 crc kubenswrapper[4612]: I0227 08:18:27.291244 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-t4m6d" Feb 27 08:18:27 crc kubenswrapper[4612]: W0227 08:18:27.853998 4612 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8439abb9_0c80_44f9_9ec4_ef0c99a50182.slice/crio-500d89d9d87dba228578d380846411e6d04943d9dfa9cea3de810aad52c046ff WatchSource:0}: Error finding container 500d89d9d87dba228578d380846411e6d04943d9dfa9cea3de810aad52c046ff: Status 404 returned error can't find the container with id 500d89d9d87dba228578d380846411e6d04943d9dfa9cea3de810aad52c046ff Feb 27 08:18:27 crc kubenswrapper[4612]: I0227 08:18:27.857447 4612 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 27 08:18:27 crc kubenswrapper[4612]: I0227 08:18:27.864210 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-t4m6d"] Feb 27 08:18:28 crc kubenswrapper[4612]: I0227 08:18:28.879782 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-t4m6d" event={"ID":"8439abb9-0c80-44f9-9ec4-ef0c99a50182","Type":"ContainerStarted","Data":"6637a81f0fe7aef324f61694cfd1ae75d9784086810b13888d2058f9f98bd97f"} Feb 27 08:18:28 crc kubenswrapper[4612]: I0227 08:18:28.880217 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-t4m6d" event={"ID":"8439abb9-0c80-44f9-9ec4-ef0c99a50182","Type":"ContainerStarted","Data":"500d89d9d87dba228578d380846411e6d04943d9dfa9cea3de810aad52c046ff"} Feb 27 08:18:28 crc kubenswrapper[4612]: I0227 08:18:28.889496 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-t4m6d" podStartSLOduration=2.716997643 podStartE2EDuration="2.889469697s" podCreationTimestamp="2026-02-27 08:18:26 +0000 UTC" firstStartedPulling="2026-02-27 08:18:27.857251082 +0000 UTC m=+1765.711181080" lastFinishedPulling="2026-02-27 08:18:28.029723126 +0000 UTC m=+1765.883653134" observedRunningTime="2026-02-27 08:18:28.879872192 +0000 UTC m=+1766.733802230" watchObservedRunningTime="2026-02-27 08:18:28.889469697 +0000 UTC m=+1766.743399705" Feb 27 08:18:29 crc kubenswrapper[4612]: I0227 08:18:29.047996 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/root-account-create-update-khpvr"] Feb 27 08:18:29 crc kubenswrapper[4612]: I0227 08:18:29.063591 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/root-account-create-update-khpvr"] Feb 27 08:18:30 crc kubenswrapper[4612]: I0227 08:18:30.883498 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="998cf80a-2cd6-4869-8967-5b0261ad2dd4" path="/var/lib/kubelet/pods/998cf80a-2cd6-4869-8967-5b0261ad2dd4/volumes" Feb 27 08:18:32 crc kubenswrapper[4612]: I0227 08:18:32.853508 4612 scope.go:117] "RemoveContainer" containerID="7c62ebf970b4c78c5d31a0ce6596e9b398afbc8a6935511ce1b20bcd4f436b08" Feb 27 08:18:32 crc kubenswrapper[4612]: E0227 08:18:32.854056 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-924vb_openshift-machine-config-operator(24599373-7adc-4a1b-8bb4-797bf726f43d)\"" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" Feb 27 08:18:42 crc kubenswrapper[4612]: I0227 08:18:42.057625 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-4ndtf"] Feb 27 08:18:42 crc kubenswrapper[4612]: I0227 08:18:42.069421 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-4ndtf"] Feb 27 08:18:42 crc kubenswrapper[4612]: I0227 08:18:42.865339 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ca30e354-dd13-4fa2-96a9-bbd6c605b1f8" path="/var/lib/kubelet/pods/ca30e354-dd13-4fa2-96a9-bbd6c605b1f8/volumes" Feb 27 08:18:44 crc kubenswrapper[4612]: I0227 08:18:44.853104 4612 scope.go:117] "RemoveContainer" containerID="7c62ebf970b4c78c5d31a0ce6596e9b398afbc8a6935511ce1b20bcd4f436b08" Feb 27 08:18:44 crc kubenswrapper[4612]: E0227 08:18:44.853602 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-924vb_openshift-machine-config-operator(24599373-7adc-4a1b-8bb4-797bf726f43d)\"" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" Feb 27 08:18:53 crc kubenswrapper[4612]: I0227 08:18:53.045387 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-qg67v"] Feb 27 08:18:53 crc kubenswrapper[4612]: I0227 08:18:53.057859 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-845d9"] Feb 27 08:18:53 crc kubenswrapper[4612]: I0227 08:18:53.072739 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-qg67v"] Feb 27 08:18:53 crc kubenswrapper[4612]: I0227 08:18:53.080607 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-bdfb-account-create-update-8rtvc"] Feb 27 08:18:53 crc kubenswrapper[4612]: I0227 08:18:53.092867 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-k4hmr"] Feb 27 08:18:53 crc kubenswrapper[4612]: I0227 08:18:53.100448 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-845d9"] Feb 27 08:18:53 crc kubenswrapper[4612]: I0227 08:18:53.107078 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-bdfb-account-create-update-8rtvc"] Feb 27 08:18:53 crc kubenswrapper[4612]: I0227 08:18:53.114082 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-3e28-account-create-update-m8rzl"] Feb 27 08:18:53 crc kubenswrapper[4612]: I0227 08:18:53.121293 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-k4hmr"] Feb 27 08:18:53 crc kubenswrapper[4612]: I0227 08:18:53.128071 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-3e28-account-create-update-m8rzl"] Feb 27 08:18:53 crc kubenswrapper[4612]: I0227 08:18:53.134456 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-a95b-account-create-update-wb47g"] Feb 27 08:18:53 crc kubenswrapper[4612]: I0227 08:18:53.140663 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-a95b-account-create-update-wb47g"] Feb 27 08:18:54 crc kubenswrapper[4612]: I0227 08:18:54.885787 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="04e60e3e-bfd4-4936-be0a-d6688312aeae" path="/var/lib/kubelet/pods/04e60e3e-bfd4-4936-be0a-d6688312aeae/volumes" Feb 27 08:18:54 crc kubenswrapper[4612]: I0227 08:18:54.888775 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4e0957b6-31f8-4664-96d9-f75bba20f23f" path="/var/lib/kubelet/pods/4e0957b6-31f8-4664-96d9-f75bba20f23f/volumes" Feb 27 08:18:54 crc kubenswrapper[4612]: I0227 08:18:54.890609 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9708a624-a4ef-4250-b3b1-2afb6b56b5c7" path="/var/lib/kubelet/pods/9708a624-a4ef-4250-b3b1-2afb6b56b5c7/volumes" Feb 27 08:18:54 crc kubenswrapper[4612]: I0227 08:18:54.892352 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c01bd6e3-e867-4506-ab09-75433d751574" path="/var/lib/kubelet/pods/c01bd6e3-e867-4506-ab09-75433d751574/volumes" Feb 27 08:18:54 crc kubenswrapper[4612]: I0227 08:18:54.895339 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cc36db3f-167a-47e6-807e-11234c86957a" path="/var/lib/kubelet/pods/cc36db3f-167a-47e6-807e-11234c86957a/volumes" Feb 27 08:18:54 crc kubenswrapper[4612]: I0227 08:18:54.898298 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fb8bfa80-78a7-46e9-acf5-d99fec1cc79c" path="/var/lib/kubelet/pods/fb8bfa80-78a7-46e9-acf5-d99fec1cc79c/volumes" Feb 27 08:18:55 crc kubenswrapper[4612]: I0227 08:18:55.852999 4612 scope.go:117] "RemoveContainer" containerID="7c62ebf970b4c78c5d31a0ce6596e9b398afbc8a6935511ce1b20bcd4f436b08" Feb 27 08:18:55 crc kubenswrapper[4612]: E0227 08:18:55.853495 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-924vb_openshift-machine-config-operator(24599373-7adc-4a1b-8bb4-797bf726f43d)\"" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" Feb 27 08:18:58 crc kubenswrapper[4612]: I0227 08:18:58.035978 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-75fb5"] Feb 27 08:18:58 crc kubenswrapper[4612]: I0227 08:18:58.046602 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-75fb5"] Feb 27 08:18:58 crc kubenswrapper[4612]: I0227 08:18:58.881871 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c1e74a83-8d08-428a-b4d7-9a0c60ca4b5e" path="/var/lib/kubelet/pods/c1e74a83-8d08-428a-b4d7-9a0c60ca4b5e/volumes" Feb 27 08:19:09 crc kubenswrapper[4612]: I0227 08:19:09.852984 4612 scope.go:117] "RemoveContainer" containerID="7c62ebf970b4c78c5d31a0ce6596e9b398afbc8a6935511ce1b20bcd4f436b08" Feb 27 08:19:09 crc kubenswrapper[4612]: E0227 08:19:09.853902 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-924vb_openshift-machine-config-operator(24599373-7adc-4a1b-8bb4-797bf726f43d)\"" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" Feb 27 08:19:15 crc kubenswrapper[4612]: I0227 08:19:15.655982 4612 scope.go:117] "RemoveContainer" containerID="da0aba518068f84f43b535ddebc9b0c49194bc74f0acf0222d57286ffda800fe" Feb 27 08:19:15 crc kubenswrapper[4612]: I0227 08:19:15.691215 4612 scope.go:117] "RemoveContainer" containerID="584b11abbdbbc425cc040fab20d9160a8e7f632e67f8118a75cea1518c3225d7" Feb 27 08:19:15 crc kubenswrapper[4612]: I0227 08:19:15.764832 4612 scope.go:117] "RemoveContainer" containerID="ff707c5ca0ead88846d488396fedd3b9b7a434450cb61ed96e59f642dbef8562" Feb 27 08:19:15 crc kubenswrapper[4612]: I0227 08:19:15.806795 4612 scope.go:117] "RemoveContainer" containerID="377a280bda8254e93f4617a17b0bf8037e8524f573ac151b55eb7e717386282e" Feb 27 08:19:15 crc kubenswrapper[4612]: I0227 08:19:15.850472 4612 scope.go:117] "RemoveContainer" containerID="2b820b796e8e94cbe25ea8ec156015b0693cd025f7c7aef8eeb8a0a9e12ebfea" Feb 27 08:19:15 crc kubenswrapper[4612]: I0227 08:19:15.886293 4612 scope.go:117] "RemoveContainer" containerID="7293eec8c40e09308544339c5238b1ec2a2fda76a771a6a729697572a772cb65" Feb 27 08:19:15 crc kubenswrapper[4612]: I0227 08:19:15.924334 4612 scope.go:117] "RemoveContainer" containerID="b11b057c5ff66e783eb8beacc2972c189703b113319ce71704cf6108ad5c60d9" Feb 27 08:19:15 crc kubenswrapper[4612]: I0227 08:19:15.947756 4612 scope.go:117] "RemoveContainer" containerID="ff6163e0bc97f75223abc25f0842bbda086dc65a63133377d47d41c542503c2f" Feb 27 08:19:15 crc kubenswrapper[4612]: I0227 08:19:15.970510 4612 scope.go:117] "RemoveContainer" containerID="8ea58c74730729970fcdaa1932750137c2afe61af32804267541e8e86631b0b8" Feb 27 08:19:15 crc kubenswrapper[4612]: I0227 08:19:15.992088 4612 scope.go:117] "RemoveContainer" containerID="4dd551c703fe08d89ae29f3a19ab83cb1dc49449642c22653f097f4a9e22bfe3" Feb 27 08:19:16 crc kubenswrapper[4612]: I0227 08:19:16.015553 4612 scope.go:117] "RemoveContainer" containerID="a5a94c230f575a9cf9c7daa80ed04a668e262a0a6dd0b39f379d608f8b63fe86" Feb 27 08:19:23 crc kubenswrapper[4612]: I0227 08:19:23.853182 4612 scope.go:117] "RemoveContainer" containerID="7c62ebf970b4c78c5d31a0ce6596e9b398afbc8a6935511ce1b20bcd4f436b08" Feb 27 08:19:23 crc kubenswrapper[4612]: E0227 08:19:23.854574 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-924vb_openshift-machine-config-operator(24599373-7adc-4a1b-8bb4-797bf726f43d)\"" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" Feb 27 08:19:34 crc kubenswrapper[4612]: I0227 08:19:34.092447 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-4ljbd"] Feb 27 08:19:34 crc kubenswrapper[4612]: I0227 08:19:34.108063 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-4ljbd"] Feb 27 08:19:34 crc kubenswrapper[4612]: I0227 08:19:34.853261 4612 scope.go:117] "RemoveContainer" containerID="7c62ebf970b4c78c5d31a0ce6596e9b398afbc8a6935511ce1b20bcd4f436b08" Feb 27 08:19:34 crc kubenswrapper[4612]: E0227 08:19:34.853893 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-924vb_openshift-machine-config-operator(24599373-7adc-4a1b-8bb4-797bf726f43d)\"" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" Feb 27 08:19:34 crc kubenswrapper[4612]: I0227 08:19:34.865647 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f63a4b12-f7e9-43a4-91d2-fd46f65daf92" path="/var/lib/kubelet/pods/f63a4b12-f7e9-43a4-91d2-fd46f65daf92/volumes" Feb 27 08:19:46 crc kubenswrapper[4612]: I0227 08:19:46.853260 4612 scope.go:117] "RemoveContainer" containerID="7c62ebf970b4c78c5d31a0ce6596e9b398afbc8a6935511ce1b20bcd4f436b08" Feb 27 08:19:46 crc kubenswrapper[4612]: E0227 08:19:46.854166 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-924vb_openshift-machine-config-operator(24599373-7adc-4a1b-8bb4-797bf726f43d)\"" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" Feb 27 08:19:51 crc kubenswrapper[4612]: I0227 08:19:51.100977 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-s5bjv"] Feb 27 08:19:51 crc kubenswrapper[4612]: I0227 08:19:51.117799 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-r8dpp"] Feb 27 08:19:51 crc kubenswrapper[4612]: I0227 08:19:51.128562 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-r8dpp"] Feb 27 08:19:51 crc kubenswrapper[4612]: I0227 08:19:51.139036 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-s5bjv"] Feb 27 08:19:51 crc kubenswrapper[4612]: I0227 08:19:51.147092 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-8zrcl"] Feb 27 08:19:51 crc kubenswrapper[4612]: I0227 08:19:51.156280 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-8zrcl"] Feb 27 08:19:52 crc kubenswrapper[4612]: I0227 08:19:52.880992 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1a05a7e5-8c46-441a-9faf-0fd3b41bbd98" path="/var/lib/kubelet/pods/1a05a7e5-8c46-441a-9faf-0fd3b41bbd98/volumes" Feb 27 08:19:52 crc kubenswrapper[4612]: I0227 08:19:52.885079 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="36b549a3-c6e9-4f7e-8cff-85216c8e0e3d" path="/var/lib/kubelet/pods/36b549a3-c6e9-4f7e-8cff-85216c8e0e3d/volumes" Feb 27 08:19:52 crc kubenswrapper[4612]: I0227 08:19:52.886756 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87df2a96-7ae3-47af-92bb-97db038b957c" path="/var/lib/kubelet/pods/87df2a96-7ae3-47af-92bb-97db038b957c/volumes" Feb 27 08:19:58 crc kubenswrapper[4612]: I0227 08:19:58.853994 4612 scope.go:117] "RemoveContainer" containerID="7c62ebf970b4c78c5d31a0ce6596e9b398afbc8a6935511ce1b20bcd4f436b08" Feb 27 08:19:58 crc kubenswrapper[4612]: E0227 08:19:58.854882 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-924vb_openshift-machine-config-operator(24599373-7adc-4a1b-8bb4-797bf726f43d)\"" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" Feb 27 08:20:00 crc kubenswrapper[4612]: I0227 08:20:00.177400 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29536340-2zmwk"] Feb 27 08:20:00 crc kubenswrapper[4612]: I0227 08:20:00.179215 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536340-2zmwk" Feb 27 08:20:00 crc kubenswrapper[4612]: I0227 08:20:00.183083 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 27 08:20:00 crc kubenswrapper[4612]: I0227 08:20:00.183875 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 27 08:20:00 crc kubenswrapper[4612]: I0227 08:20:00.185954 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-5zwfx" Feb 27 08:20:00 crc kubenswrapper[4612]: I0227 08:20:00.225148 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536340-2zmwk"] Feb 27 08:20:00 crc kubenswrapper[4612]: I0227 08:20:00.381224 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fw7fv\" (UniqueName: \"kubernetes.io/projected/ec08bd6b-d9ec-4b86-8397-9e1f709332ed-kube-api-access-fw7fv\") pod \"auto-csr-approver-29536340-2zmwk\" (UID: \"ec08bd6b-d9ec-4b86-8397-9e1f709332ed\") " pod="openshift-infra/auto-csr-approver-29536340-2zmwk" Feb 27 08:20:00 crc kubenswrapper[4612]: I0227 08:20:00.483411 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fw7fv\" (UniqueName: \"kubernetes.io/projected/ec08bd6b-d9ec-4b86-8397-9e1f709332ed-kube-api-access-fw7fv\") pod \"auto-csr-approver-29536340-2zmwk\" (UID: \"ec08bd6b-d9ec-4b86-8397-9e1f709332ed\") " pod="openshift-infra/auto-csr-approver-29536340-2zmwk" Feb 27 08:20:00 crc kubenswrapper[4612]: I0227 08:20:00.516832 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fw7fv\" (UniqueName: \"kubernetes.io/projected/ec08bd6b-d9ec-4b86-8397-9e1f709332ed-kube-api-access-fw7fv\") pod \"auto-csr-approver-29536340-2zmwk\" (UID: \"ec08bd6b-d9ec-4b86-8397-9e1f709332ed\") " pod="openshift-infra/auto-csr-approver-29536340-2zmwk" Feb 27 08:20:00 crc kubenswrapper[4612]: I0227 08:20:00.572495 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536340-2zmwk" Feb 27 08:20:01 crc kubenswrapper[4612]: I0227 08:20:01.030644 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536340-2zmwk"] Feb 27 08:20:01 crc kubenswrapper[4612]: W0227 08:20:01.031888 4612 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podec08bd6b_d9ec_4b86_8397_9e1f709332ed.slice/crio-f3d3bd344e6fdb5b310292d51e1f52aae55c02e896fae3d75c774fd7aba07200 WatchSource:0}: Error finding container f3d3bd344e6fdb5b310292d51e1f52aae55c02e896fae3d75c774fd7aba07200: Status 404 returned error can't find the container with id f3d3bd344e6fdb5b310292d51e1f52aae55c02e896fae3d75c774fd7aba07200 Feb 27 08:20:01 crc kubenswrapper[4612]: I0227 08:20:01.841615 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536340-2zmwk" event={"ID":"ec08bd6b-d9ec-4b86-8397-9e1f709332ed","Type":"ContainerStarted","Data":"f3d3bd344e6fdb5b310292d51e1f52aae55c02e896fae3d75c774fd7aba07200"} Feb 27 08:20:02 crc kubenswrapper[4612]: I0227 08:20:02.883895 4612 generic.go:334] "Generic (PLEG): container finished" podID="ec08bd6b-d9ec-4b86-8397-9e1f709332ed" containerID="66a4c28074c462834ed0ba2b9b203d2add0612d7747efadf8852535b8166500b" exitCode=0 Feb 27 08:20:02 crc kubenswrapper[4612]: I0227 08:20:02.885095 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536340-2zmwk" event={"ID":"ec08bd6b-d9ec-4b86-8397-9e1f709332ed","Type":"ContainerDied","Data":"66a4c28074c462834ed0ba2b9b203d2add0612d7747efadf8852535b8166500b"} Feb 27 08:20:04 crc kubenswrapper[4612]: I0227 08:20:04.245538 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536340-2zmwk" Feb 27 08:20:04 crc kubenswrapper[4612]: I0227 08:20:04.296326 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fw7fv\" (UniqueName: \"kubernetes.io/projected/ec08bd6b-d9ec-4b86-8397-9e1f709332ed-kube-api-access-fw7fv\") pod \"ec08bd6b-d9ec-4b86-8397-9e1f709332ed\" (UID: \"ec08bd6b-d9ec-4b86-8397-9e1f709332ed\") " Feb 27 08:20:04 crc kubenswrapper[4612]: I0227 08:20:04.304179 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ec08bd6b-d9ec-4b86-8397-9e1f709332ed-kube-api-access-fw7fv" (OuterVolumeSpecName: "kube-api-access-fw7fv") pod "ec08bd6b-d9ec-4b86-8397-9e1f709332ed" (UID: "ec08bd6b-d9ec-4b86-8397-9e1f709332ed"). InnerVolumeSpecName "kube-api-access-fw7fv". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 08:20:04 crc kubenswrapper[4612]: I0227 08:20:04.399501 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fw7fv\" (UniqueName: \"kubernetes.io/projected/ec08bd6b-d9ec-4b86-8397-9e1f709332ed-kube-api-access-fw7fv\") on node \"crc\" DevicePath \"\"" Feb 27 08:20:04 crc kubenswrapper[4612]: I0227 08:20:04.924898 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536340-2zmwk" event={"ID":"ec08bd6b-d9ec-4b86-8397-9e1f709332ed","Type":"ContainerDied","Data":"f3d3bd344e6fdb5b310292d51e1f52aae55c02e896fae3d75c774fd7aba07200"} Feb 27 08:20:04 crc kubenswrapper[4612]: I0227 08:20:04.925288 4612 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f3d3bd344e6fdb5b310292d51e1f52aae55c02e896fae3d75c774fd7aba07200" Feb 27 08:20:04 crc kubenswrapper[4612]: I0227 08:20:04.924955 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536340-2zmwk" Feb 27 08:20:05 crc kubenswrapper[4612]: I0227 08:20:05.301855 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29536334-bzlzh"] Feb 27 08:20:05 crc kubenswrapper[4612]: I0227 08:20:05.310354 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29536334-bzlzh"] Feb 27 08:20:06 crc kubenswrapper[4612]: I0227 08:20:06.876227 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="92a7b490-f9bd-4150-a650-56a3881ec95a" path="/var/lib/kubelet/pods/92a7b490-f9bd-4150-a650-56a3881ec95a/volumes" Feb 27 08:20:11 crc kubenswrapper[4612]: I0227 08:20:11.853651 4612 scope.go:117] "RemoveContainer" containerID="7c62ebf970b4c78c5d31a0ce6596e9b398afbc8a6935511ce1b20bcd4f436b08" Feb 27 08:20:11 crc kubenswrapper[4612]: E0227 08:20:11.854289 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-924vb_openshift-machine-config-operator(24599373-7adc-4a1b-8bb4-797bf726f43d)\"" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" Feb 27 08:20:13 crc kubenswrapper[4612]: I0227 08:20:13.044053 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-l5z5s"] Feb 27 08:20:13 crc kubenswrapper[4612]: I0227 08:20:13.053349 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-l5z5s"] Feb 27 08:20:14 crc kubenswrapper[4612]: I0227 08:20:14.866848 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0c9482e8-9eb1-4c30-95e9-bdaa5ccd239b" path="/var/lib/kubelet/pods/0c9482e8-9eb1-4c30-95e9-bdaa5ccd239b/volumes" Feb 27 08:20:16 crc kubenswrapper[4612]: I0227 08:20:16.236139 4612 scope.go:117] "RemoveContainer" containerID="a6d45103e694a574568b137e43fc7bcdcd0e18154a2c6b7ba6c90c509ea0da15" Feb 27 08:20:16 crc kubenswrapper[4612]: I0227 08:20:16.273475 4612 scope.go:117] "RemoveContainer" containerID="8b548933be8af10568a38da48000e3a51d2fe9ff3b04b577671c700e4196b621" Feb 27 08:20:16 crc kubenswrapper[4612]: I0227 08:20:16.339386 4612 scope.go:117] "RemoveContainer" containerID="a35ff58cdedc1d36f8345528a64eca1b073d88b1dbf4edfd2f202894c24e07b0" Feb 27 08:20:16 crc kubenswrapper[4612]: I0227 08:20:16.388247 4612 scope.go:117] "RemoveContainer" containerID="450cb4092f6cbf1bfa5907a300e2f111a1e01887f00e70a6fad596acd3495bb6" Feb 27 08:20:16 crc kubenswrapper[4612]: I0227 08:20:16.430408 4612 scope.go:117] "RemoveContainer" containerID="d7fbb0369e2f86e02466cb49c5f6a3bba32f32d2e771e988cd25937fb4d8dd80" Feb 27 08:20:16 crc kubenswrapper[4612]: I0227 08:20:16.470519 4612 scope.go:117] "RemoveContainer" containerID="7a14e07a854b7c34d66faacc2713fc09fcde83a58ea6b802875545f39c503506" Feb 27 08:20:18 crc kubenswrapper[4612]: I0227 08:20:18.066791 4612 generic.go:334] "Generic (PLEG): container finished" podID="8439abb9-0c80-44f9-9ec4-ef0c99a50182" containerID="6637a81f0fe7aef324f61694cfd1ae75d9784086810b13888d2058f9f98bd97f" exitCode=0 Feb 27 08:20:18 crc kubenswrapper[4612]: I0227 08:20:18.066879 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-t4m6d" event={"ID":"8439abb9-0c80-44f9-9ec4-ef0c99a50182","Type":"ContainerDied","Data":"6637a81f0fe7aef324f61694cfd1ae75d9784086810b13888d2058f9f98bd97f"} Feb 27 08:20:19 crc kubenswrapper[4612]: I0227 08:20:19.590430 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-t4m6d" Feb 27 08:20:19 crc kubenswrapper[4612]: I0227 08:20:19.733105 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/8439abb9-0c80-44f9-9ec4-ef0c99a50182-ssh-key-openstack-edpm-ipam\") pod \"8439abb9-0c80-44f9-9ec4-ef0c99a50182\" (UID: \"8439abb9-0c80-44f9-9ec4-ef0c99a50182\") " Feb 27 08:20:19 crc kubenswrapper[4612]: I0227 08:20:19.733929 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8439abb9-0c80-44f9-9ec4-ef0c99a50182-inventory\") pod \"8439abb9-0c80-44f9-9ec4-ef0c99a50182\" (UID: \"8439abb9-0c80-44f9-9ec4-ef0c99a50182\") " Feb 27 08:20:19 crc kubenswrapper[4612]: I0227 08:20:19.734107 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-44sqh\" (UniqueName: \"kubernetes.io/projected/8439abb9-0c80-44f9-9ec4-ef0c99a50182-kube-api-access-44sqh\") pod \"8439abb9-0c80-44f9-9ec4-ef0c99a50182\" (UID: \"8439abb9-0c80-44f9-9ec4-ef0c99a50182\") " Feb 27 08:20:19 crc kubenswrapper[4612]: I0227 08:20:19.745900 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8439abb9-0c80-44f9-9ec4-ef0c99a50182-kube-api-access-44sqh" (OuterVolumeSpecName: "kube-api-access-44sqh") pod "8439abb9-0c80-44f9-9ec4-ef0c99a50182" (UID: "8439abb9-0c80-44f9-9ec4-ef0c99a50182"). InnerVolumeSpecName "kube-api-access-44sqh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 08:20:19 crc kubenswrapper[4612]: I0227 08:20:19.764190 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8439abb9-0c80-44f9-9ec4-ef0c99a50182-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "8439abb9-0c80-44f9-9ec4-ef0c99a50182" (UID: "8439abb9-0c80-44f9-9ec4-ef0c99a50182"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:20:19 crc kubenswrapper[4612]: I0227 08:20:19.772313 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8439abb9-0c80-44f9-9ec4-ef0c99a50182-inventory" (OuterVolumeSpecName: "inventory") pod "8439abb9-0c80-44f9-9ec4-ef0c99a50182" (UID: "8439abb9-0c80-44f9-9ec4-ef0c99a50182"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:20:19 crc kubenswrapper[4612]: I0227 08:20:19.836094 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-44sqh\" (UniqueName: \"kubernetes.io/projected/8439abb9-0c80-44f9-9ec4-ef0c99a50182-kube-api-access-44sqh\") on node \"crc\" DevicePath \"\"" Feb 27 08:20:19 crc kubenswrapper[4612]: I0227 08:20:19.836141 4612 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/8439abb9-0c80-44f9-9ec4-ef0c99a50182-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 27 08:20:19 crc kubenswrapper[4612]: I0227 08:20:19.836162 4612 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8439abb9-0c80-44f9-9ec4-ef0c99a50182-inventory\") on node \"crc\" DevicePath \"\"" Feb 27 08:20:20 crc kubenswrapper[4612]: I0227 08:20:20.092033 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-t4m6d" event={"ID":"8439abb9-0c80-44f9-9ec4-ef0c99a50182","Type":"ContainerDied","Data":"500d89d9d87dba228578d380846411e6d04943d9dfa9cea3de810aad52c046ff"} Feb 27 08:20:20 crc kubenswrapper[4612]: I0227 08:20:20.092095 4612 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="500d89d9d87dba228578d380846411e6d04943d9dfa9cea3de810aad52c046ff" Feb 27 08:20:20 crc kubenswrapper[4612]: I0227 08:20:20.092194 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-t4m6d" Feb 27 08:20:20 crc kubenswrapper[4612]: I0227 08:20:20.232495 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-v4fds"] Feb 27 08:20:20 crc kubenswrapper[4612]: E0227 08:20:20.232862 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8439abb9-0c80-44f9-9ec4-ef0c99a50182" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Feb 27 08:20:20 crc kubenswrapper[4612]: I0227 08:20:20.232881 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="8439abb9-0c80-44f9-9ec4-ef0c99a50182" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Feb 27 08:20:20 crc kubenswrapper[4612]: E0227 08:20:20.232930 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ec08bd6b-d9ec-4b86-8397-9e1f709332ed" containerName="oc" Feb 27 08:20:20 crc kubenswrapper[4612]: I0227 08:20:20.232939 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="ec08bd6b-d9ec-4b86-8397-9e1f709332ed" containerName="oc" Feb 27 08:20:20 crc kubenswrapper[4612]: I0227 08:20:20.233125 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="8439abb9-0c80-44f9-9ec4-ef0c99a50182" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Feb 27 08:20:20 crc kubenswrapper[4612]: I0227 08:20:20.233164 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="ec08bd6b-d9ec-4b86-8397-9e1f709332ed" containerName="oc" Feb 27 08:20:20 crc kubenswrapper[4612]: I0227 08:20:20.234401 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-v4fds" Feb 27 08:20:20 crc kubenswrapper[4612]: I0227 08:20:20.238247 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 27 08:20:20 crc kubenswrapper[4612]: I0227 08:20:20.238986 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 27 08:20:20 crc kubenswrapper[4612]: I0227 08:20:20.239198 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 27 08:20:20 crc kubenswrapper[4612]: I0227 08:20:20.241704 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-2pwrb" Feb 27 08:20:20 crc kubenswrapper[4612]: I0227 08:20:20.270876 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-v4fds"] Feb 27 08:20:20 crc kubenswrapper[4612]: I0227 08:20:20.349227 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/c6f08c4d-80c1-49fb-b8ac-6d25b9d7f143-ssh-key-openstack-edpm-ipam\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-v4fds\" (UID: \"c6f08c4d-80c1-49fb-b8ac-6d25b9d7f143\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-v4fds" Feb 27 08:20:20 crc kubenswrapper[4612]: I0227 08:20:20.349278 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-snf4f\" (UniqueName: \"kubernetes.io/projected/c6f08c4d-80c1-49fb-b8ac-6d25b9d7f143-kube-api-access-snf4f\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-v4fds\" (UID: \"c6f08c4d-80c1-49fb-b8ac-6d25b9d7f143\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-v4fds" Feb 27 08:20:20 crc kubenswrapper[4612]: I0227 08:20:20.349382 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c6f08c4d-80c1-49fb-b8ac-6d25b9d7f143-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-v4fds\" (UID: \"c6f08c4d-80c1-49fb-b8ac-6d25b9d7f143\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-v4fds" Feb 27 08:20:20 crc kubenswrapper[4612]: I0227 08:20:20.450993 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/c6f08c4d-80c1-49fb-b8ac-6d25b9d7f143-ssh-key-openstack-edpm-ipam\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-v4fds\" (UID: \"c6f08c4d-80c1-49fb-b8ac-6d25b9d7f143\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-v4fds" Feb 27 08:20:20 crc kubenswrapper[4612]: I0227 08:20:20.451738 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-snf4f\" (UniqueName: \"kubernetes.io/projected/c6f08c4d-80c1-49fb-b8ac-6d25b9d7f143-kube-api-access-snf4f\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-v4fds\" (UID: \"c6f08c4d-80c1-49fb-b8ac-6d25b9d7f143\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-v4fds" Feb 27 08:20:20 crc kubenswrapper[4612]: I0227 08:20:20.451809 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c6f08c4d-80c1-49fb-b8ac-6d25b9d7f143-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-v4fds\" (UID: \"c6f08c4d-80c1-49fb-b8ac-6d25b9d7f143\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-v4fds" Feb 27 08:20:20 crc kubenswrapper[4612]: I0227 08:20:20.455142 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/c6f08c4d-80c1-49fb-b8ac-6d25b9d7f143-ssh-key-openstack-edpm-ipam\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-v4fds\" (UID: \"c6f08c4d-80c1-49fb-b8ac-6d25b9d7f143\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-v4fds" Feb 27 08:20:20 crc kubenswrapper[4612]: I0227 08:20:20.456125 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c6f08c4d-80c1-49fb-b8ac-6d25b9d7f143-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-v4fds\" (UID: \"c6f08c4d-80c1-49fb-b8ac-6d25b9d7f143\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-v4fds" Feb 27 08:20:20 crc kubenswrapper[4612]: I0227 08:20:20.483116 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-snf4f\" (UniqueName: \"kubernetes.io/projected/c6f08c4d-80c1-49fb-b8ac-6d25b9d7f143-kube-api-access-snf4f\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-v4fds\" (UID: \"c6f08c4d-80c1-49fb-b8ac-6d25b9d7f143\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-v4fds" Feb 27 08:20:20 crc kubenswrapper[4612]: I0227 08:20:20.555600 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-v4fds" Feb 27 08:20:21 crc kubenswrapper[4612]: I0227 08:20:21.098155 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-v4fds"] Feb 27 08:20:22 crc kubenswrapper[4612]: I0227 08:20:22.109603 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-v4fds" event={"ID":"c6f08c4d-80c1-49fb-b8ac-6d25b9d7f143","Type":"ContainerStarted","Data":"88b81298da723f61c42f4fe6809920cc76bafac665a223c2294acdc962741dd5"} Feb 27 08:20:22 crc kubenswrapper[4612]: I0227 08:20:22.110027 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-v4fds" event={"ID":"c6f08c4d-80c1-49fb-b8ac-6d25b9d7f143","Type":"ContainerStarted","Data":"39287e91abd292125f47eb1db245a17c2982b61d2e2648f25e895d46bcb22fbe"} Feb 27 08:20:22 crc kubenswrapper[4612]: I0227 08:20:22.133153 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-v4fds" podStartSLOduration=1.9543284459999999 podStartE2EDuration="2.133137558s" podCreationTimestamp="2026-02-27 08:20:20 +0000 UTC" firstStartedPulling="2026-02-27 08:20:21.116419909 +0000 UTC m=+1878.970349917" lastFinishedPulling="2026-02-27 08:20:21.295228991 +0000 UTC m=+1879.149159029" observedRunningTime="2026-02-27 08:20:22.129437262 +0000 UTC m=+1879.983367270" watchObservedRunningTime="2026-02-27 08:20:22.133137558 +0000 UTC m=+1879.987067546" Feb 27 08:20:26 crc kubenswrapper[4612]: I0227 08:20:26.854273 4612 scope.go:117] "RemoveContainer" containerID="7c62ebf970b4c78c5d31a0ce6596e9b398afbc8a6935511ce1b20bcd4f436b08" Feb 27 08:20:27 crc kubenswrapper[4612]: I0227 08:20:27.151779 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-924vb" event={"ID":"24599373-7adc-4a1b-8bb4-797bf726f43d","Type":"ContainerStarted","Data":"f80ae3b7ec2dce55e073a92197c42ea49b624e8fd3575dd785e408adfa1598e5"} Feb 27 08:20:56 crc kubenswrapper[4612]: I0227 08:20:56.041937 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-2ac7-account-create-update-hnvdn"] Feb 27 08:20:56 crc kubenswrapper[4612]: I0227 08:20:56.048050 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-36be-account-create-update-7grtc"] Feb 27 08:20:56 crc kubenswrapper[4612]: I0227 08:20:56.055715 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-tfl6p"] Feb 27 08:20:56 crc kubenswrapper[4612]: I0227 08:20:56.071040 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-36be-account-create-update-7grtc"] Feb 27 08:20:56 crc kubenswrapper[4612]: I0227 08:20:56.071114 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-2ac7-account-create-update-hnvdn"] Feb 27 08:20:56 crc kubenswrapper[4612]: I0227 08:20:56.076980 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-tfl6p"] Feb 27 08:20:56 crc kubenswrapper[4612]: I0227 08:20:56.867652 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a5162728-e496-4f30-9efb-a411e8e52ba4" path="/var/lib/kubelet/pods/a5162728-e496-4f30-9efb-a411e8e52ba4/volumes" Feb 27 08:20:56 crc kubenswrapper[4612]: I0227 08:20:56.871655 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d1b7e951-57b4-47f3-b337-b3468e7c6637" path="/var/lib/kubelet/pods/d1b7e951-57b4-47f3-b337-b3468e7c6637/volumes" Feb 27 08:20:56 crc kubenswrapper[4612]: I0227 08:20:56.872622 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e63e127a-2950-4a4c-a194-668b5142adb8" path="/var/lib/kubelet/pods/e63e127a-2950-4a4c-a194-668b5142adb8/volumes" Feb 27 08:20:57 crc kubenswrapper[4612]: I0227 08:20:57.031635 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-pqvhc"] Feb 27 08:20:57 crc kubenswrapper[4612]: I0227 08:20:57.043483 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-5f7d-account-create-update-fptbd"] Feb 27 08:20:57 crc kubenswrapper[4612]: I0227 08:20:57.058219 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-c7w6b"] Feb 27 08:20:57 crc kubenswrapper[4612]: I0227 08:20:57.067142 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-pqvhc"] Feb 27 08:20:57 crc kubenswrapper[4612]: I0227 08:20:57.106244 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-5f7d-account-create-update-fptbd"] Feb 27 08:20:57 crc kubenswrapper[4612]: I0227 08:20:57.115852 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-c7w6b"] Feb 27 08:20:58 crc kubenswrapper[4612]: I0227 08:20:58.863571 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="06a51ad9-c1fb-452b-9c2b-62e4d2c80266" path="/var/lib/kubelet/pods/06a51ad9-c1fb-452b-9c2b-62e4d2c80266/volumes" Feb 27 08:20:58 crc kubenswrapper[4612]: I0227 08:20:58.864778 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5c30c00d-ed94-4980-a218-a2e3b95bd8eb" path="/var/lib/kubelet/pods/5c30c00d-ed94-4980-a218-a2e3b95bd8eb/volumes" Feb 27 08:20:58 crc kubenswrapper[4612]: I0227 08:20:58.865635 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ed43d122-6dcc-4bce-9ed1-184dcb96647c" path="/var/lib/kubelet/pods/ed43d122-6dcc-4bce-9ed1-184dcb96647c/volumes" Feb 27 08:21:16 crc kubenswrapper[4612]: I0227 08:21:16.620516 4612 scope.go:117] "RemoveContainer" containerID="cdd9f292b431cd4aefc9a1fbba353ade1e867f18ac1d560e581e8d61e7bfc478" Feb 27 08:21:16 crc kubenswrapper[4612]: I0227 08:21:16.658756 4612 scope.go:117] "RemoveContainer" containerID="2cc2b1f42a72c486e6a061a906a7e9f77b61101b86e5eb8b8cd7dddefa341eb9" Feb 27 08:21:16 crc kubenswrapper[4612]: I0227 08:21:16.703321 4612 scope.go:117] "RemoveContainer" containerID="fa004161ac9acb310bf260abbf49e83dab05575fae08eb7b55a7941699bab387" Feb 27 08:21:16 crc kubenswrapper[4612]: I0227 08:21:16.743763 4612 scope.go:117] "RemoveContainer" containerID="76720259afaeb97d1994408be42c233f8079c49ec8b77bf9ebb83772ca20fb35" Feb 27 08:21:16 crc kubenswrapper[4612]: I0227 08:21:16.786978 4612 scope.go:117] "RemoveContainer" containerID="143969eaf93c6f453f01dc76dd51a7faea9680aaf6f7dde73247b621512f99e2" Feb 27 08:21:16 crc kubenswrapper[4612]: I0227 08:21:16.823930 4612 scope.go:117] "RemoveContainer" containerID="95d77cdca63cf202190670df3149188fa1b7295ea5a44c2f79b867a58f183e6e" Feb 27 08:21:34 crc kubenswrapper[4612]: I0227 08:21:34.061641 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-z67kr"] Feb 27 08:21:34 crc kubenswrapper[4612]: I0227 08:21:34.080906 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-z67kr"] Feb 27 08:21:34 crc kubenswrapper[4612]: I0227 08:21:34.769794 4612 generic.go:334] "Generic (PLEG): container finished" podID="c6f08c4d-80c1-49fb-b8ac-6d25b9d7f143" containerID="88b81298da723f61c42f4fe6809920cc76bafac665a223c2294acdc962741dd5" exitCode=0 Feb 27 08:21:34 crc kubenswrapper[4612]: I0227 08:21:34.769841 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-v4fds" event={"ID":"c6f08c4d-80c1-49fb-b8ac-6d25b9d7f143","Type":"ContainerDied","Data":"88b81298da723f61c42f4fe6809920cc76bafac665a223c2294acdc962741dd5"} Feb 27 08:21:34 crc kubenswrapper[4612]: I0227 08:21:34.863959 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3d52b072-aded-480f-ae46-d89c9be0c9f8" path="/var/lib/kubelet/pods/3d52b072-aded-480f-ae46-d89c9be0c9f8/volumes" Feb 27 08:21:36 crc kubenswrapper[4612]: I0227 08:21:36.167410 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-v4fds" Feb 27 08:21:36 crc kubenswrapper[4612]: I0227 08:21:36.265114 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c6f08c4d-80c1-49fb-b8ac-6d25b9d7f143-inventory\") pod \"c6f08c4d-80c1-49fb-b8ac-6d25b9d7f143\" (UID: \"c6f08c4d-80c1-49fb-b8ac-6d25b9d7f143\") " Feb 27 08:21:36 crc kubenswrapper[4612]: I0227 08:21:36.265534 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-snf4f\" (UniqueName: \"kubernetes.io/projected/c6f08c4d-80c1-49fb-b8ac-6d25b9d7f143-kube-api-access-snf4f\") pod \"c6f08c4d-80c1-49fb-b8ac-6d25b9d7f143\" (UID: \"c6f08c4d-80c1-49fb-b8ac-6d25b9d7f143\") " Feb 27 08:21:36 crc kubenswrapper[4612]: I0227 08:21:36.265729 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/c6f08c4d-80c1-49fb-b8ac-6d25b9d7f143-ssh-key-openstack-edpm-ipam\") pod \"c6f08c4d-80c1-49fb-b8ac-6d25b9d7f143\" (UID: \"c6f08c4d-80c1-49fb-b8ac-6d25b9d7f143\") " Feb 27 08:21:36 crc kubenswrapper[4612]: I0227 08:21:36.278933 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c6f08c4d-80c1-49fb-b8ac-6d25b9d7f143-kube-api-access-snf4f" (OuterVolumeSpecName: "kube-api-access-snf4f") pod "c6f08c4d-80c1-49fb-b8ac-6d25b9d7f143" (UID: "c6f08c4d-80c1-49fb-b8ac-6d25b9d7f143"). InnerVolumeSpecName "kube-api-access-snf4f". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 08:21:36 crc kubenswrapper[4612]: I0227 08:21:36.308373 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c6f08c4d-80c1-49fb-b8ac-6d25b9d7f143-inventory" (OuterVolumeSpecName: "inventory") pod "c6f08c4d-80c1-49fb-b8ac-6d25b9d7f143" (UID: "c6f08c4d-80c1-49fb-b8ac-6d25b9d7f143"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:21:36 crc kubenswrapper[4612]: I0227 08:21:36.309918 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c6f08c4d-80c1-49fb-b8ac-6d25b9d7f143-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "c6f08c4d-80c1-49fb-b8ac-6d25b9d7f143" (UID: "c6f08c4d-80c1-49fb-b8ac-6d25b9d7f143"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:21:36 crc kubenswrapper[4612]: I0227 08:21:36.367909 4612 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c6f08c4d-80c1-49fb-b8ac-6d25b9d7f143-inventory\") on node \"crc\" DevicePath \"\"" Feb 27 08:21:36 crc kubenswrapper[4612]: I0227 08:21:36.367943 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-snf4f\" (UniqueName: \"kubernetes.io/projected/c6f08c4d-80c1-49fb-b8ac-6d25b9d7f143-kube-api-access-snf4f\") on node \"crc\" DevicePath \"\"" Feb 27 08:21:36 crc kubenswrapper[4612]: I0227 08:21:36.367955 4612 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/c6f08c4d-80c1-49fb-b8ac-6d25b9d7f143-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 27 08:21:36 crc kubenswrapper[4612]: I0227 08:21:36.788763 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-v4fds" event={"ID":"c6f08c4d-80c1-49fb-b8ac-6d25b9d7f143","Type":"ContainerDied","Data":"39287e91abd292125f47eb1db245a17c2982b61d2e2648f25e895d46bcb22fbe"} Feb 27 08:21:36 crc kubenswrapper[4612]: I0227 08:21:36.788813 4612 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="39287e91abd292125f47eb1db245a17c2982b61d2e2648f25e895d46bcb22fbe" Feb 27 08:21:36 crc kubenswrapper[4612]: I0227 08:21:36.788885 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-v4fds" Feb 27 08:21:36 crc kubenswrapper[4612]: I0227 08:21:36.925506 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-vfpr6"] Feb 27 08:21:36 crc kubenswrapper[4612]: E0227 08:21:36.925915 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c6f08c4d-80c1-49fb-b8ac-6d25b9d7f143" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Feb 27 08:21:36 crc kubenswrapper[4612]: I0227 08:21:36.925936 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="c6f08c4d-80c1-49fb-b8ac-6d25b9d7f143" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Feb 27 08:21:36 crc kubenswrapper[4612]: I0227 08:21:36.926114 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="c6f08c4d-80c1-49fb-b8ac-6d25b9d7f143" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Feb 27 08:21:36 crc kubenswrapper[4612]: I0227 08:21:36.926737 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-vfpr6" Feb 27 08:21:36 crc kubenswrapper[4612]: I0227 08:21:36.929296 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 27 08:21:36 crc kubenswrapper[4612]: I0227 08:21:36.929497 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-2pwrb" Feb 27 08:21:36 crc kubenswrapper[4612]: I0227 08:21:36.929728 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 27 08:21:36 crc kubenswrapper[4612]: I0227 08:21:36.948653 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-vfpr6"] Feb 27 08:21:36 crc kubenswrapper[4612]: I0227 08:21:36.949976 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 27 08:21:37 crc kubenswrapper[4612]: I0227 08:21:37.079958 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7mbht\" (UniqueName: \"kubernetes.io/projected/ff510046-7399-4fdd-9880-67f6a51b61db-kube-api-access-7mbht\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-vfpr6\" (UID: \"ff510046-7399-4fdd-9880-67f6a51b61db\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-vfpr6" Feb 27 08:21:37 crc kubenswrapper[4612]: I0227 08:21:37.080226 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ff510046-7399-4fdd-9880-67f6a51b61db-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-vfpr6\" (UID: \"ff510046-7399-4fdd-9880-67f6a51b61db\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-vfpr6" Feb 27 08:21:37 crc kubenswrapper[4612]: I0227 08:21:37.080828 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/ff510046-7399-4fdd-9880-67f6a51b61db-ssh-key-openstack-edpm-ipam\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-vfpr6\" (UID: \"ff510046-7399-4fdd-9880-67f6a51b61db\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-vfpr6" Feb 27 08:21:37 crc kubenswrapper[4612]: I0227 08:21:37.182739 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7mbht\" (UniqueName: \"kubernetes.io/projected/ff510046-7399-4fdd-9880-67f6a51b61db-kube-api-access-7mbht\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-vfpr6\" (UID: \"ff510046-7399-4fdd-9880-67f6a51b61db\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-vfpr6" Feb 27 08:21:37 crc kubenswrapper[4612]: I0227 08:21:37.182846 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ff510046-7399-4fdd-9880-67f6a51b61db-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-vfpr6\" (UID: \"ff510046-7399-4fdd-9880-67f6a51b61db\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-vfpr6" Feb 27 08:21:37 crc kubenswrapper[4612]: I0227 08:21:37.182944 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/ff510046-7399-4fdd-9880-67f6a51b61db-ssh-key-openstack-edpm-ipam\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-vfpr6\" (UID: \"ff510046-7399-4fdd-9880-67f6a51b61db\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-vfpr6" Feb 27 08:21:37 crc kubenswrapper[4612]: I0227 08:21:37.193635 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ff510046-7399-4fdd-9880-67f6a51b61db-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-vfpr6\" (UID: \"ff510046-7399-4fdd-9880-67f6a51b61db\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-vfpr6" Feb 27 08:21:37 crc kubenswrapper[4612]: I0227 08:21:37.204373 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/ff510046-7399-4fdd-9880-67f6a51b61db-ssh-key-openstack-edpm-ipam\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-vfpr6\" (UID: \"ff510046-7399-4fdd-9880-67f6a51b61db\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-vfpr6" Feb 27 08:21:37 crc kubenswrapper[4612]: I0227 08:21:37.208974 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7mbht\" (UniqueName: \"kubernetes.io/projected/ff510046-7399-4fdd-9880-67f6a51b61db-kube-api-access-7mbht\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-vfpr6\" (UID: \"ff510046-7399-4fdd-9880-67f6a51b61db\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-vfpr6" Feb 27 08:21:37 crc kubenswrapper[4612]: I0227 08:21:37.240966 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-vfpr6" Feb 27 08:21:37 crc kubenswrapper[4612]: I0227 08:21:37.798534 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-vfpr6"] Feb 27 08:21:37 crc kubenswrapper[4612]: W0227 08:21:37.802264 4612 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podff510046_7399_4fdd_9880_67f6a51b61db.slice/crio-290706299ff5a727d416d9a09de34b602fc35f2891c6e15a5a6c4d492f002328 WatchSource:0}: Error finding container 290706299ff5a727d416d9a09de34b602fc35f2891c6e15a5a6c4d492f002328: Status 404 returned error can't find the container with id 290706299ff5a727d416d9a09de34b602fc35f2891c6e15a5a6c4d492f002328 Feb 27 08:21:38 crc kubenswrapper[4612]: I0227 08:21:38.808173 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-vfpr6" event={"ID":"ff510046-7399-4fdd-9880-67f6a51b61db","Type":"ContainerStarted","Data":"e1a9effa2786011d7c5f76f778c763560a140900321ed125b786e266f6592147"} Feb 27 08:21:38 crc kubenswrapper[4612]: I0227 08:21:38.808551 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-vfpr6" event={"ID":"ff510046-7399-4fdd-9880-67f6a51b61db","Type":"ContainerStarted","Data":"290706299ff5a727d416d9a09de34b602fc35f2891c6e15a5a6c4d492f002328"} Feb 27 08:21:38 crc kubenswrapper[4612]: I0227 08:21:38.838718 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-vfpr6" podStartSLOduration=2.638625998 podStartE2EDuration="2.838671599s" podCreationTimestamp="2026-02-27 08:21:36 +0000 UTC" firstStartedPulling="2026-02-27 08:21:37.805013815 +0000 UTC m=+1955.658943813" lastFinishedPulling="2026-02-27 08:21:38.005059416 +0000 UTC m=+1955.858989414" observedRunningTime="2026-02-27 08:21:38.830195326 +0000 UTC m=+1956.684125324" watchObservedRunningTime="2026-02-27 08:21:38.838671599 +0000 UTC m=+1956.692601617" Feb 27 08:21:43 crc kubenswrapper[4612]: I0227 08:21:43.864115 4612 generic.go:334] "Generic (PLEG): container finished" podID="ff510046-7399-4fdd-9880-67f6a51b61db" containerID="e1a9effa2786011d7c5f76f778c763560a140900321ed125b786e266f6592147" exitCode=0 Feb 27 08:21:43 crc kubenswrapper[4612]: I0227 08:21:43.865845 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-vfpr6" event={"ID":"ff510046-7399-4fdd-9880-67f6a51b61db","Type":"ContainerDied","Data":"e1a9effa2786011d7c5f76f778c763560a140900321ed125b786e266f6592147"} Feb 27 08:21:45 crc kubenswrapper[4612]: I0227 08:21:45.295667 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-vfpr6" Feb 27 08:21:45 crc kubenswrapper[4612]: I0227 08:21:45.470476 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ff510046-7399-4fdd-9880-67f6a51b61db-inventory\") pod \"ff510046-7399-4fdd-9880-67f6a51b61db\" (UID: \"ff510046-7399-4fdd-9880-67f6a51b61db\") " Feb 27 08:21:45 crc kubenswrapper[4612]: I0227 08:21:45.470569 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7mbht\" (UniqueName: \"kubernetes.io/projected/ff510046-7399-4fdd-9880-67f6a51b61db-kube-api-access-7mbht\") pod \"ff510046-7399-4fdd-9880-67f6a51b61db\" (UID: \"ff510046-7399-4fdd-9880-67f6a51b61db\") " Feb 27 08:21:45 crc kubenswrapper[4612]: I0227 08:21:45.470711 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/ff510046-7399-4fdd-9880-67f6a51b61db-ssh-key-openstack-edpm-ipam\") pod \"ff510046-7399-4fdd-9880-67f6a51b61db\" (UID: \"ff510046-7399-4fdd-9880-67f6a51b61db\") " Feb 27 08:21:45 crc kubenswrapper[4612]: I0227 08:21:45.475537 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ff510046-7399-4fdd-9880-67f6a51b61db-kube-api-access-7mbht" (OuterVolumeSpecName: "kube-api-access-7mbht") pod "ff510046-7399-4fdd-9880-67f6a51b61db" (UID: "ff510046-7399-4fdd-9880-67f6a51b61db"). InnerVolumeSpecName "kube-api-access-7mbht". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 08:21:45 crc kubenswrapper[4612]: I0227 08:21:45.495974 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ff510046-7399-4fdd-9880-67f6a51b61db-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "ff510046-7399-4fdd-9880-67f6a51b61db" (UID: "ff510046-7399-4fdd-9880-67f6a51b61db"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:21:45 crc kubenswrapper[4612]: I0227 08:21:45.518563 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ff510046-7399-4fdd-9880-67f6a51b61db-inventory" (OuterVolumeSpecName: "inventory") pod "ff510046-7399-4fdd-9880-67f6a51b61db" (UID: "ff510046-7399-4fdd-9880-67f6a51b61db"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:21:45 crc kubenswrapper[4612]: I0227 08:21:45.573341 4612 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ff510046-7399-4fdd-9880-67f6a51b61db-inventory\") on node \"crc\" DevicePath \"\"" Feb 27 08:21:45 crc kubenswrapper[4612]: I0227 08:21:45.573381 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7mbht\" (UniqueName: \"kubernetes.io/projected/ff510046-7399-4fdd-9880-67f6a51b61db-kube-api-access-7mbht\") on node \"crc\" DevicePath \"\"" Feb 27 08:21:45 crc kubenswrapper[4612]: I0227 08:21:45.573395 4612 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/ff510046-7399-4fdd-9880-67f6a51b61db-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 27 08:21:45 crc kubenswrapper[4612]: I0227 08:21:45.884891 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-vfpr6" event={"ID":"ff510046-7399-4fdd-9880-67f6a51b61db","Type":"ContainerDied","Data":"290706299ff5a727d416d9a09de34b602fc35f2891c6e15a5a6c4d492f002328"} Feb 27 08:21:45 crc kubenswrapper[4612]: I0227 08:21:45.884930 4612 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="290706299ff5a727d416d9a09de34b602fc35f2891c6e15a5a6c4d492f002328" Feb 27 08:21:45 crc kubenswrapper[4612]: I0227 08:21:45.884994 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-vfpr6" Feb 27 08:21:46 crc kubenswrapper[4612]: I0227 08:21:46.023905 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-7fwd7"] Feb 27 08:21:46 crc kubenswrapper[4612]: E0227 08:21:46.024377 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff510046-7399-4fdd-9880-67f6a51b61db" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Feb 27 08:21:46 crc kubenswrapper[4612]: I0227 08:21:46.024397 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff510046-7399-4fdd-9880-67f6a51b61db" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Feb 27 08:21:46 crc kubenswrapper[4612]: I0227 08:21:46.024647 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="ff510046-7399-4fdd-9880-67f6a51b61db" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Feb 27 08:21:46 crc kubenswrapper[4612]: I0227 08:21:46.025496 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-7fwd7" Feb 27 08:21:46 crc kubenswrapper[4612]: I0227 08:21:46.028223 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-2pwrb" Feb 27 08:21:46 crc kubenswrapper[4612]: I0227 08:21:46.028280 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 27 08:21:46 crc kubenswrapper[4612]: I0227 08:21:46.029576 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 27 08:21:46 crc kubenswrapper[4612]: I0227 08:21:46.032064 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 27 08:21:46 crc kubenswrapper[4612]: I0227 08:21:46.035283 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-7fwd7"] Feb 27 08:21:46 crc kubenswrapper[4612]: I0227 08:21:46.184032 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-96h6r\" (UniqueName: \"kubernetes.io/projected/301bc35d-5ee7-4a2e-af9b-47a57e6c5123-kube-api-access-96h6r\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-7fwd7\" (UID: \"301bc35d-5ee7-4a2e-af9b-47a57e6c5123\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-7fwd7" Feb 27 08:21:46 crc kubenswrapper[4612]: I0227 08:21:46.184346 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/301bc35d-5ee7-4a2e-af9b-47a57e6c5123-ssh-key-openstack-edpm-ipam\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-7fwd7\" (UID: \"301bc35d-5ee7-4a2e-af9b-47a57e6c5123\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-7fwd7" Feb 27 08:21:46 crc kubenswrapper[4612]: I0227 08:21:46.184480 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/301bc35d-5ee7-4a2e-af9b-47a57e6c5123-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-7fwd7\" (UID: \"301bc35d-5ee7-4a2e-af9b-47a57e6c5123\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-7fwd7" Feb 27 08:21:46 crc kubenswrapper[4612]: I0227 08:21:46.286611 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-96h6r\" (UniqueName: \"kubernetes.io/projected/301bc35d-5ee7-4a2e-af9b-47a57e6c5123-kube-api-access-96h6r\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-7fwd7\" (UID: \"301bc35d-5ee7-4a2e-af9b-47a57e6c5123\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-7fwd7" Feb 27 08:21:46 crc kubenswrapper[4612]: I0227 08:21:46.286681 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/301bc35d-5ee7-4a2e-af9b-47a57e6c5123-ssh-key-openstack-edpm-ipam\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-7fwd7\" (UID: \"301bc35d-5ee7-4a2e-af9b-47a57e6c5123\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-7fwd7" Feb 27 08:21:46 crc kubenswrapper[4612]: I0227 08:21:46.286753 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/301bc35d-5ee7-4a2e-af9b-47a57e6c5123-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-7fwd7\" (UID: \"301bc35d-5ee7-4a2e-af9b-47a57e6c5123\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-7fwd7" Feb 27 08:21:46 crc kubenswrapper[4612]: I0227 08:21:46.291121 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/301bc35d-5ee7-4a2e-af9b-47a57e6c5123-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-7fwd7\" (UID: \"301bc35d-5ee7-4a2e-af9b-47a57e6c5123\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-7fwd7" Feb 27 08:21:46 crc kubenswrapper[4612]: I0227 08:21:46.301140 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/301bc35d-5ee7-4a2e-af9b-47a57e6c5123-ssh-key-openstack-edpm-ipam\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-7fwd7\" (UID: \"301bc35d-5ee7-4a2e-af9b-47a57e6c5123\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-7fwd7" Feb 27 08:21:46 crc kubenswrapper[4612]: I0227 08:21:46.312998 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-96h6r\" (UniqueName: \"kubernetes.io/projected/301bc35d-5ee7-4a2e-af9b-47a57e6c5123-kube-api-access-96h6r\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-7fwd7\" (UID: \"301bc35d-5ee7-4a2e-af9b-47a57e6c5123\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-7fwd7" Feb 27 08:21:46 crc kubenswrapper[4612]: I0227 08:21:46.349784 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-7fwd7" Feb 27 08:21:46 crc kubenswrapper[4612]: I0227 08:21:46.939767 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-7fwd7"] Feb 27 08:21:46 crc kubenswrapper[4612]: W0227 08:21:46.946286 4612 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod301bc35d_5ee7_4a2e_af9b_47a57e6c5123.slice/crio-057446f60e92b096c31c6acd0ff8c2d9c15f75171207ebe29b2331352604ab3e WatchSource:0}: Error finding container 057446f60e92b096c31c6acd0ff8c2d9c15f75171207ebe29b2331352604ab3e: Status 404 returned error can't find the container with id 057446f60e92b096c31c6acd0ff8c2d9c15f75171207ebe29b2331352604ab3e Feb 27 08:21:47 crc kubenswrapper[4612]: I0227 08:21:47.909224 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-7fwd7" event={"ID":"301bc35d-5ee7-4a2e-af9b-47a57e6c5123","Type":"ContainerStarted","Data":"5d97fda3381a6af5498f72df28199076599f13a37f6f6c5688b3ef20aa840a04"} Feb 27 08:21:47 crc kubenswrapper[4612]: I0227 08:21:47.909781 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-7fwd7" event={"ID":"301bc35d-5ee7-4a2e-af9b-47a57e6c5123","Type":"ContainerStarted","Data":"057446f60e92b096c31c6acd0ff8c2d9c15f75171207ebe29b2331352604ab3e"} Feb 27 08:21:47 crc kubenswrapper[4612]: I0227 08:21:47.939636 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-7fwd7" podStartSLOduration=2.765138882 podStartE2EDuration="2.939612379s" podCreationTimestamp="2026-02-27 08:21:45 +0000 UTC" firstStartedPulling="2026-02-27 08:21:46.949941576 +0000 UTC m=+1964.803871574" lastFinishedPulling="2026-02-27 08:21:47.124415053 +0000 UTC m=+1964.978345071" observedRunningTime="2026-02-27 08:21:47.92951979 +0000 UTC m=+1965.783449808" watchObservedRunningTime="2026-02-27 08:21:47.939612379 +0000 UTC m=+1965.793542387" Feb 27 08:21:58 crc kubenswrapper[4612]: I0227 08:21:58.039636 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-x5mhp"] Feb 27 08:21:58 crc kubenswrapper[4612]: I0227 08:21:58.048021 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-x5mhp"] Feb 27 08:21:58 crc kubenswrapper[4612]: I0227 08:21:58.863891 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5d42ee9b-5073-4f40-8a3d-632914f14bbc" path="/var/lib/kubelet/pods/5d42ee9b-5073-4f40-8a3d-632914f14bbc/volumes" Feb 27 08:21:59 crc kubenswrapper[4612]: I0227 08:21:59.033512 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-qckvl"] Feb 27 08:21:59 crc kubenswrapper[4612]: I0227 08:21:59.044175 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-qckvl"] Feb 27 08:22:00 crc kubenswrapper[4612]: I0227 08:22:00.149014 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29536342-c9hbc"] Feb 27 08:22:00 crc kubenswrapper[4612]: I0227 08:22:00.153586 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536342-c9hbc" Feb 27 08:22:00 crc kubenswrapper[4612]: I0227 08:22:00.156859 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 27 08:22:00 crc kubenswrapper[4612]: I0227 08:22:00.157101 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 27 08:22:00 crc kubenswrapper[4612]: I0227 08:22:00.157238 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-5zwfx" Feb 27 08:22:00 crc kubenswrapper[4612]: I0227 08:22:00.157396 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536342-c9hbc"] Feb 27 08:22:00 crc kubenswrapper[4612]: I0227 08:22:00.331548 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sg5r9\" (UniqueName: \"kubernetes.io/projected/ab70fa78-06b2-4c10-81ea-18287a1d14b3-kube-api-access-sg5r9\") pod \"auto-csr-approver-29536342-c9hbc\" (UID: \"ab70fa78-06b2-4c10-81ea-18287a1d14b3\") " pod="openshift-infra/auto-csr-approver-29536342-c9hbc" Feb 27 08:22:00 crc kubenswrapper[4612]: I0227 08:22:00.433608 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sg5r9\" (UniqueName: \"kubernetes.io/projected/ab70fa78-06b2-4c10-81ea-18287a1d14b3-kube-api-access-sg5r9\") pod \"auto-csr-approver-29536342-c9hbc\" (UID: \"ab70fa78-06b2-4c10-81ea-18287a1d14b3\") " pod="openshift-infra/auto-csr-approver-29536342-c9hbc" Feb 27 08:22:00 crc kubenswrapper[4612]: I0227 08:22:00.465281 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sg5r9\" (UniqueName: \"kubernetes.io/projected/ab70fa78-06b2-4c10-81ea-18287a1d14b3-kube-api-access-sg5r9\") pod \"auto-csr-approver-29536342-c9hbc\" (UID: \"ab70fa78-06b2-4c10-81ea-18287a1d14b3\") " pod="openshift-infra/auto-csr-approver-29536342-c9hbc" Feb 27 08:22:00 crc kubenswrapper[4612]: I0227 08:22:00.474043 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536342-c9hbc" Feb 27 08:22:00 crc kubenswrapper[4612]: I0227 08:22:00.866045 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a7bca0ea-5f1b-4de7-83d1-54c76f32a813" path="/var/lib/kubelet/pods/a7bca0ea-5f1b-4de7-83d1-54c76f32a813/volumes" Feb 27 08:22:00 crc kubenswrapper[4612]: I0227 08:22:00.979868 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536342-c9hbc"] Feb 27 08:22:01 crc kubenswrapper[4612]: I0227 08:22:01.739807 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536342-c9hbc" event={"ID":"ab70fa78-06b2-4c10-81ea-18287a1d14b3","Type":"ContainerStarted","Data":"85a673f246a2601f5ee0ec52e810ea07b77175be17c53f313fe2f824ebd9753d"} Feb 27 08:22:02 crc kubenswrapper[4612]: I0227 08:22:02.751733 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536342-c9hbc" event={"ID":"ab70fa78-06b2-4c10-81ea-18287a1d14b3","Type":"ContainerStarted","Data":"9f368ab9892602f19068a88bb54bed7fbc6b535ce9a9f7df140682d7180e0498"} Feb 27 08:22:02 crc kubenswrapper[4612]: I0227 08:22:02.778275 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29536342-c9hbc" podStartSLOduration=1.537233549 podStartE2EDuration="2.778258086s" podCreationTimestamp="2026-02-27 08:22:00 +0000 UTC" firstStartedPulling="2026-02-27 08:22:00.984897197 +0000 UTC m=+1978.838827195" lastFinishedPulling="2026-02-27 08:22:02.225921724 +0000 UTC m=+1980.079851732" observedRunningTime="2026-02-27 08:22:02.776083213 +0000 UTC m=+1980.630013211" watchObservedRunningTime="2026-02-27 08:22:02.778258086 +0000 UTC m=+1980.632188084" Feb 27 08:22:03 crc kubenswrapper[4612]: I0227 08:22:03.763379 4612 generic.go:334] "Generic (PLEG): container finished" podID="ab70fa78-06b2-4c10-81ea-18287a1d14b3" containerID="9f368ab9892602f19068a88bb54bed7fbc6b535ce9a9f7df140682d7180e0498" exitCode=0 Feb 27 08:22:03 crc kubenswrapper[4612]: I0227 08:22:03.763417 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536342-c9hbc" event={"ID":"ab70fa78-06b2-4c10-81ea-18287a1d14b3","Type":"ContainerDied","Data":"9f368ab9892602f19068a88bb54bed7fbc6b535ce9a9f7df140682d7180e0498"} Feb 27 08:22:05 crc kubenswrapper[4612]: I0227 08:22:05.099250 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536342-c9hbc" Feb 27 08:22:05 crc kubenswrapper[4612]: I0227 08:22:05.240655 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sg5r9\" (UniqueName: \"kubernetes.io/projected/ab70fa78-06b2-4c10-81ea-18287a1d14b3-kube-api-access-sg5r9\") pod \"ab70fa78-06b2-4c10-81ea-18287a1d14b3\" (UID: \"ab70fa78-06b2-4c10-81ea-18287a1d14b3\") " Feb 27 08:22:05 crc kubenswrapper[4612]: I0227 08:22:05.251029 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ab70fa78-06b2-4c10-81ea-18287a1d14b3-kube-api-access-sg5r9" (OuterVolumeSpecName: "kube-api-access-sg5r9") pod "ab70fa78-06b2-4c10-81ea-18287a1d14b3" (UID: "ab70fa78-06b2-4c10-81ea-18287a1d14b3"). InnerVolumeSpecName "kube-api-access-sg5r9". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 08:22:05 crc kubenswrapper[4612]: I0227 08:22:05.347157 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sg5r9\" (UniqueName: \"kubernetes.io/projected/ab70fa78-06b2-4c10-81ea-18287a1d14b3-kube-api-access-sg5r9\") on node \"crc\" DevicePath \"\"" Feb 27 08:22:05 crc kubenswrapper[4612]: I0227 08:22:05.785133 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536342-c9hbc" event={"ID":"ab70fa78-06b2-4c10-81ea-18287a1d14b3","Type":"ContainerDied","Data":"85a673f246a2601f5ee0ec52e810ea07b77175be17c53f313fe2f824ebd9753d"} Feb 27 08:22:05 crc kubenswrapper[4612]: I0227 08:22:05.785181 4612 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="85a673f246a2601f5ee0ec52e810ea07b77175be17c53f313fe2f824ebd9753d" Feb 27 08:22:05 crc kubenswrapper[4612]: I0227 08:22:05.785639 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536342-c9hbc" Feb 27 08:22:05 crc kubenswrapper[4612]: I0227 08:22:05.840996 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29536336-49sn8"] Feb 27 08:22:05 crc kubenswrapper[4612]: I0227 08:22:05.852225 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29536336-49sn8"] Feb 27 08:22:06 crc kubenswrapper[4612]: I0227 08:22:06.865193 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="abd98e63-1fe3-4643-8df8-7512d3afb557" path="/var/lib/kubelet/pods/abd98e63-1fe3-4643-8df8-7512d3afb557/volumes" Feb 27 08:22:17 crc kubenswrapper[4612]: I0227 08:22:17.002831 4612 scope.go:117] "RemoveContainer" containerID="73e8d5c17c8a1763106d32f761ba1f2030f5bf13fe92371e47c559d774a2b505" Feb 27 08:22:17 crc kubenswrapper[4612]: I0227 08:22:17.044274 4612 scope.go:117] "RemoveContainer" containerID="55f1d5da6b2040184cc24875298b865e1ce09f44138529f1feec92112d3f4362" Feb 27 08:22:17 crc kubenswrapper[4612]: I0227 08:22:17.096837 4612 scope.go:117] "RemoveContainer" containerID="c58a73db1308d49e03d771c746d8a101c60cc387a0f267816a563635eb23794f" Feb 27 08:22:17 crc kubenswrapper[4612]: I0227 08:22:17.159958 4612 scope.go:117] "RemoveContainer" containerID="0b4e437518f07bffd24b4f73fca63432cbe5816775bc61ba305365ed6b924fe6" Feb 27 08:22:27 crc kubenswrapper[4612]: I0227 08:22:27.003031 4612 generic.go:334] "Generic (PLEG): container finished" podID="301bc35d-5ee7-4a2e-af9b-47a57e6c5123" containerID="5d97fda3381a6af5498f72df28199076599f13a37f6f6c5688b3ef20aa840a04" exitCode=0 Feb 27 08:22:27 crc kubenswrapper[4612]: I0227 08:22:27.003274 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-7fwd7" event={"ID":"301bc35d-5ee7-4a2e-af9b-47a57e6c5123","Type":"ContainerDied","Data":"5d97fda3381a6af5498f72df28199076599f13a37f6f6c5688b3ef20aa840a04"} Feb 27 08:22:28 crc kubenswrapper[4612]: I0227 08:22:28.547496 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-7fwd7" Feb 27 08:22:28 crc kubenswrapper[4612]: I0227 08:22:28.711471 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-96h6r\" (UniqueName: \"kubernetes.io/projected/301bc35d-5ee7-4a2e-af9b-47a57e6c5123-kube-api-access-96h6r\") pod \"301bc35d-5ee7-4a2e-af9b-47a57e6c5123\" (UID: \"301bc35d-5ee7-4a2e-af9b-47a57e6c5123\") " Feb 27 08:22:28 crc kubenswrapper[4612]: I0227 08:22:28.711978 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/301bc35d-5ee7-4a2e-af9b-47a57e6c5123-ssh-key-openstack-edpm-ipam\") pod \"301bc35d-5ee7-4a2e-af9b-47a57e6c5123\" (UID: \"301bc35d-5ee7-4a2e-af9b-47a57e6c5123\") " Feb 27 08:22:28 crc kubenswrapper[4612]: I0227 08:22:28.712014 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/301bc35d-5ee7-4a2e-af9b-47a57e6c5123-inventory\") pod \"301bc35d-5ee7-4a2e-af9b-47a57e6c5123\" (UID: \"301bc35d-5ee7-4a2e-af9b-47a57e6c5123\") " Feb 27 08:22:28 crc kubenswrapper[4612]: I0227 08:22:28.717734 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/301bc35d-5ee7-4a2e-af9b-47a57e6c5123-kube-api-access-96h6r" (OuterVolumeSpecName: "kube-api-access-96h6r") pod "301bc35d-5ee7-4a2e-af9b-47a57e6c5123" (UID: "301bc35d-5ee7-4a2e-af9b-47a57e6c5123"). InnerVolumeSpecName "kube-api-access-96h6r". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 08:22:28 crc kubenswrapper[4612]: I0227 08:22:28.758006 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/301bc35d-5ee7-4a2e-af9b-47a57e6c5123-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "301bc35d-5ee7-4a2e-af9b-47a57e6c5123" (UID: "301bc35d-5ee7-4a2e-af9b-47a57e6c5123"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:22:28 crc kubenswrapper[4612]: I0227 08:22:28.760630 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/301bc35d-5ee7-4a2e-af9b-47a57e6c5123-inventory" (OuterVolumeSpecName: "inventory") pod "301bc35d-5ee7-4a2e-af9b-47a57e6c5123" (UID: "301bc35d-5ee7-4a2e-af9b-47a57e6c5123"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:22:28 crc kubenswrapper[4612]: I0227 08:22:28.814919 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-96h6r\" (UniqueName: \"kubernetes.io/projected/301bc35d-5ee7-4a2e-af9b-47a57e6c5123-kube-api-access-96h6r\") on node \"crc\" DevicePath \"\"" Feb 27 08:22:28 crc kubenswrapper[4612]: I0227 08:22:28.814951 4612 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/301bc35d-5ee7-4a2e-af9b-47a57e6c5123-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 27 08:22:28 crc kubenswrapper[4612]: I0227 08:22:28.814963 4612 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/301bc35d-5ee7-4a2e-af9b-47a57e6c5123-inventory\") on node \"crc\" DevicePath \"\"" Feb 27 08:22:29 crc kubenswrapper[4612]: I0227 08:22:29.022930 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-7fwd7" event={"ID":"301bc35d-5ee7-4a2e-af9b-47a57e6c5123","Type":"ContainerDied","Data":"057446f60e92b096c31c6acd0ff8c2d9c15f75171207ebe29b2331352604ab3e"} Feb 27 08:22:29 crc kubenswrapper[4612]: I0227 08:22:29.022974 4612 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="057446f60e92b096c31c6acd0ff8c2d9c15f75171207ebe29b2331352604ab3e" Feb 27 08:22:29 crc kubenswrapper[4612]: I0227 08:22:29.023037 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-7fwd7" Feb 27 08:22:29 crc kubenswrapper[4612]: I0227 08:22:29.135769 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-w9sl2"] Feb 27 08:22:29 crc kubenswrapper[4612]: E0227 08:22:29.136298 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ab70fa78-06b2-4c10-81ea-18287a1d14b3" containerName="oc" Feb 27 08:22:29 crc kubenswrapper[4612]: I0227 08:22:29.136329 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="ab70fa78-06b2-4c10-81ea-18287a1d14b3" containerName="oc" Feb 27 08:22:29 crc kubenswrapper[4612]: E0227 08:22:29.136362 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="301bc35d-5ee7-4a2e-af9b-47a57e6c5123" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Feb 27 08:22:29 crc kubenswrapper[4612]: I0227 08:22:29.136375 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="301bc35d-5ee7-4a2e-af9b-47a57e6c5123" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Feb 27 08:22:29 crc kubenswrapper[4612]: I0227 08:22:29.136656 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="301bc35d-5ee7-4a2e-af9b-47a57e6c5123" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Feb 27 08:22:29 crc kubenswrapper[4612]: I0227 08:22:29.136680 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="ab70fa78-06b2-4c10-81ea-18287a1d14b3" containerName="oc" Feb 27 08:22:29 crc kubenswrapper[4612]: I0227 08:22:29.137400 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-w9sl2" Feb 27 08:22:29 crc kubenswrapper[4612]: I0227 08:22:29.139268 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 27 08:22:29 crc kubenswrapper[4612]: I0227 08:22:29.144574 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 27 08:22:29 crc kubenswrapper[4612]: I0227 08:22:29.144868 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 27 08:22:29 crc kubenswrapper[4612]: I0227 08:22:29.147999 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-2pwrb" Feb 27 08:22:29 crc kubenswrapper[4612]: I0227 08:22:29.156540 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-w9sl2"] Feb 27 08:22:29 crc kubenswrapper[4612]: I0227 08:22:29.325368 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lgnrq\" (UniqueName: \"kubernetes.io/projected/68ded60c-e2ee-48fb-986b-cebf9e447b82-kube-api-access-lgnrq\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-w9sl2\" (UID: \"68ded60c-e2ee-48fb-986b-cebf9e447b82\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-w9sl2" Feb 27 08:22:29 crc kubenswrapper[4612]: I0227 08:22:29.325408 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/68ded60c-e2ee-48fb-986b-cebf9e447b82-ssh-key-openstack-edpm-ipam\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-w9sl2\" (UID: \"68ded60c-e2ee-48fb-986b-cebf9e447b82\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-w9sl2" Feb 27 08:22:29 crc kubenswrapper[4612]: I0227 08:22:29.325492 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/68ded60c-e2ee-48fb-986b-cebf9e447b82-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-w9sl2\" (UID: \"68ded60c-e2ee-48fb-986b-cebf9e447b82\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-w9sl2" Feb 27 08:22:29 crc kubenswrapper[4612]: I0227 08:22:29.426950 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lgnrq\" (UniqueName: \"kubernetes.io/projected/68ded60c-e2ee-48fb-986b-cebf9e447b82-kube-api-access-lgnrq\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-w9sl2\" (UID: \"68ded60c-e2ee-48fb-986b-cebf9e447b82\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-w9sl2" Feb 27 08:22:29 crc kubenswrapper[4612]: I0227 08:22:29.426999 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/68ded60c-e2ee-48fb-986b-cebf9e447b82-ssh-key-openstack-edpm-ipam\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-w9sl2\" (UID: \"68ded60c-e2ee-48fb-986b-cebf9e447b82\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-w9sl2" Feb 27 08:22:29 crc kubenswrapper[4612]: I0227 08:22:29.427058 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/68ded60c-e2ee-48fb-986b-cebf9e447b82-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-w9sl2\" (UID: \"68ded60c-e2ee-48fb-986b-cebf9e447b82\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-w9sl2" Feb 27 08:22:29 crc kubenswrapper[4612]: I0227 08:22:29.431121 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/68ded60c-e2ee-48fb-986b-cebf9e447b82-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-w9sl2\" (UID: \"68ded60c-e2ee-48fb-986b-cebf9e447b82\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-w9sl2" Feb 27 08:22:29 crc kubenswrapper[4612]: I0227 08:22:29.431610 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/68ded60c-e2ee-48fb-986b-cebf9e447b82-ssh-key-openstack-edpm-ipam\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-w9sl2\" (UID: \"68ded60c-e2ee-48fb-986b-cebf9e447b82\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-w9sl2" Feb 27 08:22:29 crc kubenswrapper[4612]: I0227 08:22:29.444908 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lgnrq\" (UniqueName: \"kubernetes.io/projected/68ded60c-e2ee-48fb-986b-cebf9e447b82-kube-api-access-lgnrq\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-w9sl2\" (UID: \"68ded60c-e2ee-48fb-986b-cebf9e447b82\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-w9sl2" Feb 27 08:22:29 crc kubenswrapper[4612]: I0227 08:22:29.475822 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-w9sl2" Feb 27 08:22:30 crc kubenswrapper[4612]: I0227 08:22:30.084997 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-w9sl2"] Feb 27 08:22:31 crc kubenswrapper[4612]: I0227 08:22:31.044479 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-w9sl2" event={"ID":"68ded60c-e2ee-48fb-986b-cebf9e447b82","Type":"ContainerStarted","Data":"c7b6325e8a8b1794b91052eebf4e1273dfa21930b36d5fae37495cce92d7d929"} Feb 27 08:22:31 crc kubenswrapper[4612]: I0227 08:22:31.045122 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-w9sl2" event={"ID":"68ded60c-e2ee-48fb-986b-cebf9e447b82","Type":"ContainerStarted","Data":"0577bd7f64b2b8eafd769ebaab39f589f70cbeafe341c030b8c642989811748c"} Feb 27 08:22:31 crc kubenswrapper[4612]: I0227 08:22:31.078778 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-w9sl2" podStartSLOduration=1.8218018919999999 podStartE2EDuration="2.078748746s" podCreationTimestamp="2026-02-27 08:22:29 +0000 UTC" firstStartedPulling="2026-02-27 08:22:30.076941504 +0000 UTC m=+2007.930871512" lastFinishedPulling="2026-02-27 08:22:30.333888368 +0000 UTC m=+2008.187818366" observedRunningTime="2026-02-27 08:22:31.065973019 +0000 UTC m=+2008.919903017" watchObservedRunningTime="2026-02-27 08:22:31.078748746 +0000 UTC m=+2008.932678774" Feb 27 08:22:43 crc kubenswrapper[4612]: I0227 08:22:43.054856 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-7jx4l"] Feb 27 08:22:43 crc kubenswrapper[4612]: I0227 08:22:43.063707 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-7jx4l"] Feb 27 08:22:44 crc kubenswrapper[4612]: I0227 08:22:44.865984 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="16afe9fc-ff42-487d-9c1a-57f93c0a242c" path="/var/lib/kubelet/pods/16afe9fc-ff42-487d-9c1a-57f93c0a242c/volumes" Feb 27 08:22:46 crc kubenswrapper[4612]: I0227 08:22:46.028193 4612 patch_prober.go:28] interesting pod/machine-config-daemon-924vb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 27 08:22:46 crc kubenswrapper[4612]: I0227 08:22:46.028261 4612 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 27 08:23:16 crc kubenswrapper[4612]: I0227 08:23:16.027564 4612 patch_prober.go:28] interesting pod/machine-config-daemon-924vb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 27 08:23:16 crc kubenswrapper[4612]: I0227 08:23:16.030440 4612 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 27 08:23:17 crc kubenswrapper[4612]: I0227 08:23:17.263141 4612 scope.go:117] "RemoveContainer" containerID="32e74328c187d5506986ecb007765a5d4157e22064c750d89488ce3ce250584c" Feb 27 08:23:20 crc kubenswrapper[4612]: I0227 08:23:20.576468 4612 generic.go:334] "Generic (PLEG): container finished" podID="68ded60c-e2ee-48fb-986b-cebf9e447b82" containerID="c7b6325e8a8b1794b91052eebf4e1273dfa21930b36d5fae37495cce92d7d929" exitCode=0 Feb 27 08:23:20 crc kubenswrapper[4612]: I0227 08:23:20.576654 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-w9sl2" event={"ID":"68ded60c-e2ee-48fb-986b-cebf9e447b82","Type":"ContainerDied","Data":"c7b6325e8a8b1794b91052eebf4e1273dfa21930b36d5fae37495cce92d7d929"} Feb 27 08:23:22 crc kubenswrapper[4612]: I0227 08:23:22.016068 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-w9sl2" Feb 27 08:23:22 crc kubenswrapper[4612]: I0227 08:23:22.168916 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lgnrq\" (UniqueName: \"kubernetes.io/projected/68ded60c-e2ee-48fb-986b-cebf9e447b82-kube-api-access-lgnrq\") pod \"68ded60c-e2ee-48fb-986b-cebf9e447b82\" (UID: \"68ded60c-e2ee-48fb-986b-cebf9e447b82\") " Feb 27 08:23:22 crc kubenswrapper[4612]: I0227 08:23:22.169382 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/68ded60c-e2ee-48fb-986b-cebf9e447b82-inventory\") pod \"68ded60c-e2ee-48fb-986b-cebf9e447b82\" (UID: \"68ded60c-e2ee-48fb-986b-cebf9e447b82\") " Feb 27 08:23:22 crc kubenswrapper[4612]: I0227 08:23:22.169441 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/68ded60c-e2ee-48fb-986b-cebf9e447b82-ssh-key-openstack-edpm-ipam\") pod \"68ded60c-e2ee-48fb-986b-cebf9e447b82\" (UID: \"68ded60c-e2ee-48fb-986b-cebf9e447b82\") " Feb 27 08:23:22 crc kubenswrapper[4612]: I0227 08:23:22.174745 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/68ded60c-e2ee-48fb-986b-cebf9e447b82-kube-api-access-lgnrq" (OuterVolumeSpecName: "kube-api-access-lgnrq") pod "68ded60c-e2ee-48fb-986b-cebf9e447b82" (UID: "68ded60c-e2ee-48fb-986b-cebf9e447b82"). InnerVolumeSpecName "kube-api-access-lgnrq". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 08:23:22 crc kubenswrapper[4612]: I0227 08:23:22.195750 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/68ded60c-e2ee-48fb-986b-cebf9e447b82-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "68ded60c-e2ee-48fb-986b-cebf9e447b82" (UID: "68ded60c-e2ee-48fb-986b-cebf9e447b82"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:23:22 crc kubenswrapper[4612]: I0227 08:23:22.200831 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/68ded60c-e2ee-48fb-986b-cebf9e447b82-inventory" (OuterVolumeSpecName: "inventory") pod "68ded60c-e2ee-48fb-986b-cebf9e447b82" (UID: "68ded60c-e2ee-48fb-986b-cebf9e447b82"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:23:22 crc kubenswrapper[4612]: I0227 08:23:22.274398 4612 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/68ded60c-e2ee-48fb-986b-cebf9e447b82-inventory\") on node \"crc\" DevicePath \"\"" Feb 27 08:23:22 crc kubenswrapper[4612]: I0227 08:23:22.274428 4612 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/68ded60c-e2ee-48fb-986b-cebf9e447b82-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 27 08:23:22 crc kubenswrapper[4612]: I0227 08:23:22.274440 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lgnrq\" (UniqueName: \"kubernetes.io/projected/68ded60c-e2ee-48fb-986b-cebf9e447b82-kube-api-access-lgnrq\") on node \"crc\" DevicePath \"\"" Feb 27 08:23:22 crc kubenswrapper[4612]: I0227 08:23:22.598883 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-w9sl2" event={"ID":"68ded60c-e2ee-48fb-986b-cebf9e447b82","Type":"ContainerDied","Data":"0577bd7f64b2b8eafd769ebaab39f589f70cbeafe341c030b8c642989811748c"} Feb 27 08:23:22 crc kubenswrapper[4612]: I0227 08:23:22.598943 4612 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0577bd7f64b2b8eafd769ebaab39f589f70cbeafe341c030b8c642989811748c" Feb 27 08:23:22 crc kubenswrapper[4612]: I0227 08:23:22.598952 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-w9sl2" Feb 27 08:23:22 crc kubenswrapper[4612]: I0227 08:23:22.711331 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-2g86c"] Feb 27 08:23:22 crc kubenswrapper[4612]: E0227 08:23:22.711773 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="68ded60c-e2ee-48fb-986b-cebf9e447b82" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Feb 27 08:23:22 crc kubenswrapper[4612]: I0227 08:23:22.711797 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="68ded60c-e2ee-48fb-986b-cebf9e447b82" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Feb 27 08:23:22 crc kubenswrapper[4612]: I0227 08:23:22.712076 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="68ded60c-e2ee-48fb-986b-cebf9e447b82" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Feb 27 08:23:22 crc kubenswrapper[4612]: I0227 08:23:22.712881 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-2g86c" Feb 27 08:23:22 crc kubenswrapper[4612]: I0227 08:23:22.715222 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 27 08:23:22 crc kubenswrapper[4612]: I0227 08:23:22.715586 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-2pwrb" Feb 27 08:23:22 crc kubenswrapper[4612]: I0227 08:23:22.715731 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 27 08:23:22 crc kubenswrapper[4612]: I0227 08:23:22.715999 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 27 08:23:22 crc kubenswrapper[4612]: I0227 08:23:22.730045 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-2g86c"] Feb 27 08:23:22 crc kubenswrapper[4612]: I0227 08:23:22.784921 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/d99b6ff5-a143-40ce-bee0-d4300122aff2-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-2g86c\" (UID: \"d99b6ff5-a143-40ce-bee0-d4300122aff2\") " pod="openstack/ssh-known-hosts-edpm-deployment-2g86c" Feb 27 08:23:22 crc kubenswrapper[4612]: I0227 08:23:22.785052 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/d99b6ff5-a143-40ce-bee0-d4300122aff2-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-2g86c\" (UID: \"d99b6ff5-a143-40ce-bee0-d4300122aff2\") " pod="openstack/ssh-known-hosts-edpm-deployment-2g86c" Feb 27 08:23:22 crc kubenswrapper[4612]: I0227 08:23:22.785122 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-97drj\" (UniqueName: \"kubernetes.io/projected/d99b6ff5-a143-40ce-bee0-d4300122aff2-kube-api-access-97drj\") pod \"ssh-known-hosts-edpm-deployment-2g86c\" (UID: \"d99b6ff5-a143-40ce-bee0-d4300122aff2\") " pod="openstack/ssh-known-hosts-edpm-deployment-2g86c" Feb 27 08:23:22 crc kubenswrapper[4612]: I0227 08:23:22.892506 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/d99b6ff5-a143-40ce-bee0-d4300122aff2-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-2g86c\" (UID: \"d99b6ff5-a143-40ce-bee0-d4300122aff2\") " pod="openstack/ssh-known-hosts-edpm-deployment-2g86c" Feb 27 08:23:22 crc kubenswrapper[4612]: I0227 08:23:22.892704 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/d99b6ff5-a143-40ce-bee0-d4300122aff2-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-2g86c\" (UID: \"d99b6ff5-a143-40ce-bee0-d4300122aff2\") " pod="openstack/ssh-known-hosts-edpm-deployment-2g86c" Feb 27 08:23:22 crc kubenswrapper[4612]: I0227 08:23:22.892825 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-97drj\" (UniqueName: \"kubernetes.io/projected/d99b6ff5-a143-40ce-bee0-d4300122aff2-kube-api-access-97drj\") pod \"ssh-known-hosts-edpm-deployment-2g86c\" (UID: \"d99b6ff5-a143-40ce-bee0-d4300122aff2\") " pod="openstack/ssh-known-hosts-edpm-deployment-2g86c" Feb 27 08:23:22 crc kubenswrapper[4612]: I0227 08:23:22.897330 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/d99b6ff5-a143-40ce-bee0-d4300122aff2-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-2g86c\" (UID: \"d99b6ff5-a143-40ce-bee0-d4300122aff2\") " pod="openstack/ssh-known-hosts-edpm-deployment-2g86c" Feb 27 08:23:22 crc kubenswrapper[4612]: I0227 08:23:22.901306 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/d99b6ff5-a143-40ce-bee0-d4300122aff2-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-2g86c\" (UID: \"d99b6ff5-a143-40ce-bee0-d4300122aff2\") " pod="openstack/ssh-known-hosts-edpm-deployment-2g86c" Feb 27 08:23:22 crc kubenswrapper[4612]: I0227 08:23:22.916654 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-97drj\" (UniqueName: \"kubernetes.io/projected/d99b6ff5-a143-40ce-bee0-d4300122aff2-kube-api-access-97drj\") pod \"ssh-known-hosts-edpm-deployment-2g86c\" (UID: \"d99b6ff5-a143-40ce-bee0-d4300122aff2\") " pod="openstack/ssh-known-hosts-edpm-deployment-2g86c" Feb 27 08:23:23 crc kubenswrapper[4612]: I0227 08:23:23.031066 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-2g86c" Feb 27 08:23:23 crc kubenswrapper[4612]: I0227 08:23:23.602068 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-2g86c"] Feb 27 08:23:23 crc kubenswrapper[4612]: I0227 08:23:23.969814 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-2lp2b"] Feb 27 08:23:23 crc kubenswrapper[4612]: I0227 08:23:23.977211 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2lp2b" Feb 27 08:23:23 crc kubenswrapper[4612]: I0227 08:23:23.985226 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-2lp2b"] Feb 27 08:23:24 crc kubenswrapper[4612]: I0227 08:23:24.122285 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/415215a1-2901-462b-aeff-10f96bfa8344-utilities\") pod \"community-operators-2lp2b\" (UID: \"415215a1-2901-462b-aeff-10f96bfa8344\") " pod="openshift-marketplace/community-operators-2lp2b" Feb 27 08:23:24 crc kubenswrapper[4612]: I0227 08:23:24.122618 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8r8f8\" (UniqueName: \"kubernetes.io/projected/415215a1-2901-462b-aeff-10f96bfa8344-kube-api-access-8r8f8\") pod \"community-operators-2lp2b\" (UID: \"415215a1-2901-462b-aeff-10f96bfa8344\") " pod="openshift-marketplace/community-operators-2lp2b" Feb 27 08:23:24 crc kubenswrapper[4612]: I0227 08:23:24.122664 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/415215a1-2901-462b-aeff-10f96bfa8344-catalog-content\") pod \"community-operators-2lp2b\" (UID: \"415215a1-2901-462b-aeff-10f96bfa8344\") " pod="openshift-marketplace/community-operators-2lp2b" Feb 27 08:23:24 crc kubenswrapper[4612]: I0227 08:23:24.224518 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8r8f8\" (UniqueName: \"kubernetes.io/projected/415215a1-2901-462b-aeff-10f96bfa8344-kube-api-access-8r8f8\") pod \"community-operators-2lp2b\" (UID: \"415215a1-2901-462b-aeff-10f96bfa8344\") " pod="openshift-marketplace/community-operators-2lp2b" Feb 27 08:23:24 crc kubenswrapper[4612]: I0227 08:23:24.224625 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/415215a1-2901-462b-aeff-10f96bfa8344-catalog-content\") pod \"community-operators-2lp2b\" (UID: \"415215a1-2901-462b-aeff-10f96bfa8344\") " pod="openshift-marketplace/community-operators-2lp2b" Feb 27 08:23:24 crc kubenswrapper[4612]: I0227 08:23:24.224764 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/415215a1-2901-462b-aeff-10f96bfa8344-utilities\") pod \"community-operators-2lp2b\" (UID: \"415215a1-2901-462b-aeff-10f96bfa8344\") " pod="openshift-marketplace/community-operators-2lp2b" Feb 27 08:23:24 crc kubenswrapper[4612]: I0227 08:23:24.225399 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/415215a1-2901-462b-aeff-10f96bfa8344-utilities\") pod \"community-operators-2lp2b\" (UID: \"415215a1-2901-462b-aeff-10f96bfa8344\") " pod="openshift-marketplace/community-operators-2lp2b" Feb 27 08:23:24 crc kubenswrapper[4612]: I0227 08:23:24.226176 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/415215a1-2901-462b-aeff-10f96bfa8344-catalog-content\") pod \"community-operators-2lp2b\" (UID: \"415215a1-2901-462b-aeff-10f96bfa8344\") " pod="openshift-marketplace/community-operators-2lp2b" Feb 27 08:23:24 crc kubenswrapper[4612]: I0227 08:23:24.264404 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8r8f8\" (UniqueName: \"kubernetes.io/projected/415215a1-2901-462b-aeff-10f96bfa8344-kube-api-access-8r8f8\") pod \"community-operators-2lp2b\" (UID: \"415215a1-2901-462b-aeff-10f96bfa8344\") " pod="openshift-marketplace/community-operators-2lp2b" Feb 27 08:23:24 crc kubenswrapper[4612]: I0227 08:23:24.335095 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2lp2b" Feb 27 08:23:24 crc kubenswrapper[4612]: I0227 08:23:24.627341 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-2g86c" event={"ID":"d99b6ff5-a143-40ce-bee0-d4300122aff2","Type":"ContainerStarted","Data":"0e66dd4c62bb4495706bb07100657019047e3dc7c5edf4f7a6cc9950150660c4"} Feb 27 08:23:24 crc kubenswrapper[4612]: I0227 08:23:24.627386 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-2g86c" event={"ID":"d99b6ff5-a143-40ce-bee0-d4300122aff2","Type":"ContainerStarted","Data":"625f6903e977284d8e88f8ab47b2ed5b9460ce4935126675804fb68069ef3b52"} Feb 27 08:23:24 crc kubenswrapper[4612]: I0227 08:23:24.673198 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ssh-known-hosts-edpm-deployment-2g86c" podStartSLOduration=2.477846652 podStartE2EDuration="2.673176087s" podCreationTimestamp="2026-02-27 08:23:22 +0000 UTC" firstStartedPulling="2026-02-27 08:23:23.617033987 +0000 UTC m=+2061.470963995" lastFinishedPulling="2026-02-27 08:23:23.812363442 +0000 UTC m=+2061.666293430" observedRunningTime="2026-02-27 08:23:24.663824269 +0000 UTC m=+2062.517754357" watchObservedRunningTime="2026-02-27 08:23:24.673176087 +0000 UTC m=+2062.527106085" Feb 27 08:23:24 crc kubenswrapper[4612]: I0227 08:23:24.943555 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-2lp2b"] Feb 27 08:23:25 crc kubenswrapper[4612]: I0227 08:23:25.636130 4612 generic.go:334] "Generic (PLEG): container finished" podID="415215a1-2901-462b-aeff-10f96bfa8344" containerID="4588e54cc61d60c9ce5cb1e03eb9e4df62f162d261f0e0ba89dfaa765ec91fd3" exitCode=0 Feb 27 08:23:25 crc kubenswrapper[4612]: I0227 08:23:25.636221 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2lp2b" event={"ID":"415215a1-2901-462b-aeff-10f96bfa8344","Type":"ContainerDied","Data":"4588e54cc61d60c9ce5cb1e03eb9e4df62f162d261f0e0ba89dfaa765ec91fd3"} Feb 27 08:23:25 crc kubenswrapper[4612]: I0227 08:23:25.637342 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2lp2b" event={"ID":"415215a1-2901-462b-aeff-10f96bfa8344","Type":"ContainerStarted","Data":"f7d0488fe4d7c232c2197a1bed4b54fcd1cdf791a2d6a745af35c3db867a4e57"} Feb 27 08:23:26 crc kubenswrapper[4612]: I0227 08:23:26.652762 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2lp2b" event={"ID":"415215a1-2901-462b-aeff-10f96bfa8344","Type":"ContainerStarted","Data":"ea7c0654c0cc16bdd192573ab2e11f43fbd6e7bcd7d5d55720e68e9190790653"} Feb 27 08:23:29 crc kubenswrapper[4612]: I0227 08:23:29.685035 4612 generic.go:334] "Generic (PLEG): container finished" podID="415215a1-2901-462b-aeff-10f96bfa8344" containerID="ea7c0654c0cc16bdd192573ab2e11f43fbd6e7bcd7d5d55720e68e9190790653" exitCode=0 Feb 27 08:23:29 crc kubenswrapper[4612]: I0227 08:23:29.685151 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2lp2b" event={"ID":"415215a1-2901-462b-aeff-10f96bfa8344","Type":"ContainerDied","Data":"ea7c0654c0cc16bdd192573ab2e11f43fbd6e7bcd7d5d55720e68e9190790653"} Feb 27 08:23:29 crc kubenswrapper[4612]: I0227 08:23:29.688741 4612 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 27 08:23:30 crc kubenswrapper[4612]: I0227 08:23:30.697938 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2lp2b" event={"ID":"415215a1-2901-462b-aeff-10f96bfa8344","Type":"ContainerStarted","Data":"5ed88a5b3997032d25d78a971b78bfb34a4fd50c61537093407649b14d617a73"} Feb 27 08:23:30 crc kubenswrapper[4612]: I0227 08:23:30.738996 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-2lp2b" podStartSLOduration=3.263170959 podStartE2EDuration="7.73895134s" podCreationTimestamp="2026-02-27 08:23:23 +0000 UTC" firstStartedPulling="2026-02-27 08:23:25.637891834 +0000 UTC m=+2063.491821832" lastFinishedPulling="2026-02-27 08:23:30.113672175 +0000 UTC m=+2067.967602213" observedRunningTime="2026-02-27 08:23:30.721618822 +0000 UTC m=+2068.575548890" watchObservedRunningTime="2026-02-27 08:23:30.73895134 +0000 UTC m=+2068.592881358" Feb 27 08:23:31 crc kubenswrapper[4612]: I0227 08:23:31.707533 4612 generic.go:334] "Generic (PLEG): container finished" podID="d99b6ff5-a143-40ce-bee0-d4300122aff2" containerID="0e66dd4c62bb4495706bb07100657019047e3dc7c5edf4f7a6cc9950150660c4" exitCode=0 Feb 27 08:23:31 crc kubenswrapper[4612]: I0227 08:23:31.707577 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-2g86c" event={"ID":"d99b6ff5-a143-40ce-bee0-d4300122aff2","Type":"ContainerDied","Data":"0e66dd4c62bb4495706bb07100657019047e3dc7c5edf4f7a6cc9950150660c4"} Feb 27 08:23:33 crc kubenswrapper[4612]: I0227 08:23:33.112735 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-2g86c" Feb 27 08:23:33 crc kubenswrapper[4612]: I0227 08:23:33.211559 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/d99b6ff5-a143-40ce-bee0-d4300122aff2-ssh-key-openstack-edpm-ipam\") pod \"d99b6ff5-a143-40ce-bee0-d4300122aff2\" (UID: \"d99b6ff5-a143-40ce-bee0-d4300122aff2\") " Feb 27 08:23:33 crc kubenswrapper[4612]: I0227 08:23:33.211708 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-97drj\" (UniqueName: \"kubernetes.io/projected/d99b6ff5-a143-40ce-bee0-d4300122aff2-kube-api-access-97drj\") pod \"d99b6ff5-a143-40ce-bee0-d4300122aff2\" (UID: \"d99b6ff5-a143-40ce-bee0-d4300122aff2\") " Feb 27 08:23:33 crc kubenswrapper[4612]: I0227 08:23:33.211958 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/d99b6ff5-a143-40ce-bee0-d4300122aff2-inventory-0\") pod \"d99b6ff5-a143-40ce-bee0-d4300122aff2\" (UID: \"d99b6ff5-a143-40ce-bee0-d4300122aff2\") " Feb 27 08:23:33 crc kubenswrapper[4612]: I0227 08:23:33.219351 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d99b6ff5-a143-40ce-bee0-d4300122aff2-kube-api-access-97drj" (OuterVolumeSpecName: "kube-api-access-97drj") pod "d99b6ff5-a143-40ce-bee0-d4300122aff2" (UID: "d99b6ff5-a143-40ce-bee0-d4300122aff2"). InnerVolumeSpecName "kube-api-access-97drj". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 08:23:33 crc kubenswrapper[4612]: I0227 08:23:33.248540 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d99b6ff5-a143-40ce-bee0-d4300122aff2-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "d99b6ff5-a143-40ce-bee0-d4300122aff2" (UID: "d99b6ff5-a143-40ce-bee0-d4300122aff2"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:23:33 crc kubenswrapper[4612]: I0227 08:23:33.263329 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d99b6ff5-a143-40ce-bee0-d4300122aff2-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "d99b6ff5-a143-40ce-bee0-d4300122aff2" (UID: "d99b6ff5-a143-40ce-bee0-d4300122aff2"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:23:33 crc kubenswrapper[4612]: I0227 08:23:33.314326 4612 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/d99b6ff5-a143-40ce-bee0-d4300122aff2-inventory-0\") on node \"crc\" DevicePath \"\"" Feb 27 08:23:33 crc kubenswrapper[4612]: I0227 08:23:33.314579 4612 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/d99b6ff5-a143-40ce-bee0-d4300122aff2-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 27 08:23:33 crc kubenswrapper[4612]: I0227 08:23:33.314592 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-97drj\" (UniqueName: \"kubernetes.io/projected/d99b6ff5-a143-40ce-bee0-d4300122aff2-kube-api-access-97drj\") on node \"crc\" DevicePath \"\"" Feb 27 08:23:33 crc kubenswrapper[4612]: I0227 08:23:33.731892 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-2g86c" event={"ID":"d99b6ff5-a143-40ce-bee0-d4300122aff2","Type":"ContainerDied","Data":"625f6903e977284d8e88f8ab47b2ed5b9460ce4935126675804fb68069ef3b52"} Feb 27 08:23:33 crc kubenswrapper[4612]: I0227 08:23:33.731954 4612 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="625f6903e977284d8e88f8ab47b2ed5b9460ce4935126675804fb68069ef3b52" Feb 27 08:23:33 crc kubenswrapper[4612]: I0227 08:23:33.732012 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-2g86c" Feb 27 08:23:33 crc kubenswrapper[4612]: I0227 08:23:33.827978 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-psmxx"] Feb 27 08:23:33 crc kubenswrapper[4612]: E0227 08:23:33.828419 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d99b6ff5-a143-40ce-bee0-d4300122aff2" containerName="ssh-known-hosts-edpm-deployment" Feb 27 08:23:33 crc kubenswrapper[4612]: I0227 08:23:33.828439 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="d99b6ff5-a143-40ce-bee0-d4300122aff2" containerName="ssh-known-hosts-edpm-deployment" Feb 27 08:23:33 crc kubenswrapper[4612]: I0227 08:23:33.828686 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="d99b6ff5-a143-40ce-bee0-d4300122aff2" containerName="ssh-known-hosts-edpm-deployment" Feb 27 08:23:33 crc kubenswrapper[4612]: I0227 08:23:33.829394 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-psmxx" Feb 27 08:23:33 crc kubenswrapper[4612]: I0227 08:23:33.833855 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 27 08:23:33 crc kubenswrapper[4612]: I0227 08:23:33.834211 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 27 08:23:33 crc kubenswrapper[4612]: I0227 08:23:33.834638 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 27 08:23:33 crc kubenswrapper[4612]: I0227 08:23:33.834826 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-2pwrb" Feb 27 08:23:33 crc kubenswrapper[4612]: I0227 08:23:33.844764 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-psmxx"] Feb 27 08:23:33 crc kubenswrapper[4612]: I0227 08:23:33.926833 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/340ef584-802e-4696-b1b2-c7c8be815482-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-psmxx\" (UID: \"340ef584-802e-4696-b1b2-c7c8be815482\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-psmxx" Feb 27 08:23:33 crc kubenswrapper[4612]: I0227 08:23:33.927029 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/340ef584-802e-4696-b1b2-c7c8be815482-ssh-key-openstack-edpm-ipam\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-psmxx\" (UID: \"340ef584-802e-4696-b1b2-c7c8be815482\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-psmxx" Feb 27 08:23:33 crc kubenswrapper[4612]: I0227 08:23:33.927322 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ww7gc\" (UniqueName: \"kubernetes.io/projected/340ef584-802e-4696-b1b2-c7c8be815482-kube-api-access-ww7gc\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-psmxx\" (UID: \"340ef584-802e-4696-b1b2-c7c8be815482\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-psmxx" Feb 27 08:23:34 crc kubenswrapper[4612]: I0227 08:23:34.029176 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/340ef584-802e-4696-b1b2-c7c8be815482-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-psmxx\" (UID: \"340ef584-802e-4696-b1b2-c7c8be815482\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-psmxx" Feb 27 08:23:34 crc kubenswrapper[4612]: I0227 08:23:34.029496 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/340ef584-802e-4696-b1b2-c7c8be815482-ssh-key-openstack-edpm-ipam\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-psmxx\" (UID: \"340ef584-802e-4696-b1b2-c7c8be815482\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-psmxx" Feb 27 08:23:34 crc kubenswrapper[4612]: I0227 08:23:34.029678 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ww7gc\" (UniqueName: \"kubernetes.io/projected/340ef584-802e-4696-b1b2-c7c8be815482-kube-api-access-ww7gc\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-psmxx\" (UID: \"340ef584-802e-4696-b1b2-c7c8be815482\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-psmxx" Feb 27 08:23:34 crc kubenswrapper[4612]: I0227 08:23:34.034953 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/340ef584-802e-4696-b1b2-c7c8be815482-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-psmxx\" (UID: \"340ef584-802e-4696-b1b2-c7c8be815482\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-psmxx" Feb 27 08:23:34 crc kubenswrapper[4612]: I0227 08:23:34.042604 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/340ef584-802e-4696-b1b2-c7c8be815482-ssh-key-openstack-edpm-ipam\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-psmxx\" (UID: \"340ef584-802e-4696-b1b2-c7c8be815482\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-psmxx" Feb 27 08:23:34 crc kubenswrapper[4612]: I0227 08:23:34.052922 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ww7gc\" (UniqueName: \"kubernetes.io/projected/340ef584-802e-4696-b1b2-c7c8be815482-kube-api-access-ww7gc\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-psmxx\" (UID: \"340ef584-802e-4696-b1b2-c7c8be815482\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-psmxx" Feb 27 08:23:34 crc kubenswrapper[4612]: I0227 08:23:34.151594 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-psmxx" Feb 27 08:23:34 crc kubenswrapper[4612]: I0227 08:23:34.336001 4612 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-2lp2b" Feb 27 08:23:34 crc kubenswrapper[4612]: I0227 08:23:34.336312 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-2lp2b" Feb 27 08:23:34 crc kubenswrapper[4612]: I0227 08:23:34.388809 4612 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-2lp2b" Feb 27 08:23:34 crc kubenswrapper[4612]: I0227 08:23:34.783113 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-psmxx"] Feb 27 08:23:34 crc kubenswrapper[4612]: W0227 08:23:34.787778 4612 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod340ef584_802e_4696_b1b2_c7c8be815482.slice/crio-f4833e1bbeac80a5d936a2fb3abce82ba85f54a58e4ce0f8424adf74fe79eeea WatchSource:0}: Error finding container f4833e1bbeac80a5d936a2fb3abce82ba85f54a58e4ce0f8424adf74fe79eeea: Status 404 returned error can't find the container with id f4833e1bbeac80a5d936a2fb3abce82ba85f54a58e4ce0f8424adf74fe79eeea Feb 27 08:23:35 crc kubenswrapper[4612]: I0227 08:23:35.750845 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-psmxx" event={"ID":"340ef584-802e-4696-b1b2-c7c8be815482","Type":"ContainerStarted","Data":"4a09c3d75d7dd4ae786f01cf48688adfe2e3163a8f0e6b6c12aa442ffad3b49e"} Feb 27 08:23:35 crc kubenswrapper[4612]: I0227 08:23:35.751080 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-psmxx" event={"ID":"340ef584-802e-4696-b1b2-c7c8be815482","Type":"ContainerStarted","Data":"f4833e1bbeac80a5d936a2fb3abce82ba85f54a58e4ce0f8424adf74fe79eeea"} Feb 27 08:23:35 crc kubenswrapper[4612]: I0227 08:23:35.770735 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-psmxx" podStartSLOduration=2.490166276 podStartE2EDuration="2.770718328s" podCreationTimestamp="2026-02-27 08:23:33 +0000 UTC" firstStartedPulling="2026-02-27 08:23:34.800516663 +0000 UTC m=+2072.654446661" lastFinishedPulling="2026-02-27 08:23:35.081068725 +0000 UTC m=+2072.934998713" observedRunningTime="2026-02-27 08:23:35.765298792 +0000 UTC m=+2073.619228790" watchObservedRunningTime="2026-02-27 08:23:35.770718328 +0000 UTC m=+2073.624648326" Feb 27 08:23:43 crc kubenswrapper[4612]: I0227 08:23:43.844391 4612 generic.go:334] "Generic (PLEG): container finished" podID="340ef584-802e-4696-b1b2-c7c8be815482" containerID="4a09c3d75d7dd4ae786f01cf48688adfe2e3163a8f0e6b6c12aa442ffad3b49e" exitCode=0 Feb 27 08:23:43 crc kubenswrapper[4612]: I0227 08:23:43.844544 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-psmxx" event={"ID":"340ef584-802e-4696-b1b2-c7c8be815482","Type":"ContainerDied","Data":"4a09c3d75d7dd4ae786f01cf48688adfe2e3163a8f0e6b6c12aa442ffad3b49e"} Feb 27 08:23:44 crc kubenswrapper[4612]: I0227 08:23:44.391769 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-2lp2b" Feb 27 08:23:44 crc kubenswrapper[4612]: I0227 08:23:44.444185 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-2lp2b"] Feb 27 08:23:44 crc kubenswrapper[4612]: I0227 08:23:44.852501 4612 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-2lp2b" podUID="415215a1-2901-462b-aeff-10f96bfa8344" containerName="registry-server" containerID="cri-o://5ed88a5b3997032d25d78a971b78bfb34a4fd50c61537093407649b14d617a73" gracePeriod=2 Feb 27 08:23:45 crc kubenswrapper[4612]: I0227 08:23:45.430378 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-psmxx" Feb 27 08:23:45 crc kubenswrapper[4612]: I0227 08:23:45.439253 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2lp2b" Feb 27 08:23:45 crc kubenswrapper[4612]: I0227 08:23:45.583208 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ww7gc\" (UniqueName: \"kubernetes.io/projected/340ef584-802e-4696-b1b2-c7c8be815482-kube-api-access-ww7gc\") pod \"340ef584-802e-4696-b1b2-c7c8be815482\" (UID: \"340ef584-802e-4696-b1b2-c7c8be815482\") " Feb 27 08:23:45 crc kubenswrapper[4612]: I0227 08:23:45.583288 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/340ef584-802e-4696-b1b2-c7c8be815482-inventory\") pod \"340ef584-802e-4696-b1b2-c7c8be815482\" (UID: \"340ef584-802e-4696-b1b2-c7c8be815482\") " Feb 27 08:23:45 crc kubenswrapper[4612]: I0227 08:23:45.583353 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8r8f8\" (UniqueName: \"kubernetes.io/projected/415215a1-2901-462b-aeff-10f96bfa8344-kube-api-access-8r8f8\") pod \"415215a1-2901-462b-aeff-10f96bfa8344\" (UID: \"415215a1-2901-462b-aeff-10f96bfa8344\") " Feb 27 08:23:45 crc kubenswrapper[4612]: I0227 08:23:45.583386 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/415215a1-2901-462b-aeff-10f96bfa8344-catalog-content\") pod \"415215a1-2901-462b-aeff-10f96bfa8344\" (UID: \"415215a1-2901-462b-aeff-10f96bfa8344\") " Feb 27 08:23:45 crc kubenswrapper[4612]: I0227 08:23:45.583426 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/415215a1-2901-462b-aeff-10f96bfa8344-utilities\") pod \"415215a1-2901-462b-aeff-10f96bfa8344\" (UID: \"415215a1-2901-462b-aeff-10f96bfa8344\") " Feb 27 08:23:45 crc kubenswrapper[4612]: I0227 08:23:45.583580 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/340ef584-802e-4696-b1b2-c7c8be815482-ssh-key-openstack-edpm-ipam\") pod \"340ef584-802e-4696-b1b2-c7c8be815482\" (UID: \"340ef584-802e-4696-b1b2-c7c8be815482\") " Feb 27 08:23:45 crc kubenswrapper[4612]: I0227 08:23:45.584781 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/415215a1-2901-462b-aeff-10f96bfa8344-utilities" (OuterVolumeSpecName: "utilities") pod "415215a1-2901-462b-aeff-10f96bfa8344" (UID: "415215a1-2901-462b-aeff-10f96bfa8344"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 08:23:45 crc kubenswrapper[4612]: I0227 08:23:45.585063 4612 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/415215a1-2901-462b-aeff-10f96bfa8344-utilities\") on node \"crc\" DevicePath \"\"" Feb 27 08:23:45 crc kubenswrapper[4612]: I0227 08:23:45.589248 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/340ef584-802e-4696-b1b2-c7c8be815482-kube-api-access-ww7gc" (OuterVolumeSpecName: "kube-api-access-ww7gc") pod "340ef584-802e-4696-b1b2-c7c8be815482" (UID: "340ef584-802e-4696-b1b2-c7c8be815482"). InnerVolumeSpecName "kube-api-access-ww7gc". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 08:23:45 crc kubenswrapper[4612]: I0227 08:23:45.592058 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/415215a1-2901-462b-aeff-10f96bfa8344-kube-api-access-8r8f8" (OuterVolumeSpecName: "kube-api-access-8r8f8") pod "415215a1-2901-462b-aeff-10f96bfa8344" (UID: "415215a1-2901-462b-aeff-10f96bfa8344"). InnerVolumeSpecName "kube-api-access-8r8f8". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 08:23:45 crc kubenswrapper[4612]: I0227 08:23:45.624833 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/340ef584-802e-4696-b1b2-c7c8be815482-inventory" (OuterVolumeSpecName: "inventory") pod "340ef584-802e-4696-b1b2-c7c8be815482" (UID: "340ef584-802e-4696-b1b2-c7c8be815482"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:23:45 crc kubenswrapper[4612]: I0227 08:23:45.647947 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/340ef584-802e-4696-b1b2-c7c8be815482-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "340ef584-802e-4696-b1b2-c7c8be815482" (UID: "340ef584-802e-4696-b1b2-c7c8be815482"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:23:45 crc kubenswrapper[4612]: I0227 08:23:45.667284 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/415215a1-2901-462b-aeff-10f96bfa8344-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "415215a1-2901-462b-aeff-10f96bfa8344" (UID: "415215a1-2901-462b-aeff-10f96bfa8344"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 08:23:45 crc kubenswrapper[4612]: I0227 08:23:45.687302 4612 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/340ef584-802e-4696-b1b2-c7c8be815482-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 27 08:23:45 crc kubenswrapper[4612]: I0227 08:23:45.687569 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ww7gc\" (UniqueName: \"kubernetes.io/projected/340ef584-802e-4696-b1b2-c7c8be815482-kube-api-access-ww7gc\") on node \"crc\" DevicePath \"\"" Feb 27 08:23:45 crc kubenswrapper[4612]: I0227 08:23:45.687652 4612 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/340ef584-802e-4696-b1b2-c7c8be815482-inventory\") on node \"crc\" DevicePath \"\"" Feb 27 08:23:45 crc kubenswrapper[4612]: I0227 08:23:45.687835 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8r8f8\" (UniqueName: \"kubernetes.io/projected/415215a1-2901-462b-aeff-10f96bfa8344-kube-api-access-8r8f8\") on node \"crc\" DevicePath \"\"" Feb 27 08:23:45 crc kubenswrapper[4612]: I0227 08:23:45.687930 4612 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/415215a1-2901-462b-aeff-10f96bfa8344-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 27 08:23:45 crc kubenswrapper[4612]: I0227 08:23:45.860591 4612 generic.go:334] "Generic (PLEG): container finished" podID="415215a1-2901-462b-aeff-10f96bfa8344" containerID="5ed88a5b3997032d25d78a971b78bfb34a4fd50c61537093407649b14d617a73" exitCode=0 Feb 27 08:23:45 crc kubenswrapper[4612]: I0227 08:23:45.860653 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2lp2b" Feb 27 08:23:45 crc kubenswrapper[4612]: I0227 08:23:45.860667 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2lp2b" event={"ID":"415215a1-2901-462b-aeff-10f96bfa8344","Type":"ContainerDied","Data":"5ed88a5b3997032d25d78a971b78bfb34a4fd50c61537093407649b14d617a73"} Feb 27 08:23:45 crc kubenswrapper[4612]: I0227 08:23:45.860746 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2lp2b" event={"ID":"415215a1-2901-462b-aeff-10f96bfa8344","Type":"ContainerDied","Data":"f7d0488fe4d7c232c2197a1bed4b54fcd1cdf791a2d6a745af35c3db867a4e57"} Feb 27 08:23:45 crc kubenswrapper[4612]: I0227 08:23:45.860768 4612 scope.go:117] "RemoveContainer" containerID="5ed88a5b3997032d25d78a971b78bfb34a4fd50c61537093407649b14d617a73" Feb 27 08:23:45 crc kubenswrapper[4612]: I0227 08:23:45.863430 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-psmxx" event={"ID":"340ef584-802e-4696-b1b2-c7c8be815482","Type":"ContainerDied","Data":"f4833e1bbeac80a5d936a2fb3abce82ba85f54a58e4ce0f8424adf74fe79eeea"} Feb 27 08:23:45 crc kubenswrapper[4612]: I0227 08:23:45.863561 4612 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f4833e1bbeac80a5d936a2fb3abce82ba85f54a58e4ce0f8424adf74fe79eeea" Feb 27 08:23:45 crc kubenswrapper[4612]: I0227 08:23:45.863479 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-psmxx" Feb 27 08:23:45 crc kubenswrapper[4612]: I0227 08:23:45.880919 4612 scope.go:117] "RemoveContainer" containerID="ea7c0654c0cc16bdd192573ab2e11f43fbd6e7bcd7d5d55720e68e9190790653" Feb 27 08:23:45 crc kubenswrapper[4612]: I0227 08:23:45.921923 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-2lp2b"] Feb 27 08:23:45 crc kubenswrapper[4612]: I0227 08:23:45.925256 4612 scope.go:117] "RemoveContainer" containerID="4588e54cc61d60c9ce5cb1e03eb9e4df62f162d261f0e0ba89dfaa765ec91fd3" Feb 27 08:23:45 crc kubenswrapper[4612]: I0227 08:23:45.958080 4612 scope.go:117] "RemoveContainer" containerID="5ed88a5b3997032d25d78a971b78bfb34a4fd50c61537093407649b14d617a73" Feb 27 08:23:45 crc kubenswrapper[4612]: I0227 08:23:45.958201 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-2lp2b"] Feb 27 08:23:45 crc kubenswrapper[4612]: E0227 08:23:45.964305 4612 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod340ef584_802e_4696_b1b2_c7c8be815482.slice\": RecentStats: unable to find data in memory cache]" Feb 27 08:23:45 crc kubenswrapper[4612]: E0227 08:23:45.968492 4612 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5ed88a5b3997032d25d78a971b78bfb34a4fd50c61537093407649b14d617a73\": container with ID starting with 5ed88a5b3997032d25d78a971b78bfb34a4fd50c61537093407649b14d617a73 not found: ID does not exist" containerID="5ed88a5b3997032d25d78a971b78bfb34a4fd50c61537093407649b14d617a73" Feb 27 08:23:45 crc kubenswrapper[4612]: I0227 08:23:45.968617 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5ed88a5b3997032d25d78a971b78bfb34a4fd50c61537093407649b14d617a73"} err="failed to get container status \"5ed88a5b3997032d25d78a971b78bfb34a4fd50c61537093407649b14d617a73\": rpc error: code = NotFound desc = could not find container \"5ed88a5b3997032d25d78a971b78bfb34a4fd50c61537093407649b14d617a73\": container with ID starting with 5ed88a5b3997032d25d78a971b78bfb34a4fd50c61537093407649b14d617a73 not found: ID does not exist" Feb 27 08:23:45 crc kubenswrapper[4612]: I0227 08:23:45.968725 4612 scope.go:117] "RemoveContainer" containerID="ea7c0654c0cc16bdd192573ab2e11f43fbd6e7bcd7d5d55720e68e9190790653" Feb 27 08:23:45 crc kubenswrapper[4612]: E0227 08:23:45.969017 4612 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ea7c0654c0cc16bdd192573ab2e11f43fbd6e7bcd7d5d55720e68e9190790653\": container with ID starting with ea7c0654c0cc16bdd192573ab2e11f43fbd6e7bcd7d5d55720e68e9190790653 not found: ID does not exist" containerID="ea7c0654c0cc16bdd192573ab2e11f43fbd6e7bcd7d5d55720e68e9190790653" Feb 27 08:23:45 crc kubenswrapper[4612]: I0227 08:23:45.969101 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ea7c0654c0cc16bdd192573ab2e11f43fbd6e7bcd7d5d55720e68e9190790653"} err="failed to get container status \"ea7c0654c0cc16bdd192573ab2e11f43fbd6e7bcd7d5d55720e68e9190790653\": rpc error: code = NotFound desc = could not find container \"ea7c0654c0cc16bdd192573ab2e11f43fbd6e7bcd7d5d55720e68e9190790653\": container with ID starting with ea7c0654c0cc16bdd192573ab2e11f43fbd6e7bcd7d5d55720e68e9190790653 not found: ID does not exist" Feb 27 08:23:45 crc kubenswrapper[4612]: I0227 08:23:45.969175 4612 scope.go:117] "RemoveContainer" containerID="4588e54cc61d60c9ce5cb1e03eb9e4df62f162d261f0e0ba89dfaa765ec91fd3" Feb 27 08:23:45 crc kubenswrapper[4612]: E0227 08:23:45.969563 4612 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4588e54cc61d60c9ce5cb1e03eb9e4df62f162d261f0e0ba89dfaa765ec91fd3\": container with ID starting with 4588e54cc61d60c9ce5cb1e03eb9e4df62f162d261f0e0ba89dfaa765ec91fd3 not found: ID does not exist" containerID="4588e54cc61d60c9ce5cb1e03eb9e4df62f162d261f0e0ba89dfaa765ec91fd3" Feb 27 08:23:45 crc kubenswrapper[4612]: I0227 08:23:45.969604 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4588e54cc61d60c9ce5cb1e03eb9e4df62f162d261f0e0ba89dfaa765ec91fd3"} err="failed to get container status \"4588e54cc61d60c9ce5cb1e03eb9e4df62f162d261f0e0ba89dfaa765ec91fd3\": rpc error: code = NotFound desc = could not find container \"4588e54cc61d60c9ce5cb1e03eb9e4df62f162d261f0e0ba89dfaa765ec91fd3\": container with ID starting with 4588e54cc61d60c9ce5cb1e03eb9e4df62f162d261f0e0ba89dfaa765ec91fd3 not found: ID does not exist" Feb 27 08:23:45 crc kubenswrapper[4612]: I0227 08:23:45.979714 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-cl6sd"] Feb 27 08:23:45 crc kubenswrapper[4612]: E0227 08:23:45.980266 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="415215a1-2901-462b-aeff-10f96bfa8344" containerName="registry-server" Feb 27 08:23:45 crc kubenswrapper[4612]: I0227 08:23:45.980381 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="415215a1-2901-462b-aeff-10f96bfa8344" containerName="registry-server" Feb 27 08:23:45 crc kubenswrapper[4612]: E0227 08:23:45.980441 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="415215a1-2901-462b-aeff-10f96bfa8344" containerName="extract-content" Feb 27 08:23:45 crc kubenswrapper[4612]: I0227 08:23:45.980498 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="415215a1-2901-462b-aeff-10f96bfa8344" containerName="extract-content" Feb 27 08:23:45 crc kubenswrapper[4612]: E0227 08:23:45.980566 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="415215a1-2901-462b-aeff-10f96bfa8344" containerName="extract-utilities" Feb 27 08:23:45 crc kubenswrapper[4612]: I0227 08:23:45.980625 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="415215a1-2901-462b-aeff-10f96bfa8344" containerName="extract-utilities" Feb 27 08:23:45 crc kubenswrapper[4612]: E0227 08:23:45.980706 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="340ef584-802e-4696-b1b2-c7c8be815482" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Feb 27 08:23:45 crc kubenswrapper[4612]: I0227 08:23:45.980766 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="340ef584-802e-4696-b1b2-c7c8be815482" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Feb 27 08:23:45 crc kubenswrapper[4612]: I0227 08:23:45.980972 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="340ef584-802e-4696-b1b2-c7c8be815482" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Feb 27 08:23:45 crc kubenswrapper[4612]: I0227 08:23:45.981045 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="415215a1-2901-462b-aeff-10f96bfa8344" containerName="registry-server" Feb 27 08:23:45 crc kubenswrapper[4612]: I0227 08:23:45.981562 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-cl6sd"] Feb 27 08:23:45 crc kubenswrapper[4612]: I0227 08:23:45.981711 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-cl6sd" Feb 27 08:23:45 crc kubenswrapper[4612]: I0227 08:23:45.984347 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-2pwrb" Feb 27 08:23:45 crc kubenswrapper[4612]: I0227 08:23:45.989047 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 27 08:23:45 crc kubenswrapper[4612]: I0227 08:23:45.989199 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 27 08:23:45 crc kubenswrapper[4612]: I0227 08:23:45.989338 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 27 08:23:46 crc kubenswrapper[4612]: I0227 08:23:46.026535 4612 patch_prober.go:28] interesting pod/machine-config-daemon-924vb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 27 08:23:46 crc kubenswrapper[4612]: I0227 08:23:46.026701 4612 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 27 08:23:46 crc kubenswrapper[4612]: I0227 08:23:46.026801 4612 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-924vb" Feb 27 08:23:46 crc kubenswrapper[4612]: I0227 08:23:46.027567 4612 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"f80ae3b7ec2dce55e073a92197c42ea49b624e8fd3575dd785e408adfa1598e5"} pod="openshift-machine-config-operator/machine-config-daemon-924vb" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 27 08:23:46 crc kubenswrapper[4612]: I0227 08:23:46.027713 4612 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" containerName="machine-config-daemon" containerID="cri-o://f80ae3b7ec2dce55e073a92197c42ea49b624e8fd3575dd785e408adfa1598e5" gracePeriod=600 Feb 27 08:23:46 crc kubenswrapper[4612]: I0227 08:23:46.097133 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ac2d7b40-cd87-4726-8225-6c9ce8afb4d7-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-cl6sd\" (UID: \"ac2d7b40-cd87-4726-8225-6c9ce8afb4d7\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-cl6sd" Feb 27 08:23:46 crc kubenswrapper[4612]: I0227 08:23:46.097578 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/ac2d7b40-cd87-4726-8225-6c9ce8afb4d7-ssh-key-openstack-edpm-ipam\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-cl6sd\" (UID: \"ac2d7b40-cd87-4726-8225-6c9ce8afb4d7\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-cl6sd" Feb 27 08:23:46 crc kubenswrapper[4612]: I0227 08:23:46.097670 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4s246\" (UniqueName: \"kubernetes.io/projected/ac2d7b40-cd87-4726-8225-6c9ce8afb4d7-kube-api-access-4s246\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-cl6sd\" (UID: \"ac2d7b40-cd87-4726-8225-6c9ce8afb4d7\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-cl6sd" Feb 27 08:23:46 crc kubenswrapper[4612]: I0227 08:23:46.199509 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ac2d7b40-cd87-4726-8225-6c9ce8afb4d7-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-cl6sd\" (UID: \"ac2d7b40-cd87-4726-8225-6c9ce8afb4d7\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-cl6sd" Feb 27 08:23:46 crc kubenswrapper[4612]: I0227 08:23:46.199606 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/ac2d7b40-cd87-4726-8225-6c9ce8afb4d7-ssh-key-openstack-edpm-ipam\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-cl6sd\" (UID: \"ac2d7b40-cd87-4726-8225-6c9ce8afb4d7\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-cl6sd" Feb 27 08:23:46 crc kubenswrapper[4612]: I0227 08:23:46.199707 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4s246\" (UniqueName: \"kubernetes.io/projected/ac2d7b40-cd87-4726-8225-6c9ce8afb4d7-kube-api-access-4s246\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-cl6sd\" (UID: \"ac2d7b40-cd87-4726-8225-6c9ce8afb4d7\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-cl6sd" Feb 27 08:23:46 crc kubenswrapper[4612]: I0227 08:23:46.207539 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/ac2d7b40-cd87-4726-8225-6c9ce8afb4d7-ssh-key-openstack-edpm-ipam\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-cl6sd\" (UID: \"ac2d7b40-cd87-4726-8225-6c9ce8afb4d7\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-cl6sd" Feb 27 08:23:46 crc kubenswrapper[4612]: I0227 08:23:46.210617 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ac2d7b40-cd87-4726-8225-6c9ce8afb4d7-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-cl6sd\" (UID: \"ac2d7b40-cd87-4726-8225-6c9ce8afb4d7\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-cl6sd" Feb 27 08:23:46 crc kubenswrapper[4612]: I0227 08:23:46.219653 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4s246\" (UniqueName: \"kubernetes.io/projected/ac2d7b40-cd87-4726-8225-6c9ce8afb4d7-kube-api-access-4s246\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-cl6sd\" (UID: \"ac2d7b40-cd87-4726-8225-6c9ce8afb4d7\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-cl6sd" Feb 27 08:23:46 crc kubenswrapper[4612]: I0227 08:23:46.331427 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-cl6sd" Feb 27 08:23:46 crc kubenswrapper[4612]: I0227 08:23:46.863237 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="415215a1-2901-462b-aeff-10f96bfa8344" path="/var/lib/kubelet/pods/415215a1-2901-462b-aeff-10f96bfa8344/volumes" Feb 27 08:23:46 crc kubenswrapper[4612]: I0227 08:23:46.874051 4612 generic.go:334] "Generic (PLEG): container finished" podID="24599373-7adc-4a1b-8bb4-797bf726f43d" containerID="f80ae3b7ec2dce55e073a92197c42ea49b624e8fd3575dd785e408adfa1598e5" exitCode=0 Feb 27 08:23:46 crc kubenswrapper[4612]: I0227 08:23:46.874100 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-924vb" event={"ID":"24599373-7adc-4a1b-8bb4-797bf726f43d","Type":"ContainerDied","Data":"f80ae3b7ec2dce55e073a92197c42ea49b624e8fd3575dd785e408adfa1598e5"} Feb 27 08:23:46 crc kubenswrapper[4612]: I0227 08:23:46.874132 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-924vb" event={"ID":"24599373-7adc-4a1b-8bb4-797bf726f43d","Type":"ContainerStarted","Data":"76c2de0055c53fc536f48b64de557ea4f7b56a2de457f8f1b8708b99c7c612cb"} Feb 27 08:23:46 crc kubenswrapper[4612]: I0227 08:23:46.874151 4612 scope.go:117] "RemoveContainer" containerID="7c62ebf970b4c78c5d31a0ce6596e9b398afbc8a6935511ce1b20bcd4f436b08" Feb 27 08:23:47 crc kubenswrapper[4612]: I0227 08:23:46.971341 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-cl6sd"] Feb 27 08:23:47 crc kubenswrapper[4612]: W0227 08:23:46.987723 4612 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podac2d7b40_cd87_4726_8225_6c9ce8afb4d7.slice/crio-a7762d42a3382a271af10f2b05f25c82528ada204505885da7530e600f90fbfd WatchSource:0}: Error finding container a7762d42a3382a271af10f2b05f25c82528ada204505885da7530e600f90fbfd: Status 404 returned error can't find the container with id a7762d42a3382a271af10f2b05f25c82528ada204505885da7530e600f90fbfd Feb 27 08:23:47 crc kubenswrapper[4612]: I0227 08:23:47.900436 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-cl6sd" event={"ID":"ac2d7b40-cd87-4726-8225-6c9ce8afb4d7","Type":"ContainerStarted","Data":"766a07a8fd80d950bb33e8966a269c3ed563e9f0b9a001777e62d2272deab5d3"} Feb 27 08:23:47 crc kubenswrapper[4612]: I0227 08:23:47.900905 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-cl6sd" event={"ID":"ac2d7b40-cd87-4726-8225-6c9ce8afb4d7","Type":"ContainerStarted","Data":"a7762d42a3382a271af10f2b05f25c82528ada204505885da7530e600f90fbfd"} Feb 27 08:23:47 crc kubenswrapper[4612]: I0227 08:23:47.929911 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-cl6sd" podStartSLOduration=2.630190814 podStartE2EDuration="2.929880075s" podCreationTimestamp="2026-02-27 08:23:45 +0000 UTC" firstStartedPulling="2026-02-27 08:23:46.993464162 +0000 UTC m=+2084.847394160" lastFinishedPulling="2026-02-27 08:23:47.293153423 +0000 UTC m=+2085.147083421" observedRunningTime="2026-02-27 08:23:47.923485302 +0000 UTC m=+2085.777415300" watchObservedRunningTime="2026-02-27 08:23:47.929880075 +0000 UTC m=+2085.783810093" Feb 27 08:23:56 crc kubenswrapper[4612]: I0227 08:23:56.992118 4612 generic.go:334] "Generic (PLEG): container finished" podID="ac2d7b40-cd87-4726-8225-6c9ce8afb4d7" containerID="766a07a8fd80d950bb33e8966a269c3ed563e9f0b9a001777e62d2272deab5d3" exitCode=0 Feb 27 08:23:56 crc kubenswrapper[4612]: I0227 08:23:56.992192 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-cl6sd" event={"ID":"ac2d7b40-cd87-4726-8225-6c9ce8afb4d7","Type":"ContainerDied","Data":"766a07a8fd80d950bb33e8966a269c3ed563e9f0b9a001777e62d2272deab5d3"} Feb 27 08:23:58 crc kubenswrapper[4612]: I0227 08:23:58.469192 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-cl6sd" Feb 27 08:23:58 crc kubenswrapper[4612]: I0227 08:23:58.541778 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ac2d7b40-cd87-4726-8225-6c9ce8afb4d7-inventory\") pod \"ac2d7b40-cd87-4726-8225-6c9ce8afb4d7\" (UID: \"ac2d7b40-cd87-4726-8225-6c9ce8afb4d7\") " Feb 27 08:23:58 crc kubenswrapper[4612]: I0227 08:23:58.541944 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4s246\" (UniqueName: \"kubernetes.io/projected/ac2d7b40-cd87-4726-8225-6c9ce8afb4d7-kube-api-access-4s246\") pod \"ac2d7b40-cd87-4726-8225-6c9ce8afb4d7\" (UID: \"ac2d7b40-cd87-4726-8225-6c9ce8afb4d7\") " Feb 27 08:23:58 crc kubenswrapper[4612]: I0227 08:23:58.541998 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/ac2d7b40-cd87-4726-8225-6c9ce8afb4d7-ssh-key-openstack-edpm-ipam\") pod \"ac2d7b40-cd87-4726-8225-6c9ce8afb4d7\" (UID: \"ac2d7b40-cd87-4726-8225-6c9ce8afb4d7\") " Feb 27 08:23:58 crc kubenswrapper[4612]: I0227 08:23:58.547832 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ac2d7b40-cd87-4726-8225-6c9ce8afb4d7-kube-api-access-4s246" (OuterVolumeSpecName: "kube-api-access-4s246") pod "ac2d7b40-cd87-4726-8225-6c9ce8afb4d7" (UID: "ac2d7b40-cd87-4726-8225-6c9ce8afb4d7"). InnerVolumeSpecName "kube-api-access-4s246". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 08:23:58 crc kubenswrapper[4612]: I0227 08:23:58.576979 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ac2d7b40-cd87-4726-8225-6c9ce8afb4d7-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "ac2d7b40-cd87-4726-8225-6c9ce8afb4d7" (UID: "ac2d7b40-cd87-4726-8225-6c9ce8afb4d7"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:23:58 crc kubenswrapper[4612]: I0227 08:23:58.586050 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ac2d7b40-cd87-4726-8225-6c9ce8afb4d7-inventory" (OuterVolumeSpecName: "inventory") pod "ac2d7b40-cd87-4726-8225-6c9ce8afb4d7" (UID: "ac2d7b40-cd87-4726-8225-6c9ce8afb4d7"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:23:58 crc kubenswrapper[4612]: I0227 08:23:58.643784 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4s246\" (UniqueName: \"kubernetes.io/projected/ac2d7b40-cd87-4726-8225-6c9ce8afb4d7-kube-api-access-4s246\") on node \"crc\" DevicePath \"\"" Feb 27 08:23:58 crc kubenswrapper[4612]: I0227 08:23:58.643817 4612 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/ac2d7b40-cd87-4726-8225-6c9ce8afb4d7-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 27 08:23:58 crc kubenswrapper[4612]: I0227 08:23:58.643827 4612 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ac2d7b40-cd87-4726-8225-6c9ce8afb4d7-inventory\") on node \"crc\" DevicePath \"\"" Feb 27 08:23:59 crc kubenswrapper[4612]: I0227 08:23:59.009104 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-cl6sd" event={"ID":"ac2d7b40-cd87-4726-8225-6c9ce8afb4d7","Type":"ContainerDied","Data":"a7762d42a3382a271af10f2b05f25c82528ada204505885da7530e600f90fbfd"} Feb 27 08:23:59 crc kubenswrapper[4612]: I0227 08:23:59.009472 4612 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a7762d42a3382a271af10f2b05f25c82528ada204505885da7530e600f90fbfd" Feb 27 08:23:59 crc kubenswrapper[4612]: I0227 08:23:59.009137 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-cl6sd" Feb 27 08:23:59 crc kubenswrapper[4612]: I0227 08:23:59.142330 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-frrmr"] Feb 27 08:23:59 crc kubenswrapper[4612]: E0227 08:23:59.142793 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac2d7b40-cd87-4726-8225-6c9ce8afb4d7" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Feb 27 08:23:59 crc kubenswrapper[4612]: I0227 08:23:59.142814 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac2d7b40-cd87-4726-8225-6c9ce8afb4d7" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Feb 27 08:23:59 crc kubenswrapper[4612]: I0227 08:23:59.143076 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="ac2d7b40-cd87-4726-8225-6c9ce8afb4d7" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Feb 27 08:23:59 crc kubenswrapper[4612]: I0227 08:23:59.143813 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-frrmr" Feb 27 08:23:59 crc kubenswrapper[4612]: I0227 08:23:59.152066 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 27 08:23:59 crc kubenswrapper[4612]: I0227 08:23:59.152943 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-2pwrb" Feb 27 08:23:59 crc kubenswrapper[4612]: I0227 08:23:59.153446 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-neutron-metadata-default-certs-0" Feb 27 08:23:59 crc kubenswrapper[4612]: I0227 08:23:59.153809 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-libvirt-default-certs-0" Feb 27 08:23:59 crc kubenswrapper[4612]: I0227 08:23:59.153843 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 27 08:23:59 crc kubenswrapper[4612]: I0227 08:23:59.153857 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 27 08:23:59 crc kubenswrapper[4612]: I0227 08:23:59.153847 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-telemetry-default-certs-0" Feb 27 08:23:59 crc kubenswrapper[4612]: I0227 08:23:59.155488 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-ovn-default-certs-0" Feb 27 08:23:59 crc kubenswrapper[4612]: I0227 08:23:59.155812 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-frrmr"] Feb 27 08:23:59 crc kubenswrapper[4612]: I0227 08:23:59.255739 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/4f248d51-18f9-4d20-b64b-d23765054818-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-frrmr\" (UID: \"4f248d51-18f9-4d20-b64b-d23765054818\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-frrmr" Feb 27 08:23:59 crc kubenswrapper[4612]: I0227 08:23:59.255784 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f248d51-18f9-4d20-b64b-d23765054818-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-frrmr\" (UID: \"4f248d51-18f9-4d20-b64b-d23765054818\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-frrmr" Feb 27 08:23:59 crc kubenswrapper[4612]: I0227 08:23:59.255838 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f248d51-18f9-4d20-b64b-d23765054818-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-frrmr\" (UID: \"4f248d51-18f9-4d20-b64b-d23765054818\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-frrmr" Feb 27 08:23:59 crc kubenswrapper[4612]: I0227 08:23:59.255860 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vkgn2\" (UniqueName: \"kubernetes.io/projected/4f248d51-18f9-4d20-b64b-d23765054818-kube-api-access-vkgn2\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-frrmr\" (UID: \"4f248d51-18f9-4d20-b64b-d23765054818\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-frrmr" Feb 27 08:23:59 crc kubenswrapper[4612]: I0227 08:23:59.255911 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/4f248d51-18f9-4d20-b64b-d23765054818-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-frrmr\" (UID: \"4f248d51-18f9-4d20-b64b-d23765054818\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-frrmr" Feb 27 08:23:59 crc kubenswrapper[4612]: I0227 08:23:59.255950 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/4f248d51-18f9-4d20-b64b-d23765054818-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-frrmr\" (UID: \"4f248d51-18f9-4d20-b64b-d23765054818\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-frrmr" Feb 27 08:23:59 crc kubenswrapper[4612]: I0227 08:23:59.256062 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f248d51-18f9-4d20-b64b-d23765054818-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-frrmr\" (UID: \"4f248d51-18f9-4d20-b64b-d23765054818\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-frrmr" Feb 27 08:23:59 crc kubenswrapper[4612]: I0227 08:23:59.256122 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f248d51-18f9-4d20-b64b-d23765054818-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-frrmr\" (UID: \"4f248d51-18f9-4d20-b64b-d23765054818\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-frrmr" Feb 27 08:23:59 crc kubenswrapper[4612]: I0227 08:23:59.256185 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f248d51-18f9-4d20-b64b-d23765054818-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-frrmr\" (UID: \"4f248d51-18f9-4d20-b64b-d23765054818\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-frrmr" Feb 27 08:23:59 crc kubenswrapper[4612]: I0227 08:23:59.256202 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f248d51-18f9-4d20-b64b-d23765054818-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-frrmr\" (UID: \"4f248d51-18f9-4d20-b64b-d23765054818\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-frrmr" Feb 27 08:23:59 crc kubenswrapper[4612]: I0227 08:23:59.256315 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f248d51-18f9-4d20-b64b-d23765054818-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-frrmr\" (UID: \"4f248d51-18f9-4d20-b64b-d23765054818\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-frrmr" Feb 27 08:23:59 crc kubenswrapper[4612]: I0227 08:23:59.256362 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4f248d51-18f9-4d20-b64b-d23765054818-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-frrmr\" (UID: \"4f248d51-18f9-4d20-b64b-d23765054818\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-frrmr" Feb 27 08:23:59 crc kubenswrapper[4612]: I0227 08:23:59.256427 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/4f248d51-18f9-4d20-b64b-d23765054818-ssh-key-openstack-edpm-ipam\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-frrmr\" (UID: \"4f248d51-18f9-4d20-b64b-d23765054818\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-frrmr" Feb 27 08:23:59 crc kubenswrapper[4612]: I0227 08:23:59.256517 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/4f248d51-18f9-4d20-b64b-d23765054818-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-frrmr\" (UID: \"4f248d51-18f9-4d20-b64b-d23765054818\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-frrmr" Feb 27 08:23:59 crc kubenswrapper[4612]: I0227 08:23:59.358421 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vkgn2\" (UniqueName: \"kubernetes.io/projected/4f248d51-18f9-4d20-b64b-d23765054818-kube-api-access-vkgn2\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-frrmr\" (UID: \"4f248d51-18f9-4d20-b64b-d23765054818\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-frrmr" Feb 27 08:23:59 crc kubenswrapper[4612]: I0227 08:23:59.358492 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/4f248d51-18f9-4d20-b64b-d23765054818-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-frrmr\" (UID: \"4f248d51-18f9-4d20-b64b-d23765054818\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-frrmr" Feb 27 08:23:59 crc kubenswrapper[4612]: I0227 08:23:59.358535 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/4f248d51-18f9-4d20-b64b-d23765054818-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-frrmr\" (UID: \"4f248d51-18f9-4d20-b64b-d23765054818\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-frrmr" Feb 27 08:23:59 crc kubenswrapper[4612]: I0227 08:23:59.358573 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f248d51-18f9-4d20-b64b-d23765054818-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-frrmr\" (UID: \"4f248d51-18f9-4d20-b64b-d23765054818\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-frrmr" Feb 27 08:23:59 crc kubenswrapper[4612]: I0227 08:23:59.358610 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f248d51-18f9-4d20-b64b-d23765054818-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-frrmr\" (UID: \"4f248d51-18f9-4d20-b64b-d23765054818\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-frrmr" Feb 27 08:23:59 crc kubenswrapper[4612]: I0227 08:23:59.358647 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f248d51-18f9-4d20-b64b-d23765054818-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-frrmr\" (UID: \"4f248d51-18f9-4d20-b64b-d23765054818\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-frrmr" Feb 27 08:23:59 crc kubenswrapper[4612]: I0227 08:23:59.358663 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f248d51-18f9-4d20-b64b-d23765054818-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-frrmr\" (UID: \"4f248d51-18f9-4d20-b64b-d23765054818\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-frrmr" Feb 27 08:23:59 crc kubenswrapper[4612]: I0227 08:23:59.358717 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f248d51-18f9-4d20-b64b-d23765054818-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-frrmr\" (UID: \"4f248d51-18f9-4d20-b64b-d23765054818\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-frrmr" Feb 27 08:23:59 crc kubenswrapper[4612]: I0227 08:23:59.358741 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4f248d51-18f9-4d20-b64b-d23765054818-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-frrmr\" (UID: \"4f248d51-18f9-4d20-b64b-d23765054818\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-frrmr" Feb 27 08:23:59 crc kubenswrapper[4612]: I0227 08:23:59.358771 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/4f248d51-18f9-4d20-b64b-d23765054818-ssh-key-openstack-edpm-ipam\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-frrmr\" (UID: \"4f248d51-18f9-4d20-b64b-d23765054818\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-frrmr" Feb 27 08:23:59 crc kubenswrapper[4612]: I0227 08:23:59.358808 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/4f248d51-18f9-4d20-b64b-d23765054818-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-frrmr\" (UID: \"4f248d51-18f9-4d20-b64b-d23765054818\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-frrmr" Feb 27 08:23:59 crc kubenswrapper[4612]: I0227 08:23:59.358847 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/4f248d51-18f9-4d20-b64b-d23765054818-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-frrmr\" (UID: \"4f248d51-18f9-4d20-b64b-d23765054818\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-frrmr" Feb 27 08:23:59 crc kubenswrapper[4612]: I0227 08:23:59.358869 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f248d51-18f9-4d20-b64b-d23765054818-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-frrmr\" (UID: \"4f248d51-18f9-4d20-b64b-d23765054818\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-frrmr" Feb 27 08:23:59 crc kubenswrapper[4612]: I0227 08:23:59.358900 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f248d51-18f9-4d20-b64b-d23765054818-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-frrmr\" (UID: \"4f248d51-18f9-4d20-b64b-d23765054818\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-frrmr" Feb 27 08:23:59 crc kubenswrapper[4612]: I0227 08:23:59.364372 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f248d51-18f9-4d20-b64b-d23765054818-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-frrmr\" (UID: \"4f248d51-18f9-4d20-b64b-d23765054818\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-frrmr" Feb 27 08:23:59 crc kubenswrapper[4612]: I0227 08:23:59.365121 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f248d51-18f9-4d20-b64b-d23765054818-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-frrmr\" (UID: \"4f248d51-18f9-4d20-b64b-d23765054818\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-frrmr" Feb 27 08:23:59 crc kubenswrapper[4612]: I0227 08:23:59.367993 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f248d51-18f9-4d20-b64b-d23765054818-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-frrmr\" (UID: \"4f248d51-18f9-4d20-b64b-d23765054818\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-frrmr" Feb 27 08:23:59 crc kubenswrapper[4612]: I0227 08:23:59.368146 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f248d51-18f9-4d20-b64b-d23765054818-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-frrmr\" (UID: \"4f248d51-18f9-4d20-b64b-d23765054818\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-frrmr" Feb 27 08:23:59 crc kubenswrapper[4612]: I0227 08:23:59.368960 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/4f248d51-18f9-4d20-b64b-d23765054818-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-frrmr\" (UID: \"4f248d51-18f9-4d20-b64b-d23765054818\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-frrmr" Feb 27 08:23:59 crc kubenswrapper[4612]: I0227 08:23:59.373371 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f248d51-18f9-4d20-b64b-d23765054818-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-frrmr\" (UID: \"4f248d51-18f9-4d20-b64b-d23765054818\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-frrmr" Feb 27 08:23:59 crc kubenswrapper[4612]: I0227 08:23:59.373883 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/4f248d51-18f9-4d20-b64b-d23765054818-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-frrmr\" (UID: \"4f248d51-18f9-4d20-b64b-d23765054818\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-frrmr" Feb 27 08:23:59 crc kubenswrapper[4612]: I0227 08:23:59.374159 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f248d51-18f9-4d20-b64b-d23765054818-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-frrmr\" (UID: \"4f248d51-18f9-4d20-b64b-d23765054818\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-frrmr" Feb 27 08:23:59 crc kubenswrapper[4612]: I0227 08:23:59.375128 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4f248d51-18f9-4d20-b64b-d23765054818-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-frrmr\" (UID: \"4f248d51-18f9-4d20-b64b-d23765054818\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-frrmr" Feb 27 08:23:59 crc kubenswrapper[4612]: I0227 08:23:59.375348 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vkgn2\" (UniqueName: \"kubernetes.io/projected/4f248d51-18f9-4d20-b64b-d23765054818-kube-api-access-vkgn2\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-frrmr\" (UID: \"4f248d51-18f9-4d20-b64b-d23765054818\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-frrmr" Feb 27 08:23:59 crc kubenswrapper[4612]: I0227 08:23:59.384606 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f248d51-18f9-4d20-b64b-d23765054818-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-frrmr\" (UID: \"4f248d51-18f9-4d20-b64b-d23765054818\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-frrmr" Feb 27 08:23:59 crc kubenswrapper[4612]: I0227 08:23:59.385621 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/4f248d51-18f9-4d20-b64b-d23765054818-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-frrmr\" (UID: \"4f248d51-18f9-4d20-b64b-d23765054818\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-frrmr" Feb 27 08:23:59 crc kubenswrapper[4612]: I0227 08:23:59.387832 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/4f248d51-18f9-4d20-b64b-d23765054818-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-frrmr\" (UID: \"4f248d51-18f9-4d20-b64b-d23765054818\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-frrmr" Feb 27 08:23:59 crc kubenswrapper[4612]: I0227 08:23:59.393508 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/4f248d51-18f9-4d20-b64b-d23765054818-ssh-key-openstack-edpm-ipam\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-frrmr\" (UID: \"4f248d51-18f9-4d20-b64b-d23765054818\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-frrmr" Feb 27 08:23:59 crc kubenswrapper[4612]: I0227 08:23:59.462635 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-frrmr" Feb 27 08:23:59 crc kubenswrapper[4612]: I0227 08:23:59.971588 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-frrmr"] Feb 27 08:24:00 crc kubenswrapper[4612]: I0227 08:24:00.021494 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-frrmr" event={"ID":"4f248d51-18f9-4d20-b64b-d23765054818","Type":"ContainerStarted","Data":"4505e758e1ed842110e0fcdc175123fc73c4472905d84fc421c95f7021d11b7b"} Feb 27 08:24:00 crc kubenswrapper[4612]: I0227 08:24:00.153725 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29536344-pdbxv"] Feb 27 08:24:00 crc kubenswrapper[4612]: I0227 08:24:00.155168 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536344-pdbxv" Feb 27 08:24:00 crc kubenswrapper[4612]: I0227 08:24:00.159049 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-5zwfx" Feb 27 08:24:00 crc kubenswrapper[4612]: I0227 08:24:00.159247 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 27 08:24:00 crc kubenswrapper[4612]: I0227 08:24:00.159548 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 27 08:24:00 crc kubenswrapper[4612]: I0227 08:24:00.167853 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536344-pdbxv"] Feb 27 08:24:00 crc kubenswrapper[4612]: I0227 08:24:00.277525 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z5bc4\" (UniqueName: \"kubernetes.io/projected/b8b16b31-ecd0-4fb2-b827-ec0ea2f0ef03-kube-api-access-z5bc4\") pod \"auto-csr-approver-29536344-pdbxv\" (UID: \"b8b16b31-ecd0-4fb2-b827-ec0ea2f0ef03\") " pod="openshift-infra/auto-csr-approver-29536344-pdbxv" Feb 27 08:24:00 crc kubenswrapper[4612]: I0227 08:24:00.379594 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z5bc4\" (UniqueName: \"kubernetes.io/projected/b8b16b31-ecd0-4fb2-b827-ec0ea2f0ef03-kube-api-access-z5bc4\") pod \"auto-csr-approver-29536344-pdbxv\" (UID: \"b8b16b31-ecd0-4fb2-b827-ec0ea2f0ef03\") " pod="openshift-infra/auto-csr-approver-29536344-pdbxv" Feb 27 08:24:00 crc kubenswrapper[4612]: I0227 08:24:00.416826 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z5bc4\" (UniqueName: \"kubernetes.io/projected/b8b16b31-ecd0-4fb2-b827-ec0ea2f0ef03-kube-api-access-z5bc4\") pod \"auto-csr-approver-29536344-pdbxv\" (UID: \"b8b16b31-ecd0-4fb2-b827-ec0ea2f0ef03\") " pod="openshift-infra/auto-csr-approver-29536344-pdbxv" Feb 27 08:24:00 crc kubenswrapper[4612]: I0227 08:24:00.480869 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536344-pdbxv" Feb 27 08:24:01 crc kubenswrapper[4612]: I0227 08:24:01.030393 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-frrmr" event={"ID":"4f248d51-18f9-4d20-b64b-d23765054818","Type":"ContainerStarted","Data":"c2c2977108385295e9a3f2e178531ab72a589938abf030a3f318143ac8da356a"} Feb 27 08:24:01 crc kubenswrapper[4612]: I0227 08:24:01.057099 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-frrmr" podStartSLOduration=1.376407601 podStartE2EDuration="2.057081056s" podCreationTimestamp="2026-02-27 08:23:59 +0000 UTC" firstStartedPulling="2026-02-27 08:23:59.980336764 +0000 UTC m=+2097.834266762" lastFinishedPulling="2026-02-27 08:24:00.661010189 +0000 UTC m=+2098.514940217" observedRunningTime="2026-02-27 08:24:01.048065787 +0000 UTC m=+2098.901995805" watchObservedRunningTime="2026-02-27 08:24:01.057081056 +0000 UTC m=+2098.911011054" Feb 27 08:24:01 crc kubenswrapper[4612]: I0227 08:24:01.145230 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536344-pdbxv"] Feb 27 08:24:02 crc kubenswrapper[4612]: I0227 08:24:02.042547 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536344-pdbxv" event={"ID":"b8b16b31-ecd0-4fb2-b827-ec0ea2f0ef03","Type":"ContainerStarted","Data":"9d8a92b4b16b6d0e7e6033167c705f46538c2fa60fd1ce7c0d819ce719eac18a"} Feb 27 08:24:03 crc kubenswrapper[4612]: I0227 08:24:03.050225 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536344-pdbxv" event={"ID":"b8b16b31-ecd0-4fb2-b827-ec0ea2f0ef03","Type":"ContainerStarted","Data":"0023d505698c1042b5a378107945cbb81087ed8de071109ea9bcf9910711f71e"} Feb 27 08:24:04 crc kubenswrapper[4612]: I0227 08:24:04.059618 4612 generic.go:334] "Generic (PLEG): container finished" podID="b8b16b31-ecd0-4fb2-b827-ec0ea2f0ef03" containerID="0023d505698c1042b5a378107945cbb81087ed8de071109ea9bcf9910711f71e" exitCode=0 Feb 27 08:24:04 crc kubenswrapper[4612]: I0227 08:24:04.059736 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536344-pdbxv" event={"ID":"b8b16b31-ecd0-4fb2-b827-ec0ea2f0ef03","Type":"ContainerDied","Data":"0023d505698c1042b5a378107945cbb81087ed8de071109ea9bcf9910711f71e"} Feb 27 08:24:05 crc kubenswrapper[4612]: I0227 08:24:05.437917 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536344-pdbxv" Feb 27 08:24:05 crc kubenswrapper[4612]: I0227 08:24:05.505115 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z5bc4\" (UniqueName: \"kubernetes.io/projected/b8b16b31-ecd0-4fb2-b827-ec0ea2f0ef03-kube-api-access-z5bc4\") pod \"b8b16b31-ecd0-4fb2-b827-ec0ea2f0ef03\" (UID: \"b8b16b31-ecd0-4fb2-b827-ec0ea2f0ef03\") " Feb 27 08:24:05 crc kubenswrapper[4612]: I0227 08:24:05.510411 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b8b16b31-ecd0-4fb2-b827-ec0ea2f0ef03-kube-api-access-z5bc4" (OuterVolumeSpecName: "kube-api-access-z5bc4") pod "b8b16b31-ecd0-4fb2-b827-ec0ea2f0ef03" (UID: "b8b16b31-ecd0-4fb2-b827-ec0ea2f0ef03"). InnerVolumeSpecName "kube-api-access-z5bc4". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 08:24:05 crc kubenswrapper[4612]: I0227 08:24:05.607308 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z5bc4\" (UniqueName: \"kubernetes.io/projected/b8b16b31-ecd0-4fb2-b827-ec0ea2f0ef03-kube-api-access-z5bc4\") on node \"crc\" DevicePath \"\"" Feb 27 08:24:05 crc kubenswrapper[4612]: I0227 08:24:05.959656 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29536338-smcsj"] Feb 27 08:24:05 crc kubenswrapper[4612]: I0227 08:24:05.969501 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29536338-smcsj"] Feb 27 08:24:06 crc kubenswrapper[4612]: I0227 08:24:06.078349 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536344-pdbxv" event={"ID":"b8b16b31-ecd0-4fb2-b827-ec0ea2f0ef03","Type":"ContainerDied","Data":"9d8a92b4b16b6d0e7e6033167c705f46538c2fa60fd1ce7c0d819ce719eac18a"} Feb 27 08:24:06 crc kubenswrapper[4612]: I0227 08:24:06.078401 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536344-pdbxv" Feb 27 08:24:06 crc kubenswrapper[4612]: I0227 08:24:06.078405 4612 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9d8a92b4b16b6d0e7e6033167c705f46538c2fa60fd1ce7c0d819ce719eac18a" Feb 27 08:24:06 crc kubenswrapper[4612]: I0227 08:24:06.863260 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="452831e0-7617-4e8b-bb21-5740a2129fcf" path="/var/lib/kubelet/pods/452831e0-7617-4e8b-bb21-5740a2129fcf/volumes" Feb 27 08:24:17 crc kubenswrapper[4612]: I0227 08:24:17.352398 4612 scope.go:117] "RemoveContainer" containerID="85cacf47d720e93ee6405b6fe2f8eccf305b7741069ec86301c754f80389434a" Feb 27 08:24:36 crc kubenswrapper[4612]: I0227 08:24:36.375462 4612 generic.go:334] "Generic (PLEG): container finished" podID="4f248d51-18f9-4d20-b64b-d23765054818" containerID="c2c2977108385295e9a3f2e178531ab72a589938abf030a3f318143ac8da356a" exitCode=0 Feb 27 08:24:36 crc kubenswrapper[4612]: I0227 08:24:36.375536 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-frrmr" event={"ID":"4f248d51-18f9-4d20-b64b-d23765054818","Type":"ContainerDied","Data":"c2c2977108385295e9a3f2e178531ab72a589938abf030a3f318143ac8da356a"} Feb 27 08:24:37 crc kubenswrapper[4612]: I0227 08:24:37.791890 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-frrmr" Feb 27 08:24:37 crc kubenswrapper[4612]: I0227 08:24:37.932488 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vkgn2\" (UniqueName: \"kubernetes.io/projected/4f248d51-18f9-4d20-b64b-d23765054818-kube-api-access-vkgn2\") pod \"4f248d51-18f9-4d20-b64b-d23765054818\" (UID: \"4f248d51-18f9-4d20-b64b-d23765054818\") " Feb 27 08:24:37 crc kubenswrapper[4612]: I0227 08:24:37.932551 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f248d51-18f9-4d20-b64b-d23765054818-ovn-combined-ca-bundle\") pod \"4f248d51-18f9-4d20-b64b-d23765054818\" (UID: \"4f248d51-18f9-4d20-b64b-d23765054818\") " Feb 27 08:24:37 crc kubenswrapper[4612]: I0227 08:24:37.932571 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/4f248d51-18f9-4d20-b64b-d23765054818-ssh-key-openstack-edpm-ipam\") pod \"4f248d51-18f9-4d20-b64b-d23765054818\" (UID: \"4f248d51-18f9-4d20-b64b-d23765054818\") " Feb 27 08:24:37 crc kubenswrapper[4612]: I0227 08:24:37.932634 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f248d51-18f9-4d20-b64b-d23765054818-nova-combined-ca-bundle\") pod \"4f248d51-18f9-4d20-b64b-d23765054818\" (UID: \"4f248d51-18f9-4d20-b64b-d23765054818\") " Feb 27 08:24:37 crc kubenswrapper[4612]: I0227 08:24:37.932683 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/4f248d51-18f9-4d20-b64b-d23765054818-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"4f248d51-18f9-4d20-b64b-d23765054818\" (UID: \"4f248d51-18f9-4d20-b64b-d23765054818\") " Feb 27 08:24:37 crc kubenswrapper[4612]: I0227 08:24:37.932745 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f248d51-18f9-4d20-b64b-d23765054818-libvirt-combined-ca-bundle\") pod \"4f248d51-18f9-4d20-b64b-d23765054818\" (UID: \"4f248d51-18f9-4d20-b64b-d23765054818\") " Feb 27 08:24:37 crc kubenswrapper[4612]: I0227 08:24:37.932761 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4f248d51-18f9-4d20-b64b-d23765054818-inventory\") pod \"4f248d51-18f9-4d20-b64b-d23765054818\" (UID: \"4f248d51-18f9-4d20-b64b-d23765054818\") " Feb 27 08:24:37 crc kubenswrapper[4612]: I0227 08:24:37.932780 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f248d51-18f9-4d20-b64b-d23765054818-bootstrap-combined-ca-bundle\") pod \"4f248d51-18f9-4d20-b64b-d23765054818\" (UID: \"4f248d51-18f9-4d20-b64b-d23765054818\") " Feb 27 08:24:37 crc kubenswrapper[4612]: I0227 08:24:37.932801 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/4f248d51-18f9-4d20-b64b-d23765054818-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"4f248d51-18f9-4d20-b64b-d23765054818\" (UID: \"4f248d51-18f9-4d20-b64b-d23765054818\") " Feb 27 08:24:37 crc kubenswrapper[4612]: I0227 08:24:37.932884 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f248d51-18f9-4d20-b64b-d23765054818-neutron-metadata-combined-ca-bundle\") pod \"4f248d51-18f9-4d20-b64b-d23765054818\" (UID: \"4f248d51-18f9-4d20-b64b-d23765054818\") " Feb 27 08:24:37 crc kubenswrapper[4612]: I0227 08:24:37.932908 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/4f248d51-18f9-4d20-b64b-d23765054818-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"4f248d51-18f9-4d20-b64b-d23765054818\" (UID: \"4f248d51-18f9-4d20-b64b-d23765054818\") " Feb 27 08:24:37 crc kubenswrapper[4612]: I0227 08:24:37.932924 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f248d51-18f9-4d20-b64b-d23765054818-repo-setup-combined-ca-bundle\") pod \"4f248d51-18f9-4d20-b64b-d23765054818\" (UID: \"4f248d51-18f9-4d20-b64b-d23765054818\") " Feb 27 08:24:37 crc kubenswrapper[4612]: I0227 08:24:37.932942 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f248d51-18f9-4d20-b64b-d23765054818-telemetry-combined-ca-bundle\") pod \"4f248d51-18f9-4d20-b64b-d23765054818\" (UID: \"4f248d51-18f9-4d20-b64b-d23765054818\") " Feb 27 08:24:37 crc kubenswrapper[4612]: I0227 08:24:37.933001 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/4f248d51-18f9-4d20-b64b-d23765054818-openstack-edpm-ipam-ovn-default-certs-0\") pod \"4f248d51-18f9-4d20-b64b-d23765054818\" (UID: \"4f248d51-18f9-4d20-b64b-d23765054818\") " Feb 27 08:24:37 crc kubenswrapper[4612]: I0227 08:24:37.944239 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4f248d51-18f9-4d20-b64b-d23765054818-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "4f248d51-18f9-4d20-b64b-d23765054818" (UID: "4f248d51-18f9-4d20-b64b-d23765054818"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:24:37 crc kubenswrapper[4612]: I0227 08:24:37.946022 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4f248d51-18f9-4d20-b64b-d23765054818-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "4f248d51-18f9-4d20-b64b-d23765054818" (UID: "4f248d51-18f9-4d20-b64b-d23765054818"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:24:37 crc kubenswrapper[4612]: I0227 08:24:37.946205 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4f248d51-18f9-4d20-b64b-d23765054818-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "4f248d51-18f9-4d20-b64b-d23765054818" (UID: "4f248d51-18f9-4d20-b64b-d23765054818"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:24:37 crc kubenswrapper[4612]: I0227 08:24:37.947090 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4f248d51-18f9-4d20-b64b-d23765054818-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "4f248d51-18f9-4d20-b64b-d23765054818" (UID: "4f248d51-18f9-4d20-b64b-d23765054818"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:24:37 crc kubenswrapper[4612]: I0227 08:24:37.947982 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4f248d51-18f9-4d20-b64b-d23765054818-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "4f248d51-18f9-4d20-b64b-d23765054818" (UID: "4f248d51-18f9-4d20-b64b-d23765054818"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:24:37 crc kubenswrapper[4612]: I0227 08:24:37.948786 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4f248d51-18f9-4d20-b64b-d23765054818-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "4f248d51-18f9-4d20-b64b-d23765054818" (UID: "4f248d51-18f9-4d20-b64b-d23765054818"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:24:37 crc kubenswrapper[4612]: I0227 08:24:37.952102 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4f248d51-18f9-4d20-b64b-d23765054818-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "4f248d51-18f9-4d20-b64b-d23765054818" (UID: "4f248d51-18f9-4d20-b64b-d23765054818"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:24:37 crc kubenswrapper[4612]: I0227 08:24:37.968530 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4f248d51-18f9-4d20-b64b-d23765054818-openstack-edpm-ipam-neutron-metadata-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-neutron-metadata-default-certs-0") pod "4f248d51-18f9-4d20-b64b-d23765054818" (UID: "4f248d51-18f9-4d20-b64b-d23765054818"). InnerVolumeSpecName "openstack-edpm-ipam-neutron-metadata-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 08:24:37 crc kubenswrapper[4612]: I0227 08:24:37.968836 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4f248d51-18f9-4d20-b64b-d23765054818-openstack-edpm-ipam-telemetry-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-telemetry-default-certs-0") pod "4f248d51-18f9-4d20-b64b-d23765054818" (UID: "4f248d51-18f9-4d20-b64b-d23765054818"). InnerVolumeSpecName "openstack-edpm-ipam-telemetry-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 08:24:37 crc kubenswrapper[4612]: I0227 08:24:37.968641 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4f248d51-18f9-4d20-b64b-d23765054818-kube-api-access-vkgn2" (OuterVolumeSpecName: "kube-api-access-vkgn2") pod "4f248d51-18f9-4d20-b64b-d23765054818" (UID: "4f248d51-18f9-4d20-b64b-d23765054818"). InnerVolumeSpecName "kube-api-access-vkgn2". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 08:24:37 crc kubenswrapper[4612]: I0227 08:24:37.968683 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4f248d51-18f9-4d20-b64b-d23765054818-openstack-edpm-ipam-libvirt-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-libvirt-default-certs-0") pod "4f248d51-18f9-4d20-b64b-d23765054818" (UID: "4f248d51-18f9-4d20-b64b-d23765054818"). InnerVolumeSpecName "openstack-edpm-ipam-libvirt-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 08:24:37 crc kubenswrapper[4612]: I0227 08:24:37.968759 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4f248d51-18f9-4d20-b64b-d23765054818-openstack-edpm-ipam-ovn-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-ovn-default-certs-0") pod "4f248d51-18f9-4d20-b64b-d23765054818" (UID: "4f248d51-18f9-4d20-b64b-d23765054818"). InnerVolumeSpecName "openstack-edpm-ipam-ovn-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 08:24:37 crc kubenswrapper[4612]: I0227 08:24:37.978735 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4f248d51-18f9-4d20-b64b-d23765054818-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "4f248d51-18f9-4d20-b64b-d23765054818" (UID: "4f248d51-18f9-4d20-b64b-d23765054818"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:24:37 crc kubenswrapper[4612]: I0227 08:24:37.980969 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4f248d51-18f9-4d20-b64b-d23765054818-inventory" (OuterVolumeSpecName: "inventory") pod "4f248d51-18f9-4d20-b64b-d23765054818" (UID: "4f248d51-18f9-4d20-b64b-d23765054818"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:24:38 crc kubenswrapper[4612]: I0227 08:24:38.035176 4612 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/4f248d51-18f9-4d20-b64b-d23765054818-openstack-edpm-ipam-ovn-default-certs-0\") on node \"crc\" DevicePath \"\"" Feb 27 08:24:38 crc kubenswrapper[4612]: I0227 08:24:38.035205 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vkgn2\" (UniqueName: \"kubernetes.io/projected/4f248d51-18f9-4d20-b64b-d23765054818-kube-api-access-vkgn2\") on node \"crc\" DevicePath \"\"" Feb 27 08:24:38 crc kubenswrapper[4612]: I0227 08:24:38.035216 4612 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f248d51-18f9-4d20-b64b-d23765054818-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 08:24:38 crc kubenswrapper[4612]: I0227 08:24:38.035225 4612 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/4f248d51-18f9-4d20-b64b-d23765054818-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 27 08:24:38 crc kubenswrapper[4612]: I0227 08:24:38.035234 4612 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f248d51-18f9-4d20-b64b-d23765054818-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 08:24:38 crc kubenswrapper[4612]: I0227 08:24:38.035242 4612 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/4f248d51-18f9-4d20-b64b-d23765054818-openstack-edpm-ipam-telemetry-default-certs-0\") on node \"crc\" DevicePath \"\"" Feb 27 08:24:38 crc kubenswrapper[4612]: I0227 08:24:38.035251 4612 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f248d51-18f9-4d20-b64b-d23765054818-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 08:24:38 crc kubenswrapper[4612]: I0227 08:24:38.035260 4612 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4f248d51-18f9-4d20-b64b-d23765054818-inventory\") on node \"crc\" DevicePath \"\"" Feb 27 08:24:38 crc kubenswrapper[4612]: I0227 08:24:38.035268 4612 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f248d51-18f9-4d20-b64b-d23765054818-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 08:24:38 crc kubenswrapper[4612]: I0227 08:24:38.035276 4612 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/4f248d51-18f9-4d20-b64b-d23765054818-openstack-edpm-ipam-libvirt-default-certs-0\") on node \"crc\" DevicePath \"\"" Feb 27 08:24:38 crc kubenswrapper[4612]: I0227 08:24:38.035285 4612 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f248d51-18f9-4d20-b64b-d23765054818-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 08:24:38 crc kubenswrapper[4612]: I0227 08:24:38.035294 4612 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/4f248d51-18f9-4d20-b64b-d23765054818-openstack-edpm-ipam-neutron-metadata-default-certs-0\") on node \"crc\" DevicePath \"\"" Feb 27 08:24:38 crc kubenswrapper[4612]: I0227 08:24:38.035304 4612 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f248d51-18f9-4d20-b64b-d23765054818-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 08:24:38 crc kubenswrapper[4612]: I0227 08:24:38.035314 4612 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f248d51-18f9-4d20-b64b-d23765054818-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 08:24:38 crc kubenswrapper[4612]: I0227 08:24:38.397965 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-frrmr" event={"ID":"4f248d51-18f9-4d20-b64b-d23765054818","Type":"ContainerDied","Data":"4505e758e1ed842110e0fcdc175123fc73c4472905d84fc421c95f7021d11b7b"} Feb 27 08:24:38 crc kubenswrapper[4612]: I0227 08:24:38.398042 4612 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4505e758e1ed842110e0fcdc175123fc73c4472905d84fc421c95f7021d11b7b" Feb 27 08:24:38 crc kubenswrapper[4612]: I0227 08:24:38.398077 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-frrmr" Feb 27 08:24:38 crc kubenswrapper[4612]: I0227 08:24:38.548783 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-xhcrn"] Feb 27 08:24:38 crc kubenswrapper[4612]: E0227 08:24:38.549546 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4f248d51-18f9-4d20-b64b-d23765054818" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Feb 27 08:24:38 crc kubenswrapper[4612]: I0227 08:24:38.549576 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="4f248d51-18f9-4d20-b64b-d23765054818" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Feb 27 08:24:38 crc kubenswrapper[4612]: E0227 08:24:38.549610 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b8b16b31-ecd0-4fb2-b827-ec0ea2f0ef03" containerName="oc" Feb 27 08:24:38 crc kubenswrapper[4612]: I0227 08:24:38.549620 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="b8b16b31-ecd0-4fb2-b827-ec0ea2f0ef03" containerName="oc" Feb 27 08:24:38 crc kubenswrapper[4612]: I0227 08:24:38.549833 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="4f248d51-18f9-4d20-b64b-d23765054818" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Feb 27 08:24:38 crc kubenswrapper[4612]: I0227 08:24:38.549854 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="b8b16b31-ecd0-4fb2-b827-ec0ea2f0ef03" containerName="oc" Feb 27 08:24:38 crc kubenswrapper[4612]: I0227 08:24:38.550481 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-xhcrn" Feb 27 08:24:38 crc kubenswrapper[4612]: I0227 08:24:38.554225 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-config" Feb 27 08:24:38 crc kubenswrapper[4612]: I0227 08:24:38.554547 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 27 08:24:38 crc kubenswrapper[4612]: I0227 08:24:38.556623 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 27 08:24:38 crc kubenswrapper[4612]: I0227 08:24:38.557168 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 27 08:24:38 crc kubenswrapper[4612]: I0227 08:24:38.565835 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-2pwrb" Feb 27 08:24:38 crc kubenswrapper[4612]: I0227 08:24:38.566712 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-xhcrn"] Feb 27 08:24:38 crc kubenswrapper[4612]: I0227 08:24:38.650154 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sgp5v\" (UniqueName: \"kubernetes.io/projected/dfdf6a2b-0ca5-4ea9-b87c-fd7a61f4981c-kube-api-access-sgp5v\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-xhcrn\" (UID: \"dfdf6a2b-0ca5-4ea9-b87c-fd7a61f4981c\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-xhcrn" Feb 27 08:24:38 crc kubenswrapper[4612]: I0227 08:24:38.650214 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/dfdf6a2b-0ca5-4ea9-b87c-fd7a61f4981c-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-xhcrn\" (UID: \"dfdf6a2b-0ca5-4ea9-b87c-fd7a61f4981c\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-xhcrn" Feb 27 08:24:38 crc kubenswrapper[4612]: I0227 08:24:38.650253 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dfdf6a2b-0ca5-4ea9-b87c-fd7a61f4981c-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-xhcrn\" (UID: \"dfdf6a2b-0ca5-4ea9-b87c-fd7a61f4981c\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-xhcrn" Feb 27 08:24:38 crc kubenswrapper[4612]: I0227 08:24:38.650366 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/dfdf6a2b-0ca5-4ea9-b87c-fd7a61f4981c-ssh-key-openstack-edpm-ipam\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-xhcrn\" (UID: \"dfdf6a2b-0ca5-4ea9-b87c-fd7a61f4981c\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-xhcrn" Feb 27 08:24:38 crc kubenswrapper[4612]: I0227 08:24:38.650408 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/dfdf6a2b-0ca5-4ea9-b87c-fd7a61f4981c-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-xhcrn\" (UID: \"dfdf6a2b-0ca5-4ea9-b87c-fd7a61f4981c\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-xhcrn" Feb 27 08:24:38 crc kubenswrapper[4612]: I0227 08:24:38.752623 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/dfdf6a2b-0ca5-4ea9-b87c-fd7a61f4981c-ssh-key-openstack-edpm-ipam\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-xhcrn\" (UID: \"dfdf6a2b-0ca5-4ea9-b87c-fd7a61f4981c\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-xhcrn" Feb 27 08:24:38 crc kubenswrapper[4612]: I0227 08:24:38.752762 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/dfdf6a2b-0ca5-4ea9-b87c-fd7a61f4981c-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-xhcrn\" (UID: \"dfdf6a2b-0ca5-4ea9-b87c-fd7a61f4981c\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-xhcrn" Feb 27 08:24:38 crc kubenswrapper[4612]: I0227 08:24:38.752967 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sgp5v\" (UniqueName: \"kubernetes.io/projected/dfdf6a2b-0ca5-4ea9-b87c-fd7a61f4981c-kube-api-access-sgp5v\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-xhcrn\" (UID: \"dfdf6a2b-0ca5-4ea9-b87c-fd7a61f4981c\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-xhcrn" Feb 27 08:24:38 crc kubenswrapper[4612]: I0227 08:24:38.753018 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/dfdf6a2b-0ca5-4ea9-b87c-fd7a61f4981c-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-xhcrn\" (UID: \"dfdf6a2b-0ca5-4ea9-b87c-fd7a61f4981c\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-xhcrn" Feb 27 08:24:38 crc kubenswrapper[4612]: I0227 08:24:38.753082 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dfdf6a2b-0ca5-4ea9-b87c-fd7a61f4981c-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-xhcrn\" (UID: \"dfdf6a2b-0ca5-4ea9-b87c-fd7a61f4981c\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-xhcrn" Feb 27 08:24:38 crc kubenswrapper[4612]: I0227 08:24:38.754025 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/dfdf6a2b-0ca5-4ea9-b87c-fd7a61f4981c-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-xhcrn\" (UID: \"dfdf6a2b-0ca5-4ea9-b87c-fd7a61f4981c\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-xhcrn" Feb 27 08:24:38 crc kubenswrapper[4612]: I0227 08:24:38.758068 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/dfdf6a2b-0ca5-4ea9-b87c-fd7a61f4981c-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-xhcrn\" (UID: \"dfdf6a2b-0ca5-4ea9-b87c-fd7a61f4981c\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-xhcrn" Feb 27 08:24:38 crc kubenswrapper[4612]: I0227 08:24:38.758305 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/dfdf6a2b-0ca5-4ea9-b87c-fd7a61f4981c-ssh-key-openstack-edpm-ipam\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-xhcrn\" (UID: \"dfdf6a2b-0ca5-4ea9-b87c-fd7a61f4981c\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-xhcrn" Feb 27 08:24:38 crc kubenswrapper[4612]: I0227 08:24:38.759896 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dfdf6a2b-0ca5-4ea9-b87c-fd7a61f4981c-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-xhcrn\" (UID: \"dfdf6a2b-0ca5-4ea9-b87c-fd7a61f4981c\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-xhcrn" Feb 27 08:24:38 crc kubenswrapper[4612]: I0227 08:24:38.777114 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sgp5v\" (UniqueName: \"kubernetes.io/projected/dfdf6a2b-0ca5-4ea9-b87c-fd7a61f4981c-kube-api-access-sgp5v\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-xhcrn\" (UID: \"dfdf6a2b-0ca5-4ea9-b87c-fd7a61f4981c\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-xhcrn" Feb 27 08:24:38 crc kubenswrapper[4612]: I0227 08:24:38.916813 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-xhcrn" Feb 27 08:24:39 crc kubenswrapper[4612]: I0227 08:24:39.482100 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-xhcrn"] Feb 27 08:24:40 crc kubenswrapper[4612]: I0227 08:24:40.429931 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-xhcrn" event={"ID":"dfdf6a2b-0ca5-4ea9-b87c-fd7a61f4981c","Type":"ContainerStarted","Data":"27ec6f685d7b9a699df1b6325897c2133b0e8eadd1579c1405cf884d40a31d91"} Feb 27 08:24:40 crc kubenswrapper[4612]: I0227 08:24:40.430240 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-xhcrn" event={"ID":"dfdf6a2b-0ca5-4ea9-b87c-fd7a61f4981c","Type":"ContainerStarted","Data":"889b74e01e806ba955d88047bf3550daf9409f1ef1ec9488ce6942705419a8fd"} Feb 27 08:24:40 crc kubenswrapper[4612]: I0227 08:24:40.454061 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-xhcrn" podStartSLOduration=2.275090188 podStartE2EDuration="2.454031014s" podCreationTimestamp="2026-02-27 08:24:38 +0000 UTC" firstStartedPulling="2026-02-27 08:24:39.494911928 +0000 UTC m=+2137.348841926" lastFinishedPulling="2026-02-27 08:24:39.673852754 +0000 UTC m=+2137.527782752" observedRunningTime="2026-02-27 08:24:40.448279089 +0000 UTC m=+2138.302209097" watchObservedRunningTime="2026-02-27 08:24:40.454031014 +0000 UTC m=+2138.307961042" Feb 27 08:24:42 crc kubenswrapper[4612]: I0227 08:24:42.704597 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-jdqbp"] Feb 27 08:24:42 crc kubenswrapper[4612]: I0227 08:24:42.712474 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jdqbp" Feb 27 08:24:42 crc kubenswrapper[4612]: I0227 08:24:42.725736 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-jdqbp"] Feb 27 08:24:42 crc kubenswrapper[4612]: I0227 08:24:42.842411 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tqq65\" (UniqueName: \"kubernetes.io/projected/cc8baf5d-91f1-4fed-836d-d3eb2acd6854-kube-api-access-tqq65\") pod \"redhat-operators-jdqbp\" (UID: \"cc8baf5d-91f1-4fed-836d-d3eb2acd6854\") " pod="openshift-marketplace/redhat-operators-jdqbp" Feb 27 08:24:42 crc kubenswrapper[4612]: I0227 08:24:42.842454 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cc8baf5d-91f1-4fed-836d-d3eb2acd6854-utilities\") pod \"redhat-operators-jdqbp\" (UID: \"cc8baf5d-91f1-4fed-836d-d3eb2acd6854\") " pod="openshift-marketplace/redhat-operators-jdqbp" Feb 27 08:24:42 crc kubenswrapper[4612]: I0227 08:24:42.842640 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cc8baf5d-91f1-4fed-836d-d3eb2acd6854-catalog-content\") pod \"redhat-operators-jdqbp\" (UID: \"cc8baf5d-91f1-4fed-836d-d3eb2acd6854\") " pod="openshift-marketplace/redhat-operators-jdqbp" Feb 27 08:24:42 crc kubenswrapper[4612]: I0227 08:24:42.944525 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tqq65\" (UniqueName: \"kubernetes.io/projected/cc8baf5d-91f1-4fed-836d-d3eb2acd6854-kube-api-access-tqq65\") pod \"redhat-operators-jdqbp\" (UID: \"cc8baf5d-91f1-4fed-836d-d3eb2acd6854\") " pod="openshift-marketplace/redhat-operators-jdqbp" Feb 27 08:24:42 crc kubenswrapper[4612]: I0227 08:24:42.944582 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cc8baf5d-91f1-4fed-836d-d3eb2acd6854-utilities\") pod \"redhat-operators-jdqbp\" (UID: \"cc8baf5d-91f1-4fed-836d-d3eb2acd6854\") " pod="openshift-marketplace/redhat-operators-jdqbp" Feb 27 08:24:42 crc kubenswrapper[4612]: I0227 08:24:42.944632 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cc8baf5d-91f1-4fed-836d-d3eb2acd6854-catalog-content\") pod \"redhat-operators-jdqbp\" (UID: \"cc8baf5d-91f1-4fed-836d-d3eb2acd6854\") " pod="openshift-marketplace/redhat-operators-jdqbp" Feb 27 08:24:42 crc kubenswrapper[4612]: I0227 08:24:42.945209 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cc8baf5d-91f1-4fed-836d-d3eb2acd6854-utilities\") pod \"redhat-operators-jdqbp\" (UID: \"cc8baf5d-91f1-4fed-836d-d3eb2acd6854\") " pod="openshift-marketplace/redhat-operators-jdqbp" Feb 27 08:24:42 crc kubenswrapper[4612]: I0227 08:24:42.945296 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cc8baf5d-91f1-4fed-836d-d3eb2acd6854-catalog-content\") pod \"redhat-operators-jdqbp\" (UID: \"cc8baf5d-91f1-4fed-836d-d3eb2acd6854\") " pod="openshift-marketplace/redhat-operators-jdqbp" Feb 27 08:24:42 crc kubenswrapper[4612]: I0227 08:24:42.974808 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tqq65\" (UniqueName: \"kubernetes.io/projected/cc8baf5d-91f1-4fed-836d-d3eb2acd6854-kube-api-access-tqq65\") pod \"redhat-operators-jdqbp\" (UID: \"cc8baf5d-91f1-4fed-836d-d3eb2acd6854\") " pod="openshift-marketplace/redhat-operators-jdqbp" Feb 27 08:24:43 crc kubenswrapper[4612]: I0227 08:24:43.044302 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jdqbp" Feb 27 08:24:43 crc kubenswrapper[4612]: W0227 08:24:43.576318 4612 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcc8baf5d_91f1_4fed_836d_d3eb2acd6854.slice/crio-4b3b52a1c7ffb57a9197507b9bdeca573bb7cd15af01cb679425b90f656eeb1a WatchSource:0}: Error finding container 4b3b52a1c7ffb57a9197507b9bdeca573bb7cd15af01cb679425b90f656eeb1a: Status 404 returned error can't find the container with id 4b3b52a1c7ffb57a9197507b9bdeca573bb7cd15af01cb679425b90f656eeb1a Feb 27 08:24:43 crc kubenswrapper[4612]: I0227 08:24:43.591255 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-jdqbp"] Feb 27 08:24:44 crc kubenswrapper[4612]: I0227 08:24:44.466409 4612 generic.go:334] "Generic (PLEG): container finished" podID="cc8baf5d-91f1-4fed-836d-d3eb2acd6854" containerID="220c30df5c00042ad8b9eefe0a39591caf49685b318eb386322ed07d2bdbcb94" exitCode=0 Feb 27 08:24:44 crc kubenswrapper[4612]: I0227 08:24:44.466564 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jdqbp" event={"ID":"cc8baf5d-91f1-4fed-836d-d3eb2acd6854","Type":"ContainerDied","Data":"220c30df5c00042ad8b9eefe0a39591caf49685b318eb386322ed07d2bdbcb94"} Feb 27 08:24:44 crc kubenswrapper[4612]: I0227 08:24:44.466729 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jdqbp" event={"ID":"cc8baf5d-91f1-4fed-836d-d3eb2acd6854","Type":"ContainerStarted","Data":"4b3b52a1c7ffb57a9197507b9bdeca573bb7cd15af01cb679425b90f656eeb1a"} Feb 27 08:24:46 crc kubenswrapper[4612]: I0227 08:24:46.492663 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jdqbp" event={"ID":"cc8baf5d-91f1-4fed-836d-d3eb2acd6854","Type":"ContainerStarted","Data":"4f7fb07b1662532350bed1c58e2175b77fc29ef44a333d70dc9ccc77074f01ee"} Feb 27 08:24:51 crc kubenswrapper[4612]: I0227 08:24:51.537425 4612 generic.go:334] "Generic (PLEG): container finished" podID="cc8baf5d-91f1-4fed-836d-d3eb2acd6854" containerID="4f7fb07b1662532350bed1c58e2175b77fc29ef44a333d70dc9ccc77074f01ee" exitCode=0 Feb 27 08:24:51 crc kubenswrapper[4612]: I0227 08:24:51.537518 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jdqbp" event={"ID":"cc8baf5d-91f1-4fed-836d-d3eb2acd6854","Type":"ContainerDied","Data":"4f7fb07b1662532350bed1c58e2175b77fc29ef44a333d70dc9ccc77074f01ee"} Feb 27 08:24:52 crc kubenswrapper[4612]: I0227 08:24:52.547093 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jdqbp" event={"ID":"cc8baf5d-91f1-4fed-836d-d3eb2acd6854","Type":"ContainerStarted","Data":"0630a0b7713cbeb711b50f8e42f6c495aca8cc9e1a942a420753cb223fca9ea2"} Feb 27 08:24:52 crc kubenswrapper[4612]: I0227 08:24:52.573567 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-jdqbp" podStartSLOduration=3.138659439 podStartE2EDuration="10.573546214s" podCreationTimestamp="2026-02-27 08:24:42 +0000 UTC" firstStartedPulling="2026-02-27 08:24:44.46825469 +0000 UTC m=+2142.322184688" lastFinishedPulling="2026-02-27 08:24:51.903141475 +0000 UTC m=+2149.757071463" observedRunningTime="2026-02-27 08:24:52.565534964 +0000 UTC m=+2150.419464962" watchObservedRunningTime="2026-02-27 08:24:52.573546214 +0000 UTC m=+2150.427476212" Feb 27 08:24:53 crc kubenswrapper[4612]: I0227 08:24:53.044491 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-jdqbp" Feb 27 08:24:53 crc kubenswrapper[4612]: I0227 08:24:53.045023 4612 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-jdqbp" Feb 27 08:24:54 crc kubenswrapper[4612]: I0227 08:24:54.099226 4612 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-jdqbp" podUID="cc8baf5d-91f1-4fed-836d-d3eb2acd6854" containerName="registry-server" probeResult="failure" output=< Feb 27 08:24:54 crc kubenswrapper[4612]: timeout: failed to connect service ":50051" within 1s Feb 27 08:24:54 crc kubenswrapper[4612]: > Feb 27 08:24:54 crc kubenswrapper[4612]: I0227 08:24:54.316032 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-2cvlb"] Feb 27 08:24:54 crc kubenswrapper[4612]: I0227 08:24:54.318540 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2cvlb" Feb 27 08:24:54 crc kubenswrapper[4612]: I0227 08:24:54.336608 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-2cvlb"] Feb 27 08:24:54 crc kubenswrapper[4612]: I0227 08:24:54.418562 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1ccadfc0-4f98-4981-b1cb-db16b255e529-utilities\") pod \"redhat-marketplace-2cvlb\" (UID: \"1ccadfc0-4f98-4981-b1cb-db16b255e529\") " pod="openshift-marketplace/redhat-marketplace-2cvlb" Feb 27 08:24:54 crc kubenswrapper[4612]: I0227 08:24:54.418994 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1ccadfc0-4f98-4981-b1cb-db16b255e529-catalog-content\") pod \"redhat-marketplace-2cvlb\" (UID: \"1ccadfc0-4f98-4981-b1cb-db16b255e529\") " pod="openshift-marketplace/redhat-marketplace-2cvlb" Feb 27 08:24:54 crc kubenswrapper[4612]: I0227 08:24:54.419021 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zz9kl\" (UniqueName: \"kubernetes.io/projected/1ccadfc0-4f98-4981-b1cb-db16b255e529-kube-api-access-zz9kl\") pod \"redhat-marketplace-2cvlb\" (UID: \"1ccadfc0-4f98-4981-b1cb-db16b255e529\") " pod="openshift-marketplace/redhat-marketplace-2cvlb" Feb 27 08:24:54 crc kubenswrapper[4612]: I0227 08:24:54.520542 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1ccadfc0-4f98-4981-b1cb-db16b255e529-utilities\") pod \"redhat-marketplace-2cvlb\" (UID: \"1ccadfc0-4f98-4981-b1cb-db16b255e529\") " pod="openshift-marketplace/redhat-marketplace-2cvlb" Feb 27 08:24:54 crc kubenswrapper[4612]: I0227 08:24:54.520625 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1ccadfc0-4f98-4981-b1cb-db16b255e529-catalog-content\") pod \"redhat-marketplace-2cvlb\" (UID: \"1ccadfc0-4f98-4981-b1cb-db16b255e529\") " pod="openshift-marketplace/redhat-marketplace-2cvlb" Feb 27 08:24:54 crc kubenswrapper[4612]: I0227 08:24:54.520646 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zz9kl\" (UniqueName: \"kubernetes.io/projected/1ccadfc0-4f98-4981-b1cb-db16b255e529-kube-api-access-zz9kl\") pod \"redhat-marketplace-2cvlb\" (UID: \"1ccadfc0-4f98-4981-b1cb-db16b255e529\") " pod="openshift-marketplace/redhat-marketplace-2cvlb" Feb 27 08:24:54 crc kubenswrapper[4612]: I0227 08:24:54.521019 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1ccadfc0-4f98-4981-b1cb-db16b255e529-utilities\") pod \"redhat-marketplace-2cvlb\" (UID: \"1ccadfc0-4f98-4981-b1cb-db16b255e529\") " pod="openshift-marketplace/redhat-marketplace-2cvlb" Feb 27 08:24:54 crc kubenswrapper[4612]: I0227 08:24:54.521147 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1ccadfc0-4f98-4981-b1cb-db16b255e529-catalog-content\") pod \"redhat-marketplace-2cvlb\" (UID: \"1ccadfc0-4f98-4981-b1cb-db16b255e529\") " pod="openshift-marketplace/redhat-marketplace-2cvlb" Feb 27 08:24:54 crc kubenswrapper[4612]: I0227 08:24:54.552919 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zz9kl\" (UniqueName: \"kubernetes.io/projected/1ccadfc0-4f98-4981-b1cb-db16b255e529-kube-api-access-zz9kl\") pod \"redhat-marketplace-2cvlb\" (UID: \"1ccadfc0-4f98-4981-b1cb-db16b255e529\") " pod="openshift-marketplace/redhat-marketplace-2cvlb" Feb 27 08:24:54 crc kubenswrapper[4612]: I0227 08:24:54.637442 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2cvlb" Feb 27 08:24:55 crc kubenswrapper[4612]: I0227 08:24:55.236641 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-2cvlb"] Feb 27 08:24:55 crc kubenswrapper[4612]: I0227 08:24:55.575126 4612 generic.go:334] "Generic (PLEG): container finished" podID="1ccadfc0-4f98-4981-b1cb-db16b255e529" containerID="e6ba5d3f85cdc42928efac6eaf4b253947c321887b7063b40c56c897464c3f04" exitCode=0 Feb 27 08:24:55 crc kubenswrapper[4612]: I0227 08:24:55.575303 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2cvlb" event={"ID":"1ccadfc0-4f98-4981-b1cb-db16b255e529","Type":"ContainerDied","Data":"e6ba5d3f85cdc42928efac6eaf4b253947c321887b7063b40c56c897464c3f04"} Feb 27 08:24:55 crc kubenswrapper[4612]: I0227 08:24:55.575471 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2cvlb" event={"ID":"1ccadfc0-4f98-4981-b1cb-db16b255e529","Type":"ContainerStarted","Data":"0532d9008e239bc55b242336dc501d517d911c830216662b51268448e0eda93c"} Feb 27 08:24:56 crc kubenswrapper[4612]: I0227 08:24:56.585647 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2cvlb" event={"ID":"1ccadfc0-4f98-4981-b1cb-db16b255e529","Type":"ContainerStarted","Data":"62951ea4bedbcbfe9dfb70febd1ff488d8be7f6bea8606518d737930e01c18dd"} Feb 27 08:24:57 crc kubenswrapper[4612]: I0227 08:24:57.594378 4612 generic.go:334] "Generic (PLEG): container finished" podID="1ccadfc0-4f98-4981-b1cb-db16b255e529" containerID="62951ea4bedbcbfe9dfb70febd1ff488d8be7f6bea8606518d737930e01c18dd" exitCode=0 Feb 27 08:24:57 crc kubenswrapper[4612]: I0227 08:24:57.594476 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2cvlb" event={"ID":"1ccadfc0-4f98-4981-b1cb-db16b255e529","Type":"ContainerDied","Data":"62951ea4bedbcbfe9dfb70febd1ff488d8be7f6bea8606518d737930e01c18dd"} Feb 27 08:24:58 crc kubenswrapper[4612]: I0227 08:24:58.605619 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2cvlb" event={"ID":"1ccadfc0-4f98-4981-b1cb-db16b255e529","Type":"ContainerStarted","Data":"0be7c0828c7d8d39cf02bec49f6ad2bc3139c7d473084fbac40872ef067ec239"} Feb 27 08:24:58 crc kubenswrapper[4612]: I0227 08:24:58.630883 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-2cvlb" podStartSLOduration=2.233629956 podStartE2EDuration="4.630858385s" podCreationTimestamp="2026-02-27 08:24:54 +0000 UTC" firstStartedPulling="2026-02-27 08:24:55.576828146 +0000 UTC m=+2153.430758144" lastFinishedPulling="2026-02-27 08:24:57.974056575 +0000 UTC m=+2155.827986573" observedRunningTime="2026-02-27 08:24:58.625421789 +0000 UTC m=+2156.479351787" watchObservedRunningTime="2026-02-27 08:24:58.630858385 +0000 UTC m=+2156.484788383" Feb 27 08:25:04 crc kubenswrapper[4612]: I0227 08:25:04.106401 4612 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-jdqbp" podUID="cc8baf5d-91f1-4fed-836d-d3eb2acd6854" containerName="registry-server" probeResult="failure" output=< Feb 27 08:25:04 crc kubenswrapper[4612]: timeout: failed to connect service ":50051" within 1s Feb 27 08:25:04 crc kubenswrapper[4612]: > Feb 27 08:25:04 crc kubenswrapper[4612]: I0227 08:25:04.637965 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-2cvlb" Feb 27 08:25:04 crc kubenswrapper[4612]: I0227 08:25:04.638028 4612 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-2cvlb" Feb 27 08:25:04 crc kubenswrapper[4612]: I0227 08:25:04.719136 4612 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-2cvlb" Feb 27 08:25:04 crc kubenswrapper[4612]: I0227 08:25:04.813080 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-2cvlb" Feb 27 08:25:04 crc kubenswrapper[4612]: I0227 08:25:04.964730 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-2cvlb"] Feb 27 08:25:06 crc kubenswrapper[4612]: I0227 08:25:06.667270 4612 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-2cvlb" podUID="1ccadfc0-4f98-4981-b1cb-db16b255e529" containerName="registry-server" containerID="cri-o://0be7c0828c7d8d39cf02bec49f6ad2bc3139c7d473084fbac40872ef067ec239" gracePeriod=2 Feb 27 08:25:07 crc kubenswrapper[4612]: I0227 08:25:07.133086 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2cvlb" Feb 27 08:25:07 crc kubenswrapper[4612]: I0227 08:25:07.254916 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1ccadfc0-4f98-4981-b1cb-db16b255e529-utilities\") pod \"1ccadfc0-4f98-4981-b1cb-db16b255e529\" (UID: \"1ccadfc0-4f98-4981-b1cb-db16b255e529\") " Feb 27 08:25:07 crc kubenswrapper[4612]: I0227 08:25:07.254972 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zz9kl\" (UniqueName: \"kubernetes.io/projected/1ccadfc0-4f98-4981-b1cb-db16b255e529-kube-api-access-zz9kl\") pod \"1ccadfc0-4f98-4981-b1cb-db16b255e529\" (UID: \"1ccadfc0-4f98-4981-b1cb-db16b255e529\") " Feb 27 08:25:07 crc kubenswrapper[4612]: I0227 08:25:07.255109 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1ccadfc0-4f98-4981-b1cb-db16b255e529-catalog-content\") pod \"1ccadfc0-4f98-4981-b1cb-db16b255e529\" (UID: \"1ccadfc0-4f98-4981-b1cb-db16b255e529\") " Feb 27 08:25:07 crc kubenswrapper[4612]: I0227 08:25:07.256010 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1ccadfc0-4f98-4981-b1cb-db16b255e529-utilities" (OuterVolumeSpecName: "utilities") pod "1ccadfc0-4f98-4981-b1cb-db16b255e529" (UID: "1ccadfc0-4f98-4981-b1cb-db16b255e529"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 08:25:07 crc kubenswrapper[4612]: I0227 08:25:07.269881 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1ccadfc0-4f98-4981-b1cb-db16b255e529-kube-api-access-zz9kl" (OuterVolumeSpecName: "kube-api-access-zz9kl") pod "1ccadfc0-4f98-4981-b1cb-db16b255e529" (UID: "1ccadfc0-4f98-4981-b1cb-db16b255e529"). InnerVolumeSpecName "kube-api-access-zz9kl". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 08:25:07 crc kubenswrapper[4612]: I0227 08:25:07.276711 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1ccadfc0-4f98-4981-b1cb-db16b255e529-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1ccadfc0-4f98-4981-b1cb-db16b255e529" (UID: "1ccadfc0-4f98-4981-b1cb-db16b255e529"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 08:25:07 crc kubenswrapper[4612]: I0227 08:25:07.357337 4612 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1ccadfc0-4f98-4981-b1cb-db16b255e529-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 27 08:25:07 crc kubenswrapper[4612]: I0227 08:25:07.357379 4612 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1ccadfc0-4f98-4981-b1cb-db16b255e529-utilities\") on node \"crc\" DevicePath \"\"" Feb 27 08:25:07 crc kubenswrapper[4612]: I0227 08:25:07.357393 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zz9kl\" (UniqueName: \"kubernetes.io/projected/1ccadfc0-4f98-4981-b1cb-db16b255e529-kube-api-access-zz9kl\") on node \"crc\" DevicePath \"\"" Feb 27 08:25:07 crc kubenswrapper[4612]: I0227 08:25:07.681051 4612 generic.go:334] "Generic (PLEG): container finished" podID="1ccadfc0-4f98-4981-b1cb-db16b255e529" containerID="0be7c0828c7d8d39cf02bec49f6ad2bc3139c7d473084fbac40872ef067ec239" exitCode=0 Feb 27 08:25:07 crc kubenswrapper[4612]: I0227 08:25:07.681108 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2cvlb" event={"ID":"1ccadfc0-4f98-4981-b1cb-db16b255e529","Type":"ContainerDied","Data":"0be7c0828c7d8d39cf02bec49f6ad2bc3139c7d473084fbac40872ef067ec239"} Feb 27 08:25:07 crc kubenswrapper[4612]: I0227 08:25:07.681154 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2cvlb" event={"ID":"1ccadfc0-4f98-4981-b1cb-db16b255e529","Type":"ContainerDied","Data":"0532d9008e239bc55b242336dc501d517d911c830216662b51268448e0eda93c"} Feb 27 08:25:07 crc kubenswrapper[4612]: I0227 08:25:07.681179 4612 scope.go:117] "RemoveContainer" containerID="0be7c0828c7d8d39cf02bec49f6ad2bc3139c7d473084fbac40872ef067ec239" Feb 27 08:25:07 crc kubenswrapper[4612]: I0227 08:25:07.681213 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2cvlb" Feb 27 08:25:07 crc kubenswrapper[4612]: I0227 08:25:07.710862 4612 scope.go:117] "RemoveContainer" containerID="62951ea4bedbcbfe9dfb70febd1ff488d8be7f6bea8606518d737930e01c18dd" Feb 27 08:25:07 crc kubenswrapper[4612]: I0227 08:25:07.742189 4612 scope.go:117] "RemoveContainer" containerID="e6ba5d3f85cdc42928efac6eaf4b253947c321887b7063b40c56c897464c3f04" Feb 27 08:25:07 crc kubenswrapper[4612]: I0227 08:25:07.750488 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-2cvlb"] Feb 27 08:25:07 crc kubenswrapper[4612]: I0227 08:25:07.760643 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-2cvlb"] Feb 27 08:25:07 crc kubenswrapper[4612]: I0227 08:25:07.786921 4612 scope.go:117] "RemoveContainer" containerID="0be7c0828c7d8d39cf02bec49f6ad2bc3139c7d473084fbac40872ef067ec239" Feb 27 08:25:07 crc kubenswrapper[4612]: E0227 08:25:07.789876 4612 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0be7c0828c7d8d39cf02bec49f6ad2bc3139c7d473084fbac40872ef067ec239\": container with ID starting with 0be7c0828c7d8d39cf02bec49f6ad2bc3139c7d473084fbac40872ef067ec239 not found: ID does not exist" containerID="0be7c0828c7d8d39cf02bec49f6ad2bc3139c7d473084fbac40872ef067ec239" Feb 27 08:25:07 crc kubenswrapper[4612]: I0227 08:25:07.789923 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0be7c0828c7d8d39cf02bec49f6ad2bc3139c7d473084fbac40872ef067ec239"} err="failed to get container status \"0be7c0828c7d8d39cf02bec49f6ad2bc3139c7d473084fbac40872ef067ec239\": rpc error: code = NotFound desc = could not find container \"0be7c0828c7d8d39cf02bec49f6ad2bc3139c7d473084fbac40872ef067ec239\": container with ID starting with 0be7c0828c7d8d39cf02bec49f6ad2bc3139c7d473084fbac40872ef067ec239 not found: ID does not exist" Feb 27 08:25:07 crc kubenswrapper[4612]: I0227 08:25:07.789950 4612 scope.go:117] "RemoveContainer" containerID="62951ea4bedbcbfe9dfb70febd1ff488d8be7f6bea8606518d737930e01c18dd" Feb 27 08:25:07 crc kubenswrapper[4612]: E0227 08:25:07.790226 4612 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"62951ea4bedbcbfe9dfb70febd1ff488d8be7f6bea8606518d737930e01c18dd\": container with ID starting with 62951ea4bedbcbfe9dfb70febd1ff488d8be7f6bea8606518d737930e01c18dd not found: ID does not exist" containerID="62951ea4bedbcbfe9dfb70febd1ff488d8be7f6bea8606518d737930e01c18dd" Feb 27 08:25:07 crc kubenswrapper[4612]: I0227 08:25:07.790267 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"62951ea4bedbcbfe9dfb70febd1ff488d8be7f6bea8606518d737930e01c18dd"} err="failed to get container status \"62951ea4bedbcbfe9dfb70febd1ff488d8be7f6bea8606518d737930e01c18dd\": rpc error: code = NotFound desc = could not find container \"62951ea4bedbcbfe9dfb70febd1ff488d8be7f6bea8606518d737930e01c18dd\": container with ID starting with 62951ea4bedbcbfe9dfb70febd1ff488d8be7f6bea8606518d737930e01c18dd not found: ID does not exist" Feb 27 08:25:07 crc kubenswrapper[4612]: I0227 08:25:07.790281 4612 scope.go:117] "RemoveContainer" containerID="e6ba5d3f85cdc42928efac6eaf4b253947c321887b7063b40c56c897464c3f04" Feb 27 08:25:07 crc kubenswrapper[4612]: E0227 08:25:07.790524 4612 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e6ba5d3f85cdc42928efac6eaf4b253947c321887b7063b40c56c897464c3f04\": container with ID starting with e6ba5d3f85cdc42928efac6eaf4b253947c321887b7063b40c56c897464c3f04 not found: ID does not exist" containerID="e6ba5d3f85cdc42928efac6eaf4b253947c321887b7063b40c56c897464c3f04" Feb 27 08:25:07 crc kubenswrapper[4612]: I0227 08:25:07.790543 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e6ba5d3f85cdc42928efac6eaf4b253947c321887b7063b40c56c897464c3f04"} err="failed to get container status \"e6ba5d3f85cdc42928efac6eaf4b253947c321887b7063b40c56c897464c3f04\": rpc error: code = NotFound desc = could not find container \"e6ba5d3f85cdc42928efac6eaf4b253947c321887b7063b40c56c897464c3f04\": container with ID starting with e6ba5d3f85cdc42928efac6eaf4b253947c321887b7063b40c56c897464c3f04 not found: ID does not exist" Feb 27 08:25:08 crc kubenswrapper[4612]: I0227 08:25:08.871452 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1ccadfc0-4f98-4981-b1cb-db16b255e529" path="/var/lib/kubelet/pods/1ccadfc0-4f98-4981-b1cb-db16b255e529/volumes" Feb 27 08:25:14 crc kubenswrapper[4612]: I0227 08:25:14.143148 4612 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-jdqbp" podUID="cc8baf5d-91f1-4fed-836d-d3eb2acd6854" containerName="registry-server" probeResult="failure" output=< Feb 27 08:25:14 crc kubenswrapper[4612]: timeout: failed to connect service ":50051" within 1s Feb 27 08:25:14 crc kubenswrapper[4612]: > Feb 27 08:25:24 crc kubenswrapper[4612]: I0227 08:25:24.101978 4612 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-jdqbp" podUID="cc8baf5d-91f1-4fed-836d-d3eb2acd6854" containerName="registry-server" probeResult="failure" output=< Feb 27 08:25:24 crc kubenswrapper[4612]: timeout: failed to connect service ":50051" within 1s Feb 27 08:25:24 crc kubenswrapper[4612]: > Feb 27 08:25:27 crc kubenswrapper[4612]: I0227 08:25:27.223793 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-rrgzc"] Feb 27 08:25:27 crc kubenswrapper[4612]: E0227 08:25:27.226228 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ccadfc0-4f98-4981-b1cb-db16b255e529" containerName="registry-server" Feb 27 08:25:27 crc kubenswrapper[4612]: I0227 08:25:27.226348 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ccadfc0-4f98-4981-b1cb-db16b255e529" containerName="registry-server" Feb 27 08:25:27 crc kubenswrapper[4612]: E0227 08:25:27.226520 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ccadfc0-4f98-4981-b1cb-db16b255e529" containerName="extract-content" Feb 27 08:25:27 crc kubenswrapper[4612]: I0227 08:25:27.226635 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ccadfc0-4f98-4981-b1cb-db16b255e529" containerName="extract-content" Feb 27 08:25:27 crc kubenswrapper[4612]: E0227 08:25:27.226769 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ccadfc0-4f98-4981-b1cb-db16b255e529" containerName="extract-utilities" Feb 27 08:25:27 crc kubenswrapper[4612]: I0227 08:25:27.227168 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ccadfc0-4f98-4981-b1cb-db16b255e529" containerName="extract-utilities" Feb 27 08:25:27 crc kubenswrapper[4612]: I0227 08:25:27.228909 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="1ccadfc0-4f98-4981-b1cb-db16b255e529" containerName="registry-server" Feb 27 08:25:27 crc kubenswrapper[4612]: I0227 08:25:27.237416 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rrgzc" Feb 27 08:25:27 crc kubenswrapper[4612]: I0227 08:25:27.253012 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-rrgzc"] Feb 27 08:25:27 crc kubenswrapper[4612]: I0227 08:25:27.264949 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qm869\" (UniqueName: \"kubernetes.io/projected/48d52567-f8d6-4070-9739-43f9df081d20-kube-api-access-qm869\") pod \"certified-operators-rrgzc\" (UID: \"48d52567-f8d6-4070-9739-43f9df081d20\") " pod="openshift-marketplace/certified-operators-rrgzc" Feb 27 08:25:27 crc kubenswrapper[4612]: I0227 08:25:27.265002 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/48d52567-f8d6-4070-9739-43f9df081d20-utilities\") pod \"certified-operators-rrgzc\" (UID: \"48d52567-f8d6-4070-9739-43f9df081d20\") " pod="openshift-marketplace/certified-operators-rrgzc" Feb 27 08:25:27 crc kubenswrapper[4612]: I0227 08:25:27.265081 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/48d52567-f8d6-4070-9739-43f9df081d20-catalog-content\") pod \"certified-operators-rrgzc\" (UID: \"48d52567-f8d6-4070-9739-43f9df081d20\") " pod="openshift-marketplace/certified-operators-rrgzc" Feb 27 08:25:27 crc kubenswrapper[4612]: I0227 08:25:27.366581 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/48d52567-f8d6-4070-9739-43f9df081d20-catalog-content\") pod \"certified-operators-rrgzc\" (UID: \"48d52567-f8d6-4070-9739-43f9df081d20\") " pod="openshift-marketplace/certified-operators-rrgzc" Feb 27 08:25:27 crc kubenswrapper[4612]: I0227 08:25:27.366727 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qm869\" (UniqueName: \"kubernetes.io/projected/48d52567-f8d6-4070-9739-43f9df081d20-kube-api-access-qm869\") pod \"certified-operators-rrgzc\" (UID: \"48d52567-f8d6-4070-9739-43f9df081d20\") " pod="openshift-marketplace/certified-operators-rrgzc" Feb 27 08:25:27 crc kubenswrapper[4612]: I0227 08:25:27.366750 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/48d52567-f8d6-4070-9739-43f9df081d20-utilities\") pod \"certified-operators-rrgzc\" (UID: \"48d52567-f8d6-4070-9739-43f9df081d20\") " pod="openshift-marketplace/certified-operators-rrgzc" Feb 27 08:25:27 crc kubenswrapper[4612]: I0227 08:25:27.367223 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/48d52567-f8d6-4070-9739-43f9df081d20-utilities\") pod \"certified-operators-rrgzc\" (UID: \"48d52567-f8d6-4070-9739-43f9df081d20\") " pod="openshift-marketplace/certified-operators-rrgzc" Feb 27 08:25:27 crc kubenswrapper[4612]: I0227 08:25:27.367439 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/48d52567-f8d6-4070-9739-43f9df081d20-catalog-content\") pod \"certified-operators-rrgzc\" (UID: \"48d52567-f8d6-4070-9739-43f9df081d20\") " pod="openshift-marketplace/certified-operators-rrgzc" Feb 27 08:25:27 crc kubenswrapper[4612]: I0227 08:25:27.405198 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qm869\" (UniqueName: \"kubernetes.io/projected/48d52567-f8d6-4070-9739-43f9df081d20-kube-api-access-qm869\") pod \"certified-operators-rrgzc\" (UID: \"48d52567-f8d6-4070-9739-43f9df081d20\") " pod="openshift-marketplace/certified-operators-rrgzc" Feb 27 08:25:27 crc kubenswrapper[4612]: I0227 08:25:27.608481 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rrgzc" Feb 27 08:25:28 crc kubenswrapper[4612]: I0227 08:25:28.101039 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-rrgzc"] Feb 27 08:25:28 crc kubenswrapper[4612]: I0227 08:25:28.882785 4612 generic.go:334] "Generic (PLEG): container finished" podID="48d52567-f8d6-4070-9739-43f9df081d20" containerID="91c4f5cea9f4458bb8b898e8be2c2b529fe980d2c5259ba859f2815a6286a055" exitCode=0 Feb 27 08:25:28 crc kubenswrapper[4612]: I0227 08:25:28.882866 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rrgzc" event={"ID":"48d52567-f8d6-4070-9739-43f9df081d20","Type":"ContainerDied","Data":"91c4f5cea9f4458bb8b898e8be2c2b529fe980d2c5259ba859f2815a6286a055"} Feb 27 08:25:28 crc kubenswrapper[4612]: I0227 08:25:28.883410 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rrgzc" event={"ID":"48d52567-f8d6-4070-9739-43f9df081d20","Type":"ContainerStarted","Data":"7a56b8bd24b67c30cddf4ab0a738eb01d99fc7a629a7d5c7dcd9716817b48b6b"} Feb 27 08:25:29 crc kubenswrapper[4612]: I0227 08:25:29.893509 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rrgzc" event={"ID":"48d52567-f8d6-4070-9739-43f9df081d20","Type":"ContainerStarted","Data":"d4514d5817780b2d1fda354f54f719ed3f0b8599d16ab0a4242c57fe5c594932"} Feb 27 08:25:32 crc kubenswrapper[4612]: I0227 08:25:32.931685 4612 generic.go:334] "Generic (PLEG): container finished" podID="48d52567-f8d6-4070-9739-43f9df081d20" containerID="d4514d5817780b2d1fda354f54f719ed3f0b8599d16ab0a4242c57fe5c594932" exitCode=0 Feb 27 08:25:32 crc kubenswrapper[4612]: I0227 08:25:32.932112 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rrgzc" event={"ID":"48d52567-f8d6-4070-9739-43f9df081d20","Type":"ContainerDied","Data":"d4514d5817780b2d1fda354f54f719ed3f0b8599d16ab0a4242c57fe5c594932"} Feb 27 08:25:33 crc kubenswrapper[4612]: I0227 08:25:33.090053 4612 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-jdqbp" Feb 27 08:25:33 crc kubenswrapper[4612]: I0227 08:25:33.148645 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-jdqbp" Feb 27 08:25:33 crc kubenswrapper[4612]: I0227 08:25:33.588862 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-jdqbp"] Feb 27 08:25:34 crc kubenswrapper[4612]: I0227 08:25:34.954717 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rrgzc" event={"ID":"48d52567-f8d6-4070-9739-43f9df081d20","Type":"ContainerStarted","Data":"587347ac26807084596cbb37f3b758c41b49fff7c59a89175fc458cc181cb3f9"} Feb 27 08:25:34 crc kubenswrapper[4612]: I0227 08:25:34.954983 4612 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-jdqbp" podUID="cc8baf5d-91f1-4fed-836d-d3eb2acd6854" containerName="registry-server" containerID="cri-o://0630a0b7713cbeb711b50f8e42f6c495aca8cc9e1a942a420753cb223fca9ea2" gracePeriod=2 Feb 27 08:25:34 crc kubenswrapper[4612]: I0227 08:25:34.986008 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-rrgzc" podStartSLOduration=2.927835437 podStartE2EDuration="7.985985663s" podCreationTimestamp="2026-02-27 08:25:27 +0000 UTC" firstStartedPulling="2026-02-27 08:25:28.886620316 +0000 UTC m=+2186.740550314" lastFinishedPulling="2026-02-27 08:25:33.944770492 +0000 UTC m=+2191.798700540" observedRunningTime="2026-02-27 08:25:34.977050147 +0000 UTC m=+2192.830980155" watchObservedRunningTime="2026-02-27 08:25:34.985985663 +0000 UTC m=+2192.839915661" Feb 27 08:25:35 crc kubenswrapper[4612]: I0227 08:25:35.409390 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jdqbp" Feb 27 08:25:35 crc kubenswrapper[4612]: I0227 08:25:35.454318 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cc8baf5d-91f1-4fed-836d-d3eb2acd6854-catalog-content\") pod \"cc8baf5d-91f1-4fed-836d-d3eb2acd6854\" (UID: \"cc8baf5d-91f1-4fed-836d-d3eb2acd6854\") " Feb 27 08:25:35 crc kubenswrapper[4612]: I0227 08:25:35.454384 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tqq65\" (UniqueName: \"kubernetes.io/projected/cc8baf5d-91f1-4fed-836d-d3eb2acd6854-kube-api-access-tqq65\") pod \"cc8baf5d-91f1-4fed-836d-d3eb2acd6854\" (UID: \"cc8baf5d-91f1-4fed-836d-d3eb2acd6854\") " Feb 27 08:25:35 crc kubenswrapper[4612]: I0227 08:25:35.454464 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cc8baf5d-91f1-4fed-836d-d3eb2acd6854-utilities\") pod \"cc8baf5d-91f1-4fed-836d-d3eb2acd6854\" (UID: \"cc8baf5d-91f1-4fed-836d-d3eb2acd6854\") " Feb 27 08:25:35 crc kubenswrapper[4612]: I0227 08:25:35.455626 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cc8baf5d-91f1-4fed-836d-d3eb2acd6854-utilities" (OuterVolumeSpecName: "utilities") pod "cc8baf5d-91f1-4fed-836d-d3eb2acd6854" (UID: "cc8baf5d-91f1-4fed-836d-d3eb2acd6854"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 08:25:35 crc kubenswrapper[4612]: I0227 08:25:35.461289 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cc8baf5d-91f1-4fed-836d-d3eb2acd6854-kube-api-access-tqq65" (OuterVolumeSpecName: "kube-api-access-tqq65") pod "cc8baf5d-91f1-4fed-836d-d3eb2acd6854" (UID: "cc8baf5d-91f1-4fed-836d-d3eb2acd6854"). InnerVolumeSpecName "kube-api-access-tqq65". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 08:25:35 crc kubenswrapper[4612]: I0227 08:25:35.556803 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tqq65\" (UniqueName: \"kubernetes.io/projected/cc8baf5d-91f1-4fed-836d-d3eb2acd6854-kube-api-access-tqq65\") on node \"crc\" DevicePath \"\"" Feb 27 08:25:35 crc kubenswrapper[4612]: I0227 08:25:35.556829 4612 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cc8baf5d-91f1-4fed-836d-d3eb2acd6854-utilities\") on node \"crc\" DevicePath \"\"" Feb 27 08:25:35 crc kubenswrapper[4612]: I0227 08:25:35.581773 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cc8baf5d-91f1-4fed-836d-d3eb2acd6854-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "cc8baf5d-91f1-4fed-836d-d3eb2acd6854" (UID: "cc8baf5d-91f1-4fed-836d-d3eb2acd6854"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 08:25:35 crc kubenswrapper[4612]: I0227 08:25:35.658248 4612 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cc8baf5d-91f1-4fed-836d-d3eb2acd6854-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 27 08:25:35 crc kubenswrapper[4612]: I0227 08:25:35.962531 4612 generic.go:334] "Generic (PLEG): container finished" podID="cc8baf5d-91f1-4fed-836d-d3eb2acd6854" containerID="0630a0b7713cbeb711b50f8e42f6c495aca8cc9e1a942a420753cb223fca9ea2" exitCode=0 Feb 27 08:25:35 crc kubenswrapper[4612]: I0227 08:25:35.962578 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jdqbp" event={"ID":"cc8baf5d-91f1-4fed-836d-d3eb2acd6854","Type":"ContainerDied","Data":"0630a0b7713cbeb711b50f8e42f6c495aca8cc9e1a942a420753cb223fca9ea2"} Feb 27 08:25:35 crc kubenswrapper[4612]: I0227 08:25:35.962596 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jdqbp" Feb 27 08:25:35 crc kubenswrapper[4612]: I0227 08:25:35.962615 4612 scope.go:117] "RemoveContainer" containerID="0630a0b7713cbeb711b50f8e42f6c495aca8cc9e1a942a420753cb223fca9ea2" Feb 27 08:25:35 crc kubenswrapper[4612]: I0227 08:25:35.962604 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jdqbp" event={"ID":"cc8baf5d-91f1-4fed-836d-d3eb2acd6854","Type":"ContainerDied","Data":"4b3b52a1c7ffb57a9197507b9bdeca573bb7cd15af01cb679425b90f656eeb1a"} Feb 27 08:25:35 crc kubenswrapper[4612]: I0227 08:25:35.983152 4612 scope.go:117] "RemoveContainer" containerID="4f7fb07b1662532350bed1c58e2175b77fc29ef44a333d70dc9ccc77074f01ee" Feb 27 08:25:36 crc kubenswrapper[4612]: I0227 08:25:36.013342 4612 scope.go:117] "RemoveContainer" containerID="220c30df5c00042ad8b9eefe0a39591caf49685b318eb386322ed07d2bdbcb94" Feb 27 08:25:36 crc kubenswrapper[4612]: I0227 08:25:36.017768 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-jdqbp"] Feb 27 08:25:36 crc kubenswrapper[4612]: I0227 08:25:36.022555 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-jdqbp"] Feb 27 08:25:36 crc kubenswrapper[4612]: I0227 08:25:36.062558 4612 scope.go:117] "RemoveContainer" containerID="0630a0b7713cbeb711b50f8e42f6c495aca8cc9e1a942a420753cb223fca9ea2" Feb 27 08:25:36 crc kubenswrapper[4612]: E0227 08:25:36.062973 4612 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0630a0b7713cbeb711b50f8e42f6c495aca8cc9e1a942a420753cb223fca9ea2\": container with ID starting with 0630a0b7713cbeb711b50f8e42f6c495aca8cc9e1a942a420753cb223fca9ea2 not found: ID does not exist" containerID="0630a0b7713cbeb711b50f8e42f6c495aca8cc9e1a942a420753cb223fca9ea2" Feb 27 08:25:36 crc kubenswrapper[4612]: I0227 08:25:36.063028 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0630a0b7713cbeb711b50f8e42f6c495aca8cc9e1a942a420753cb223fca9ea2"} err="failed to get container status \"0630a0b7713cbeb711b50f8e42f6c495aca8cc9e1a942a420753cb223fca9ea2\": rpc error: code = NotFound desc = could not find container \"0630a0b7713cbeb711b50f8e42f6c495aca8cc9e1a942a420753cb223fca9ea2\": container with ID starting with 0630a0b7713cbeb711b50f8e42f6c495aca8cc9e1a942a420753cb223fca9ea2 not found: ID does not exist" Feb 27 08:25:36 crc kubenswrapper[4612]: I0227 08:25:36.063050 4612 scope.go:117] "RemoveContainer" containerID="4f7fb07b1662532350bed1c58e2175b77fc29ef44a333d70dc9ccc77074f01ee" Feb 27 08:25:36 crc kubenswrapper[4612]: E0227 08:25:36.063280 4612 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4f7fb07b1662532350bed1c58e2175b77fc29ef44a333d70dc9ccc77074f01ee\": container with ID starting with 4f7fb07b1662532350bed1c58e2175b77fc29ef44a333d70dc9ccc77074f01ee not found: ID does not exist" containerID="4f7fb07b1662532350bed1c58e2175b77fc29ef44a333d70dc9ccc77074f01ee" Feb 27 08:25:36 crc kubenswrapper[4612]: I0227 08:25:36.063298 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4f7fb07b1662532350bed1c58e2175b77fc29ef44a333d70dc9ccc77074f01ee"} err="failed to get container status \"4f7fb07b1662532350bed1c58e2175b77fc29ef44a333d70dc9ccc77074f01ee\": rpc error: code = NotFound desc = could not find container \"4f7fb07b1662532350bed1c58e2175b77fc29ef44a333d70dc9ccc77074f01ee\": container with ID starting with 4f7fb07b1662532350bed1c58e2175b77fc29ef44a333d70dc9ccc77074f01ee not found: ID does not exist" Feb 27 08:25:36 crc kubenswrapper[4612]: I0227 08:25:36.063310 4612 scope.go:117] "RemoveContainer" containerID="220c30df5c00042ad8b9eefe0a39591caf49685b318eb386322ed07d2bdbcb94" Feb 27 08:25:36 crc kubenswrapper[4612]: E0227 08:25:36.063537 4612 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"220c30df5c00042ad8b9eefe0a39591caf49685b318eb386322ed07d2bdbcb94\": container with ID starting with 220c30df5c00042ad8b9eefe0a39591caf49685b318eb386322ed07d2bdbcb94 not found: ID does not exist" containerID="220c30df5c00042ad8b9eefe0a39591caf49685b318eb386322ed07d2bdbcb94" Feb 27 08:25:36 crc kubenswrapper[4612]: I0227 08:25:36.063557 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"220c30df5c00042ad8b9eefe0a39591caf49685b318eb386322ed07d2bdbcb94"} err="failed to get container status \"220c30df5c00042ad8b9eefe0a39591caf49685b318eb386322ed07d2bdbcb94\": rpc error: code = NotFound desc = could not find container \"220c30df5c00042ad8b9eefe0a39591caf49685b318eb386322ed07d2bdbcb94\": container with ID starting with 220c30df5c00042ad8b9eefe0a39591caf49685b318eb386322ed07d2bdbcb94 not found: ID does not exist" Feb 27 08:25:36 crc kubenswrapper[4612]: I0227 08:25:36.863522 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cc8baf5d-91f1-4fed-836d-d3eb2acd6854" path="/var/lib/kubelet/pods/cc8baf5d-91f1-4fed-836d-d3eb2acd6854/volumes" Feb 27 08:25:37 crc kubenswrapper[4612]: I0227 08:25:37.608893 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-rrgzc" Feb 27 08:25:37 crc kubenswrapper[4612]: I0227 08:25:37.609866 4612 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-rrgzc" Feb 27 08:25:38 crc kubenswrapper[4612]: I0227 08:25:38.657718 4612 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-rrgzc" podUID="48d52567-f8d6-4070-9739-43f9df081d20" containerName="registry-server" probeResult="failure" output=< Feb 27 08:25:38 crc kubenswrapper[4612]: timeout: failed to connect service ":50051" within 1s Feb 27 08:25:38 crc kubenswrapper[4612]: > Feb 27 08:25:45 crc kubenswrapper[4612]: I0227 08:25:45.056561 4612 generic.go:334] "Generic (PLEG): container finished" podID="dfdf6a2b-0ca5-4ea9-b87c-fd7a61f4981c" containerID="27ec6f685d7b9a699df1b6325897c2133b0e8eadd1579c1405cf884d40a31d91" exitCode=0 Feb 27 08:25:45 crc kubenswrapper[4612]: I0227 08:25:45.056727 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-xhcrn" event={"ID":"dfdf6a2b-0ca5-4ea9-b87c-fd7a61f4981c","Type":"ContainerDied","Data":"27ec6f685d7b9a699df1b6325897c2133b0e8eadd1579c1405cf884d40a31d91"} Feb 27 08:25:46 crc kubenswrapper[4612]: I0227 08:25:46.027064 4612 patch_prober.go:28] interesting pod/machine-config-daemon-924vb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 27 08:25:46 crc kubenswrapper[4612]: I0227 08:25:46.027581 4612 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 27 08:25:46 crc kubenswrapper[4612]: I0227 08:25:46.489274 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-xhcrn" Feb 27 08:25:46 crc kubenswrapper[4612]: I0227 08:25:46.602091 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/dfdf6a2b-0ca5-4ea9-b87c-fd7a61f4981c-ovncontroller-config-0\") pod \"dfdf6a2b-0ca5-4ea9-b87c-fd7a61f4981c\" (UID: \"dfdf6a2b-0ca5-4ea9-b87c-fd7a61f4981c\") " Feb 27 08:25:46 crc kubenswrapper[4612]: I0227 08:25:46.602160 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/dfdf6a2b-0ca5-4ea9-b87c-fd7a61f4981c-inventory\") pod \"dfdf6a2b-0ca5-4ea9-b87c-fd7a61f4981c\" (UID: \"dfdf6a2b-0ca5-4ea9-b87c-fd7a61f4981c\") " Feb 27 08:25:46 crc kubenswrapper[4612]: I0227 08:25:46.602987 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sgp5v\" (UniqueName: \"kubernetes.io/projected/dfdf6a2b-0ca5-4ea9-b87c-fd7a61f4981c-kube-api-access-sgp5v\") pod \"dfdf6a2b-0ca5-4ea9-b87c-fd7a61f4981c\" (UID: \"dfdf6a2b-0ca5-4ea9-b87c-fd7a61f4981c\") " Feb 27 08:25:46 crc kubenswrapper[4612]: I0227 08:25:46.603036 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/dfdf6a2b-0ca5-4ea9-b87c-fd7a61f4981c-ssh-key-openstack-edpm-ipam\") pod \"dfdf6a2b-0ca5-4ea9-b87c-fd7a61f4981c\" (UID: \"dfdf6a2b-0ca5-4ea9-b87c-fd7a61f4981c\") " Feb 27 08:25:46 crc kubenswrapper[4612]: I0227 08:25:46.603103 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dfdf6a2b-0ca5-4ea9-b87c-fd7a61f4981c-ovn-combined-ca-bundle\") pod \"dfdf6a2b-0ca5-4ea9-b87c-fd7a61f4981c\" (UID: \"dfdf6a2b-0ca5-4ea9-b87c-fd7a61f4981c\") " Feb 27 08:25:46 crc kubenswrapper[4612]: I0227 08:25:46.609815 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dfdf6a2b-0ca5-4ea9-b87c-fd7a61f4981c-kube-api-access-sgp5v" (OuterVolumeSpecName: "kube-api-access-sgp5v") pod "dfdf6a2b-0ca5-4ea9-b87c-fd7a61f4981c" (UID: "dfdf6a2b-0ca5-4ea9-b87c-fd7a61f4981c"). InnerVolumeSpecName "kube-api-access-sgp5v". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 08:25:46 crc kubenswrapper[4612]: I0227 08:25:46.620666 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dfdf6a2b-0ca5-4ea9-b87c-fd7a61f4981c-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "dfdf6a2b-0ca5-4ea9-b87c-fd7a61f4981c" (UID: "dfdf6a2b-0ca5-4ea9-b87c-fd7a61f4981c"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:25:46 crc kubenswrapper[4612]: I0227 08:25:46.632527 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dfdf6a2b-0ca5-4ea9-b87c-fd7a61f4981c-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "dfdf6a2b-0ca5-4ea9-b87c-fd7a61f4981c" (UID: "dfdf6a2b-0ca5-4ea9-b87c-fd7a61f4981c"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:25:46 crc kubenswrapper[4612]: I0227 08:25:46.637041 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dfdf6a2b-0ca5-4ea9-b87c-fd7a61f4981c-inventory" (OuterVolumeSpecName: "inventory") pod "dfdf6a2b-0ca5-4ea9-b87c-fd7a61f4981c" (UID: "dfdf6a2b-0ca5-4ea9-b87c-fd7a61f4981c"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:25:46 crc kubenswrapper[4612]: I0227 08:25:46.652265 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dfdf6a2b-0ca5-4ea9-b87c-fd7a61f4981c-ovncontroller-config-0" (OuterVolumeSpecName: "ovncontroller-config-0") pod "dfdf6a2b-0ca5-4ea9-b87c-fd7a61f4981c" (UID: "dfdf6a2b-0ca5-4ea9-b87c-fd7a61f4981c"). InnerVolumeSpecName "ovncontroller-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 08:25:46 crc kubenswrapper[4612]: I0227 08:25:46.710235 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sgp5v\" (UniqueName: \"kubernetes.io/projected/dfdf6a2b-0ca5-4ea9-b87c-fd7a61f4981c-kube-api-access-sgp5v\") on node \"crc\" DevicePath \"\"" Feb 27 08:25:46 crc kubenswrapper[4612]: I0227 08:25:46.710285 4612 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/dfdf6a2b-0ca5-4ea9-b87c-fd7a61f4981c-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 27 08:25:46 crc kubenswrapper[4612]: I0227 08:25:46.710300 4612 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dfdf6a2b-0ca5-4ea9-b87c-fd7a61f4981c-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 08:25:46 crc kubenswrapper[4612]: I0227 08:25:46.710315 4612 reconciler_common.go:293] "Volume detached for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/dfdf6a2b-0ca5-4ea9-b87c-fd7a61f4981c-ovncontroller-config-0\") on node \"crc\" DevicePath \"\"" Feb 27 08:25:46 crc kubenswrapper[4612]: I0227 08:25:46.710330 4612 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/dfdf6a2b-0ca5-4ea9-b87c-fd7a61f4981c-inventory\") on node \"crc\" DevicePath \"\"" Feb 27 08:25:47 crc kubenswrapper[4612]: I0227 08:25:47.079454 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-xhcrn" event={"ID":"dfdf6a2b-0ca5-4ea9-b87c-fd7a61f4981c","Type":"ContainerDied","Data":"889b74e01e806ba955d88047bf3550daf9409f1ef1ec9488ce6942705419a8fd"} Feb 27 08:25:47 crc kubenswrapper[4612]: I0227 08:25:47.079502 4612 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="889b74e01e806ba955d88047bf3550daf9409f1ef1ec9488ce6942705419a8fd" Feb 27 08:25:47 crc kubenswrapper[4612]: I0227 08:25:47.079519 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-xhcrn" Feb 27 08:25:47 crc kubenswrapper[4612]: I0227 08:25:47.201290 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-x52ht"] Feb 27 08:25:47 crc kubenswrapper[4612]: E0227 08:25:47.201811 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc8baf5d-91f1-4fed-836d-d3eb2acd6854" containerName="extract-content" Feb 27 08:25:47 crc kubenswrapper[4612]: I0227 08:25:47.201836 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc8baf5d-91f1-4fed-836d-d3eb2acd6854" containerName="extract-content" Feb 27 08:25:47 crc kubenswrapper[4612]: E0227 08:25:47.201870 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc8baf5d-91f1-4fed-836d-d3eb2acd6854" containerName="registry-server" Feb 27 08:25:47 crc kubenswrapper[4612]: I0227 08:25:47.201882 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc8baf5d-91f1-4fed-836d-d3eb2acd6854" containerName="registry-server" Feb 27 08:25:47 crc kubenswrapper[4612]: E0227 08:25:47.201906 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc8baf5d-91f1-4fed-836d-d3eb2acd6854" containerName="extract-utilities" Feb 27 08:25:47 crc kubenswrapper[4612]: I0227 08:25:47.201917 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc8baf5d-91f1-4fed-836d-d3eb2acd6854" containerName="extract-utilities" Feb 27 08:25:47 crc kubenswrapper[4612]: E0227 08:25:47.201940 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dfdf6a2b-0ca5-4ea9-b87c-fd7a61f4981c" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Feb 27 08:25:47 crc kubenswrapper[4612]: I0227 08:25:47.201949 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="dfdf6a2b-0ca5-4ea9-b87c-fd7a61f4981c" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Feb 27 08:25:47 crc kubenswrapper[4612]: I0227 08:25:47.202222 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="dfdf6a2b-0ca5-4ea9-b87c-fd7a61f4981c" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Feb 27 08:25:47 crc kubenswrapper[4612]: I0227 08:25:47.202261 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="cc8baf5d-91f1-4fed-836d-d3eb2acd6854" containerName="registry-server" Feb 27 08:25:47 crc kubenswrapper[4612]: I0227 08:25:47.203032 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-x52ht" Feb 27 08:25:47 crc kubenswrapper[4612]: I0227 08:25:47.206851 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-neutron-config" Feb 27 08:25:47 crc kubenswrapper[4612]: I0227 08:25:47.210084 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 27 08:25:47 crc kubenswrapper[4612]: I0227 08:25:47.210430 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-ovn-metadata-agent-neutron-config" Feb 27 08:25:47 crc kubenswrapper[4612]: I0227 08:25:47.210601 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 27 08:25:47 crc kubenswrapper[4612]: I0227 08:25:47.210928 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 27 08:25:47 crc kubenswrapper[4612]: I0227 08:25:47.212622 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-2pwrb" Feb 27 08:25:47 crc kubenswrapper[4612]: I0227 08:25:47.221791 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-x52ht"] Feb 27 08:25:47 crc kubenswrapper[4612]: I0227 08:25:47.321742 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9f8tj\" (UniqueName: \"kubernetes.io/projected/a213a1c0-9409-487d-834a-a60e11fd8a66-kube-api-access-9f8tj\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-x52ht\" (UID: \"a213a1c0-9409-487d-834a-a60e11fd8a66\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-x52ht" Feb 27 08:25:47 crc kubenswrapper[4612]: I0227 08:25:47.322111 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/a213a1c0-9409-487d-834a-a60e11fd8a66-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-x52ht\" (UID: \"a213a1c0-9409-487d-834a-a60e11fd8a66\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-x52ht" Feb 27 08:25:47 crc kubenswrapper[4612]: I0227 08:25:47.322530 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/a213a1c0-9409-487d-834a-a60e11fd8a66-ssh-key-openstack-edpm-ipam\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-x52ht\" (UID: \"a213a1c0-9409-487d-834a-a60e11fd8a66\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-x52ht" Feb 27 08:25:47 crc kubenswrapper[4612]: I0227 08:25:47.322684 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a213a1c0-9409-487d-834a-a60e11fd8a66-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-x52ht\" (UID: \"a213a1c0-9409-487d-834a-a60e11fd8a66\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-x52ht" Feb 27 08:25:47 crc kubenswrapper[4612]: I0227 08:25:47.322846 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a213a1c0-9409-487d-834a-a60e11fd8a66-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-x52ht\" (UID: \"a213a1c0-9409-487d-834a-a60e11fd8a66\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-x52ht" Feb 27 08:25:47 crc kubenswrapper[4612]: I0227 08:25:47.322921 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/a213a1c0-9409-487d-834a-a60e11fd8a66-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-x52ht\" (UID: \"a213a1c0-9409-487d-834a-a60e11fd8a66\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-x52ht" Feb 27 08:25:47 crc kubenswrapper[4612]: I0227 08:25:47.425131 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/a213a1c0-9409-487d-834a-a60e11fd8a66-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-x52ht\" (UID: \"a213a1c0-9409-487d-834a-a60e11fd8a66\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-x52ht" Feb 27 08:25:47 crc kubenswrapper[4612]: I0227 08:25:47.425254 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9f8tj\" (UniqueName: \"kubernetes.io/projected/a213a1c0-9409-487d-834a-a60e11fd8a66-kube-api-access-9f8tj\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-x52ht\" (UID: \"a213a1c0-9409-487d-834a-a60e11fd8a66\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-x52ht" Feb 27 08:25:47 crc kubenswrapper[4612]: I0227 08:25:47.425321 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/a213a1c0-9409-487d-834a-a60e11fd8a66-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-x52ht\" (UID: \"a213a1c0-9409-487d-834a-a60e11fd8a66\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-x52ht" Feb 27 08:25:47 crc kubenswrapper[4612]: I0227 08:25:47.425469 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/a213a1c0-9409-487d-834a-a60e11fd8a66-ssh-key-openstack-edpm-ipam\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-x52ht\" (UID: \"a213a1c0-9409-487d-834a-a60e11fd8a66\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-x52ht" Feb 27 08:25:47 crc kubenswrapper[4612]: I0227 08:25:47.425516 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a213a1c0-9409-487d-834a-a60e11fd8a66-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-x52ht\" (UID: \"a213a1c0-9409-487d-834a-a60e11fd8a66\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-x52ht" Feb 27 08:25:47 crc kubenswrapper[4612]: I0227 08:25:47.425561 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a213a1c0-9409-487d-834a-a60e11fd8a66-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-x52ht\" (UID: \"a213a1c0-9409-487d-834a-a60e11fd8a66\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-x52ht" Feb 27 08:25:47 crc kubenswrapper[4612]: I0227 08:25:47.431191 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/a213a1c0-9409-487d-834a-a60e11fd8a66-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-x52ht\" (UID: \"a213a1c0-9409-487d-834a-a60e11fd8a66\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-x52ht" Feb 27 08:25:47 crc kubenswrapper[4612]: I0227 08:25:47.431840 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/a213a1c0-9409-487d-834a-a60e11fd8a66-ssh-key-openstack-edpm-ipam\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-x52ht\" (UID: \"a213a1c0-9409-487d-834a-a60e11fd8a66\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-x52ht" Feb 27 08:25:47 crc kubenswrapper[4612]: I0227 08:25:47.432152 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/a213a1c0-9409-487d-834a-a60e11fd8a66-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-x52ht\" (UID: \"a213a1c0-9409-487d-834a-a60e11fd8a66\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-x52ht" Feb 27 08:25:47 crc kubenswrapper[4612]: I0227 08:25:47.433029 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a213a1c0-9409-487d-834a-a60e11fd8a66-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-x52ht\" (UID: \"a213a1c0-9409-487d-834a-a60e11fd8a66\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-x52ht" Feb 27 08:25:47 crc kubenswrapper[4612]: I0227 08:25:47.433726 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a213a1c0-9409-487d-834a-a60e11fd8a66-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-x52ht\" (UID: \"a213a1c0-9409-487d-834a-a60e11fd8a66\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-x52ht" Feb 27 08:25:47 crc kubenswrapper[4612]: I0227 08:25:47.456458 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9f8tj\" (UniqueName: \"kubernetes.io/projected/a213a1c0-9409-487d-834a-a60e11fd8a66-kube-api-access-9f8tj\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-x52ht\" (UID: \"a213a1c0-9409-487d-834a-a60e11fd8a66\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-x52ht" Feb 27 08:25:47 crc kubenswrapper[4612]: I0227 08:25:47.581106 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-x52ht" Feb 27 08:25:47 crc kubenswrapper[4612]: I0227 08:25:47.687984 4612 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-rrgzc" Feb 27 08:25:47 crc kubenswrapper[4612]: I0227 08:25:47.772834 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-rrgzc" Feb 27 08:25:48 crc kubenswrapper[4612]: I0227 08:25:47.930161 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-rrgzc"] Feb 27 08:25:48 crc kubenswrapper[4612]: I0227 08:25:48.235808 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-x52ht"] Feb 27 08:25:49 crc kubenswrapper[4612]: I0227 08:25:49.097065 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-x52ht" event={"ID":"a213a1c0-9409-487d-834a-a60e11fd8a66","Type":"ContainerStarted","Data":"7181ed7bee7378f3a23dd7d8df1c2f43ecf310bf461cd3323b3727f5cb458b5f"} Feb 27 08:25:49 crc kubenswrapper[4612]: I0227 08:25:49.097436 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-x52ht" event={"ID":"a213a1c0-9409-487d-834a-a60e11fd8a66","Type":"ContainerStarted","Data":"101802c5e706e590e09c02cc021707d3b1368fdee83e9faa0741b52c52516775"} Feb 27 08:25:49 crc kubenswrapper[4612]: I0227 08:25:49.097311 4612 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-rrgzc" podUID="48d52567-f8d6-4070-9739-43f9df081d20" containerName="registry-server" containerID="cri-o://587347ac26807084596cbb37f3b758c41b49fff7c59a89175fc458cc181cb3f9" gracePeriod=2 Feb 27 08:25:49 crc kubenswrapper[4612]: I0227 08:25:49.114921 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-x52ht" podStartSLOduration=1.909071343 podStartE2EDuration="2.11489968s" podCreationTimestamp="2026-02-27 08:25:47 +0000 UTC" firstStartedPulling="2026-02-27 08:25:48.233371611 +0000 UTC m=+2206.087301609" lastFinishedPulling="2026-02-27 08:25:48.439199928 +0000 UTC m=+2206.293129946" observedRunningTime="2026-02-27 08:25:49.112235764 +0000 UTC m=+2206.966165772" watchObservedRunningTime="2026-02-27 08:25:49.11489968 +0000 UTC m=+2206.968829678" Feb 27 08:25:49 crc kubenswrapper[4612]: I0227 08:25:49.584262 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rrgzc" Feb 27 08:25:49 crc kubenswrapper[4612]: I0227 08:25:49.710802 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qm869\" (UniqueName: \"kubernetes.io/projected/48d52567-f8d6-4070-9739-43f9df081d20-kube-api-access-qm869\") pod \"48d52567-f8d6-4070-9739-43f9df081d20\" (UID: \"48d52567-f8d6-4070-9739-43f9df081d20\") " Feb 27 08:25:49 crc kubenswrapper[4612]: I0227 08:25:49.711316 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/48d52567-f8d6-4070-9739-43f9df081d20-utilities\") pod \"48d52567-f8d6-4070-9739-43f9df081d20\" (UID: \"48d52567-f8d6-4070-9739-43f9df081d20\") " Feb 27 08:25:49 crc kubenswrapper[4612]: I0227 08:25:49.711346 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/48d52567-f8d6-4070-9739-43f9df081d20-catalog-content\") pod \"48d52567-f8d6-4070-9739-43f9df081d20\" (UID: \"48d52567-f8d6-4070-9739-43f9df081d20\") " Feb 27 08:25:49 crc kubenswrapper[4612]: I0227 08:25:49.713247 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/48d52567-f8d6-4070-9739-43f9df081d20-utilities" (OuterVolumeSpecName: "utilities") pod "48d52567-f8d6-4070-9739-43f9df081d20" (UID: "48d52567-f8d6-4070-9739-43f9df081d20"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 08:25:49 crc kubenswrapper[4612]: I0227 08:25:49.734812 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/48d52567-f8d6-4070-9739-43f9df081d20-kube-api-access-qm869" (OuterVolumeSpecName: "kube-api-access-qm869") pod "48d52567-f8d6-4070-9739-43f9df081d20" (UID: "48d52567-f8d6-4070-9739-43f9df081d20"). InnerVolumeSpecName "kube-api-access-qm869". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 08:25:49 crc kubenswrapper[4612]: I0227 08:25:49.775065 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/48d52567-f8d6-4070-9739-43f9df081d20-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "48d52567-f8d6-4070-9739-43f9df081d20" (UID: "48d52567-f8d6-4070-9739-43f9df081d20"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 08:25:49 crc kubenswrapper[4612]: I0227 08:25:49.813215 4612 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/48d52567-f8d6-4070-9739-43f9df081d20-utilities\") on node \"crc\" DevicePath \"\"" Feb 27 08:25:49 crc kubenswrapper[4612]: I0227 08:25:49.813241 4612 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/48d52567-f8d6-4070-9739-43f9df081d20-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 27 08:25:49 crc kubenswrapper[4612]: I0227 08:25:49.813251 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qm869\" (UniqueName: \"kubernetes.io/projected/48d52567-f8d6-4070-9739-43f9df081d20-kube-api-access-qm869\") on node \"crc\" DevicePath \"\"" Feb 27 08:25:50 crc kubenswrapper[4612]: I0227 08:25:50.109028 4612 generic.go:334] "Generic (PLEG): container finished" podID="48d52567-f8d6-4070-9739-43f9df081d20" containerID="587347ac26807084596cbb37f3b758c41b49fff7c59a89175fc458cc181cb3f9" exitCode=0 Feb 27 08:25:50 crc kubenswrapper[4612]: I0227 08:25:50.109117 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rrgzc" Feb 27 08:25:50 crc kubenswrapper[4612]: I0227 08:25:50.109188 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rrgzc" event={"ID":"48d52567-f8d6-4070-9739-43f9df081d20","Type":"ContainerDied","Data":"587347ac26807084596cbb37f3b758c41b49fff7c59a89175fc458cc181cb3f9"} Feb 27 08:25:50 crc kubenswrapper[4612]: I0227 08:25:50.109230 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rrgzc" event={"ID":"48d52567-f8d6-4070-9739-43f9df081d20","Type":"ContainerDied","Data":"7a56b8bd24b67c30cddf4ab0a738eb01d99fc7a629a7d5c7dcd9716817b48b6b"} Feb 27 08:25:50 crc kubenswrapper[4612]: I0227 08:25:50.109255 4612 scope.go:117] "RemoveContainer" containerID="587347ac26807084596cbb37f3b758c41b49fff7c59a89175fc458cc181cb3f9" Feb 27 08:25:50 crc kubenswrapper[4612]: I0227 08:25:50.151124 4612 scope.go:117] "RemoveContainer" containerID="d4514d5817780b2d1fda354f54f719ed3f0b8599d16ab0a4242c57fe5c594932" Feb 27 08:25:50 crc kubenswrapper[4612]: I0227 08:25:50.171295 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-rrgzc"] Feb 27 08:25:50 crc kubenswrapper[4612]: I0227 08:25:50.180861 4612 scope.go:117] "RemoveContainer" containerID="91c4f5cea9f4458bb8b898e8be2c2b529fe980d2c5259ba859f2815a6286a055" Feb 27 08:25:50 crc kubenswrapper[4612]: I0227 08:25:50.181328 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-rrgzc"] Feb 27 08:25:50 crc kubenswrapper[4612]: I0227 08:25:50.232878 4612 scope.go:117] "RemoveContainer" containerID="587347ac26807084596cbb37f3b758c41b49fff7c59a89175fc458cc181cb3f9" Feb 27 08:25:50 crc kubenswrapper[4612]: E0227 08:25:50.233371 4612 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"587347ac26807084596cbb37f3b758c41b49fff7c59a89175fc458cc181cb3f9\": container with ID starting with 587347ac26807084596cbb37f3b758c41b49fff7c59a89175fc458cc181cb3f9 not found: ID does not exist" containerID="587347ac26807084596cbb37f3b758c41b49fff7c59a89175fc458cc181cb3f9" Feb 27 08:25:50 crc kubenswrapper[4612]: I0227 08:25:50.233413 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"587347ac26807084596cbb37f3b758c41b49fff7c59a89175fc458cc181cb3f9"} err="failed to get container status \"587347ac26807084596cbb37f3b758c41b49fff7c59a89175fc458cc181cb3f9\": rpc error: code = NotFound desc = could not find container \"587347ac26807084596cbb37f3b758c41b49fff7c59a89175fc458cc181cb3f9\": container with ID starting with 587347ac26807084596cbb37f3b758c41b49fff7c59a89175fc458cc181cb3f9 not found: ID does not exist" Feb 27 08:25:50 crc kubenswrapper[4612]: I0227 08:25:50.233443 4612 scope.go:117] "RemoveContainer" containerID="d4514d5817780b2d1fda354f54f719ed3f0b8599d16ab0a4242c57fe5c594932" Feb 27 08:25:50 crc kubenswrapper[4612]: E0227 08:25:50.233994 4612 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d4514d5817780b2d1fda354f54f719ed3f0b8599d16ab0a4242c57fe5c594932\": container with ID starting with d4514d5817780b2d1fda354f54f719ed3f0b8599d16ab0a4242c57fe5c594932 not found: ID does not exist" containerID="d4514d5817780b2d1fda354f54f719ed3f0b8599d16ab0a4242c57fe5c594932" Feb 27 08:25:50 crc kubenswrapper[4612]: I0227 08:25:50.234035 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d4514d5817780b2d1fda354f54f719ed3f0b8599d16ab0a4242c57fe5c594932"} err="failed to get container status \"d4514d5817780b2d1fda354f54f719ed3f0b8599d16ab0a4242c57fe5c594932\": rpc error: code = NotFound desc = could not find container \"d4514d5817780b2d1fda354f54f719ed3f0b8599d16ab0a4242c57fe5c594932\": container with ID starting with d4514d5817780b2d1fda354f54f719ed3f0b8599d16ab0a4242c57fe5c594932 not found: ID does not exist" Feb 27 08:25:50 crc kubenswrapper[4612]: I0227 08:25:50.234063 4612 scope.go:117] "RemoveContainer" containerID="91c4f5cea9f4458bb8b898e8be2c2b529fe980d2c5259ba859f2815a6286a055" Feb 27 08:25:50 crc kubenswrapper[4612]: E0227 08:25:50.234682 4612 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"91c4f5cea9f4458bb8b898e8be2c2b529fe980d2c5259ba859f2815a6286a055\": container with ID starting with 91c4f5cea9f4458bb8b898e8be2c2b529fe980d2c5259ba859f2815a6286a055 not found: ID does not exist" containerID="91c4f5cea9f4458bb8b898e8be2c2b529fe980d2c5259ba859f2815a6286a055" Feb 27 08:25:50 crc kubenswrapper[4612]: I0227 08:25:50.234760 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"91c4f5cea9f4458bb8b898e8be2c2b529fe980d2c5259ba859f2815a6286a055"} err="failed to get container status \"91c4f5cea9f4458bb8b898e8be2c2b529fe980d2c5259ba859f2815a6286a055\": rpc error: code = NotFound desc = could not find container \"91c4f5cea9f4458bb8b898e8be2c2b529fe980d2c5259ba859f2815a6286a055\": container with ID starting with 91c4f5cea9f4458bb8b898e8be2c2b529fe980d2c5259ba859f2815a6286a055 not found: ID does not exist" Feb 27 08:25:50 crc kubenswrapper[4612]: I0227 08:25:50.868040 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="48d52567-f8d6-4070-9739-43f9df081d20" path="/var/lib/kubelet/pods/48d52567-f8d6-4070-9739-43f9df081d20/volumes" Feb 27 08:26:00 crc kubenswrapper[4612]: I0227 08:26:00.160076 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29536346-c26b4"] Feb 27 08:26:00 crc kubenswrapper[4612]: E0227 08:26:00.161074 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="48d52567-f8d6-4070-9739-43f9df081d20" containerName="registry-server" Feb 27 08:26:00 crc kubenswrapper[4612]: I0227 08:26:00.161092 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="48d52567-f8d6-4070-9739-43f9df081d20" containerName="registry-server" Feb 27 08:26:00 crc kubenswrapper[4612]: E0227 08:26:00.161155 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="48d52567-f8d6-4070-9739-43f9df081d20" containerName="extract-content" Feb 27 08:26:00 crc kubenswrapper[4612]: I0227 08:26:00.161166 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="48d52567-f8d6-4070-9739-43f9df081d20" containerName="extract-content" Feb 27 08:26:00 crc kubenswrapper[4612]: E0227 08:26:00.161190 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="48d52567-f8d6-4070-9739-43f9df081d20" containerName="extract-utilities" Feb 27 08:26:00 crc kubenswrapper[4612]: I0227 08:26:00.161201 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="48d52567-f8d6-4070-9739-43f9df081d20" containerName="extract-utilities" Feb 27 08:26:00 crc kubenswrapper[4612]: I0227 08:26:00.161438 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="48d52567-f8d6-4070-9739-43f9df081d20" containerName="registry-server" Feb 27 08:26:00 crc kubenswrapper[4612]: I0227 08:26:00.162156 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536346-c26b4" Feb 27 08:26:00 crc kubenswrapper[4612]: I0227 08:26:00.168262 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 27 08:26:00 crc kubenswrapper[4612]: I0227 08:26:00.168874 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-5zwfx" Feb 27 08:26:00 crc kubenswrapper[4612]: I0227 08:26:00.169007 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 27 08:26:00 crc kubenswrapper[4612]: I0227 08:26:00.173351 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536346-c26b4"] Feb 27 08:26:00 crc kubenswrapper[4612]: I0227 08:26:00.297828 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ggmrc\" (UniqueName: \"kubernetes.io/projected/fba772a3-9097-4773-bff1-98ee5ebaa81c-kube-api-access-ggmrc\") pod \"auto-csr-approver-29536346-c26b4\" (UID: \"fba772a3-9097-4773-bff1-98ee5ebaa81c\") " pod="openshift-infra/auto-csr-approver-29536346-c26b4" Feb 27 08:26:00 crc kubenswrapper[4612]: I0227 08:26:00.399741 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ggmrc\" (UniqueName: \"kubernetes.io/projected/fba772a3-9097-4773-bff1-98ee5ebaa81c-kube-api-access-ggmrc\") pod \"auto-csr-approver-29536346-c26b4\" (UID: \"fba772a3-9097-4773-bff1-98ee5ebaa81c\") " pod="openshift-infra/auto-csr-approver-29536346-c26b4" Feb 27 08:26:00 crc kubenswrapper[4612]: I0227 08:26:00.424347 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ggmrc\" (UniqueName: \"kubernetes.io/projected/fba772a3-9097-4773-bff1-98ee5ebaa81c-kube-api-access-ggmrc\") pod \"auto-csr-approver-29536346-c26b4\" (UID: \"fba772a3-9097-4773-bff1-98ee5ebaa81c\") " pod="openshift-infra/auto-csr-approver-29536346-c26b4" Feb 27 08:26:00 crc kubenswrapper[4612]: I0227 08:26:00.485788 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536346-c26b4" Feb 27 08:26:00 crc kubenswrapper[4612]: I0227 08:26:00.937251 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536346-c26b4"] Feb 27 08:26:01 crc kubenswrapper[4612]: I0227 08:26:01.242661 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536346-c26b4" event={"ID":"fba772a3-9097-4773-bff1-98ee5ebaa81c","Type":"ContainerStarted","Data":"0a53ee378a048146686cb54fcb072f265247244dcf5448b2a62c1ec99c6ca6ea"} Feb 27 08:26:02 crc kubenswrapper[4612]: I0227 08:26:02.260388 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536346-c26b4" event={"ID":"fba772a3-9097-4773-bff1-98ee5ebaa81c","Type":"ContainerStarted","Data":"455208a77a54fb810ab226d17541bdb62481523d5d2a486b7aff76c2d10e2427"} Feb 27 08:26:02 crc kubenswrapper[4612]: I0227 08:26:02.280131 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29536346-c26b4" podStartSLOduration=1.481739538 podStartE2EDuration="2.280111361s" podCreationTimestamp="2026-02-27 08:26:00 +0000 UTC" firstStartedPulling="2026-02-27 08:26:00.948077202 +0000 UTC m=+2218.802007200" lastFinishedPulling="2026-02-27 08:26:01.746449015 +0000 UTC m=+2219.600379023" observedRunningTime="2026-02-27 08:26:02.279416921 +0000 UTC m=+2220.133346919" watchObservedRunningTime="2026-02-27 08:26:02.280111361 +0000 UTC m=+2220.134041369" Feb 27 08:26:03 crc kubenswrapper[4612]: I0227 08:26:03.270552 4612 generic.go:334] "Generic (PLEG): container finished" podID="fba772a3-9097-4773-bff1-98ee5ebaa81c" containerID="455208a77a54fb810ab226d17541bdb62481523d5d2a486b7aff76c2d10e2427" exitCode=0 Feb 27 08:26:03 crc kubenswrapper[4612]: I0227 08:26:03.270639 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536346-c26b4" event={"ID":"fba772a3-9097-4773-bff1-98ee5ebaa81c","Type":"ContainerDied","Data":"455208a77a54fb810ab226d17541bdb62481523d5d2a486b7aff76c2d10e2427"} Feb 27 08:26:04 crc kubenswrapper[4612]: I0227 08:26:04.592287 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536346-c26b4" Feb 27 08:26:04 crc kubenswrapper[4612]: I0227 08:26:04.689864 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ggmrc\" (UniqueName: \"kubernetes.io/projected/fba772a3-9097-4773-bff1-98ee5ebaa81c-kube-api-access-ggmrc\") pod \"fba772a3-9097-4773-bff1-98ee5ebaa81c\" (UID: \"fba772a3-9097-4773-bff1-98ee5ebaa81c\") " Feb 27 08:26:04 crc kubenswrapper[4612]: I0227 08:26:04.702002 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fba772a3-9097-4773-bff1-98ee5ebaa81c-kube-api-access-ggmrc" (OuterVolumeSpecName: "kube-api-access-ggmrc") pod "fba772a3-9097-4773-bff1-98ee5ebaa81c" (UID: "fba772a3-9097-4773-bff1-98ee5ebaa81c"). InnerVolumeSpecName "kube-api-access-ggmrc". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 08:26:04 crc kubenswrapper[4612]: I0227 08:26:04.791855 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ggmrc\" (UniqueName: \"kubernetes.io/projected/fba772a3-9097-4773-bff1-98ee5ebaa81c-kube-api-access-ggmrc\") on node \"crc\" DevicePath \"\"" Feb 27 08:26:05 crc kubenswrapper[4612]: I0227 08:26:05.329679 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536346-c26b4" event={"ID":"fba772a3-9097-4773-bff1-98ee5ebaa81c","Type":"ContainerDied","Data":"0a53ee378a048146686cb54fcb072f265247244dcf5448b2a62c1ec99c6ca6ea"} Feb 27 08:26:05 crc kubenswrapper[4612]: I0227 08:26:05.329747 4612 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0a53ee378a048146686cb54fcb072f265247244dcf5448b2a62c1ec99c6ca6ea" Feb 27 08:26:05 crc kubenswrapper[4612]: I0227 08:26:05.329838 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536346-c26b4" Feb 27 08:26:05 crc kubenswrapper[4612]: I0227 08:26:05.371631 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29536340-2zmwk"] Feb 27 08:26:05 crc kubenswrapper[4612]: I0227 08:26:05.381840 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29536340-2zmwk"] Feb 27 08:26:06 crc kubenswrapper[4612]: I0227 08:26:06.868443 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ec08bd6b-d9ec-4b86-8397-9e1f709332ed" path="/var/lib/kubelet/pods/ec08bd6b-d9ec-4b86-8397-9e1f709332ed/volumes" Feb 27 08:26:16 crc kubenswrapper[4612]: I0227 08:26:16.027548 4612 patch_prober.go:28] interesting pod/machine-config-daemon-924vb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 27 08:26:16 crc kubenswrapper[4612]: I0227 08:26:16.028211 4612 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 27 08:26:17 crc kubenswrapper[4612]: I0227 08:26:17.473559 4612 scope.go:117] "RemoveContainer" containerID="66a4c28074c462834ed0ba2b9b203d2add0612d7747efadf8852535b8166500b" Feb 27 08:26:38 crc kubenswrapper[4612]: I0227 08:26:38.675767 4612 generic.go:334] "Generic (PLEG): container finished" podID="a213a1c0-9409-487d-834a-a60e11fd8a66" containerID="7181ed7bee7378f3a23dd7d8df1c2f43ecf310bf461cd3323b3727f5cb458b5f" exitCode=0 Feb 27 08:26:38 crc kubenswrapper[4612]: I0227 08:26:38.676171 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-x52ht" event={"ID":"a213a1c0-9409-487d-834a-a60e11fd8a66","Type":"ContainerDied","Data":"7181ed7bee7378f3a23dd7d8df1c2f43ecf310bf461cd3323b3727f5cb458b5f"} Feb 27 08:26:40 crc kubenswrapper[4612]: I0227 08:26:40.112331 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-x52ht" Feb 27 08:26:40 crc kubenswrapper[4612]: I0227 08:26:40.242486 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a213a1c0-9409-487d-834a-a60e11fd8a66-inventory\") pod \"a213a1c0-9409-487d-834a-a60e11fd8a66\" (UID: \"a213a1c0-9409-487d-834a-a60e11fd8a66\") " Feb 27 08:26:40 crc kubenswrapper[4612]: I0227 08:26:40.242818 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/a213a1c0-9409-487d-834a-a60e11fd8a66-neutron-ovn-metadata-agent-neutron-config-0\") pod \"a213a1c0-9409-487d-834a-a60e11fd8a66\" (UID: \"a213a1c0-9409-487d-834a-a60e11fd8a66\") " Feb 27 08:26:40 crc kubenswrapper[4612]: I0227 08:26:40.243110 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9f8tj\" (UniqueName: \"kubernetes.io/projected/a213a1c0-9409-487d-834a-a60e11fd8a66-kube-api-access-9f8tj\") pod \"a213a1c0-9409-487d-834a-a60e11fd8a66\" (UID: \"a213a1c0-9409-487d-834a-a60e11fd8a66\") " Feb 27 08:26:40 crc kubenswrapper[4612]: I0227 08:26:40.243165 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/a213a1c0-9409-487d-834a-a60e11fd8a66-nova-metadata-neutron-config-0\") pod \"a213a1c0-9409-487d-834a-a60e11fd8a66\" (UID: \"a213a1c0-9409-487d-834a-a60e11fd8a66\") " Feb 27 08:26:40 crc kubenswrapper[4612]: I0227 08:26:40.243324 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/a213a1c0-9409-487d-834a-a60e11fd8a66-ssh-key-openstack-edpm-ipam\") pod \"a213a1c0-9409-487d-834a-a60e11fd8a66\" (UID: \"a213a1c0-9409-487d-834a-a60e11fd8a66\") " Feb 27 08:26:40 crc kubenswrapper[4612]: I0227 08:26:40.243372 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a213a1c0-9409-487d-834a-a60e11fd8a66-neutron-metadata-combined-ca-bundle\") pod \"a213a1c0-9409-487d-834a-a60e11fd8a66\" (UID: \"a213a1c0-9409-487d-834a-a60e11fd8a66\") " Feb 27 08:26:40 crc kubenswrapper[4612]: I0227 08:26:40.268645 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a213a1c0-9409-487d-834a-a60e11fd8a66-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "a213a1c0-9409-487d-834a-a60e11fd8a66" (UID: "a213a1c0-9409-487d-834a-a60e11fd8a66"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:26:40 crc kubenswrapper[4612]: I0227 08:26:40.279633 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a213a1c0-9409-487d-834a-a60e11fd8a66-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "a213a1c0-9409-487d-834a-a60e11fd8a66" (UID: "a213a1c0-9409-487d-834a-a60e11fd8a66"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:26:40 crc kubenswrapper[4612]: I0227 08:26:40.279814 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a213a1c0-9409-487d-834a-a60e11fd8a66-neutron-ovn-metadata-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-ovn-metadata-agent-neutron-config-0") pod "a213a1c0-9409-487d-834a-a60e11fd8a66" (UID: "a213a1c0-9409-487d-834a-a60e11fd8a66"). InnerVolumeSpecName "neutron-ovn-metadata-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:26:40 crc kubenswrapper[4612]: I0227 08:26:40.279851 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a213a1c0-9409-487d-834a-a60e11fd8a66-inventory" (OuterVolumeSpecName: "inventory") pod "a213a1c0-9409-487d-834a-a60e11fd8a66" (UID: "a213a1c0-9409-487d-834a-a60e11fd8a66"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:26:40 crc kubenswrapper[4612]: I0227 08:26:40.279981 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a213a1c0-9409-487d-834a-a60e11fd8a66-nova-metadata-neutron-config-0" (OuterVolumeSpecName: "nova-metadata-neutron-config-0") pod "a213a1c0-9409-487d-834a-a60e11fd8a66" (UID: "a213a1c0-9409-487d-834a-a60e11fd8a66"). InnerVolumeSpecName "nova-metadata-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:26:40 crc kubenswrapper[4612]: I0227 08:26:40.281872 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a213a1c0-9409-487d-834a-a60e11fd8a66-kube-api-access-9f8tj" (OuterVolumeSpecName: "kube-api-access-9f8tj") pod "a213a1c0-9409-487d-834a-a60e11fd8a66" (UID: "a213a1c0-9409-487d-834a-a60e11fd8a66"). InnerVolumeSpecName "kube-api-access-9f8tj". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 08:26:40 crc kubenswrapper[4612]: I0227 08:26:40.345353 4612 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/a213a1c0-9409-487d-834a-a60e11fd8a66-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 27 08:26:40 crc kubenswrapper[4612]: I0227 08:26:40.345393 4612 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a213a1c0-9409-487d-834a-a60e11fd8a66-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 08:26:40 crc kubenswrapper[4612]: I0227 08:26:40.345405 4612 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a213a1c0-9409-487d-834a-a60e11fd8a66-inventory\") on node \"crc\" DevicePath \"\"" Feb 27 08:26:40 crc kubenswrapper[4612]: I0227 08:26:40.345415 4612 reconciler_common.go:293] "Volume detached for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/a213a1c0-9409-487d-834a-a60e11fd8a66-neutron-ovn-metadata-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Feb 27 08:26:40 crc kubenswrapper[4612]: I0227 08:26:40.345426 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9f8tj\" (UniqueName: \"kubernetes.io/projected/a213a1c0-9409-487d-834a-a60e11fd8a66-kube-api-access-9f8tj\") on node \"crc\" DevicePath \"\"" Feb 27 08:26:40 crc kubenswrapper[4612]: I0227 08:26:40.345434 4612 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/a213a1c0-9409-487d-834a-a60e11fd8a66-nova-metadata-neutron-config-0\") on node \"crc\" DevicePath \"\"" Feb 27 08:26:40 crc kubenswrapper[4612]: I0227 08:26:40.699102 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-x52ht" event={"ID":"a213a1c0-9409-487d-834a-a60e11fd8a66","Type":"ContainerDied","Data":"101802c5e706e590e09c02cc021707d3b1368fdee83e9faa0741b52c52516775"} Feb 27 08:26:40 crc kubenswrapper[4612]: I0227 08:26:40.699174 4612 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="101802c5e706e590e09c02cc021707d3b1368fdee83e9faa0741b52c52516775" Feb 27 08:26:40 crc kubenswrapper[4612]: I0227 08:26:40.699235 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-x52ht" Feb 27 08:26:40 crc kubenswrapper[4612]: I0227 08:26:40.803983 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-27c56"] Feb 27 08:26:40 crc kubenswrapper[4612]: E0227 08:26:40.804319 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fba772a3-9097-4773-bff1-98ee5ebaa81c" containerName="oc" Feb 27 08:26:40 crc kubenswrapper[4612]: I0227 08:26:40.804335 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="fba772a3-9097-4773-bff1-98ee5ebaa81c" containerName="oc" Feb 27 08:26:40 crc kubenswrapper[4612]: E0227 08:26:40.804355 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a213a1c0-9409-487d-834a-a60e11fd8a66" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Feb 27 08:26:40 crc kubenswrapper[4612]: I0227 08:26:40.804362 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="a213a1c0-9409-487d-834a-a60e11fd8a66" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Feb 27 08:26:40 crc kubenswrapper[4612]: I0227 08:26:40.804539 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="a213a1c0-9409-487d-834a-a60e11fd8a66" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Feb 27 08:26:40 crc kubenswrapper[4612]: I0227 08:26:40.804563 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="fba772a3-9097-4773-bff1-98ee5ebaa81c" containerName="oc" Feb 27 08:26:40 crc kubenswrapper[4612]: I0227 08:26:40.805427 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-27c56" Feb 27 08:26:40 crc kubenswrapper[4612]: I0227 08:26:40.809589 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 27 08:26:40 crc kubenswrapper[4612]: I0227 08:26:40.809610 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"libvirt-secret" Feb 27 08:26:40 crc kubenswrapper[4612]: I0227 08:26:40.809858 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 27 08:26:40 crc kubenswrapper[4612]: I0227 08:26:40.810840 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-2pwrb" Feb 27 08:26:40 crc kubenswrapper[4612]: I0227 08:26:40.811011 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 27 08:26:40 crc kubenswrapper[4612]: I0227 08:26:40.819358 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-27c56"] Feb 27 08:26:40 crc kubenswrapper[4612]: I0227 08:26:40.955842 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f4308354-0bda-4fb3-976f-cc2fa471dcb4-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-27c56\" (UID: \"f4308354-0bda-4fb3-976f-cc2fa471dcb4\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-27c56" Feb 27 08:26:40 crc kubenswrapper[4612]: I0227 08:26:40.955888 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f4308354-0bda-4fb3-976f-cc2fa471dcb4-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-27c56\" (UID: \"f4308354-0bda-4fb3-976f-cc2fa471dcb4\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-27c56" Feb 27 08:26:40 crc kubenswrapper[4612]: I0227 08:26:40.956777 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/f4308354-0bda-4fb3-976f-cc2fa471dcb4-ssh-key-openstack-edpm-ipam\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-27c56\" (UID: \"f4308354-0bda-4fb3-976f-cc2fa471dcb4\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-27c56" Feb 27 08:26:40 crc kubenswrapper[4612]: I0227 08:26:40.956852 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g8qlt\" (UniqueName: \"kubernetes.io/projected/f4308354-0bda-4fb3-976f-cc2fa471dcb4-kube-api-access-g8qlt\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-27c56\" (UID: \"f4308354-0bda-4fb3-976f-cc2fa471dcb4\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-27c56" Feb 27 08:26:40 crc kubenswrapper[4612]: I0227 08:26:40.956903 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/f4308354-0bda-4fb3-976f-cc2fa471dcb4-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-27c56\" (UID: \"f4308354-0bda-4fb3-976f-cc2fa471dcb4\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-27c56" Feb 27 08:26:41 crc kubenswrapper[4612]: I0227 08:26:41.059586 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/f4308354-0bda-4fb3-976f-cc2fa471dcb4-ssh-key-openstack-edpm-ipam\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-27c56\" (UID: \"f4308354-0bda-4fb3-976f-cc2fa471dcb4\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-27c56" Feb 27 08:26:41 crc kubenswrapper[4612]: I0227 08:26:41.059631 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g8qlt\" (UniqueName: \"kubernetes.io/projected/f4308354-0bda-4fb3-976f-cc2fa471dcb4-kube-api-access-g8qlt\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-27c56\" (UID: \"f4308354-0bda-4fb3-976f-cc2fa471dcb4\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-27c56" Feb 27 08:26:41 crc kubenswrapper[4612]: I0227 08:26:41.059652 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/f4308354-0bda-4fb3-976f-cc2fa471dcb4-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-27c56\" (UID: \"f4308354-0bda-4fb3-976f-cc2fa471dcb4\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-27c56" Feb 27 08:26:41 crc kubenswrapper[4612]: I0227 08:26:41.059764 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f4308354-0bda-4fb3-976f-cc2fa471dcb4-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-27c56\" (UID: \"f4308354-0bda-4fb3-976f-cc2fa471dcb4\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-27c56" Feb 27 08:26:41 crc kubenswrapper[4612]: I0227 08:26:41.059784 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f4308354-0bda-4fb3-976f-cc2fa471dcb4-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-27c56\" (UID: \"f4308354-0bda-4fb3-976f-cc2fa471dcb4\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-27c56" Feb 27 08:26:41 crc kubenswrapper[4612]: I0227 08:26:41.070460 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/f4308354-0bda-4fb3-976f-cc2fa471dcb4-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-27c56\" (UID: \"f4308354-0bda-4fb3-976f-cc2fa471dcb4\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-27c56" Feb 27 08:26:41 crc kubenswrapper[4612]: I0227 08:26:41.070516 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f4308354-0bda-4fb3-976f-cc2fa471dcb4-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-27c56\" (UID: \"f4308354-0bda-4fb3-976f-cc2fa471dcb4\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-27c56" Feb 27 08:26:41 crc kubenswrapper[4612]: I0227 08:26:41.071476 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f4308354-0bda-4fb3-976f-cc2fa471dcb4-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-27c56\" (UID: \"f4308354-0bda-4fb3-976f-cc2fa471dcb4\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-27c56" Feb 27 08:26:41 crc kubenswrapper[4612]: I0227 08:26:41.072105 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/f4308354-0bda-4fb3-976f-cc2fa471dcb4-ssh-key-openstack-edpm-ipam\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-27c56\" (UID: \"f4308354-0bda-4fb3-976f-cc2fa471dcb4\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-27c56" Feb 27 08:26:41 crc kubenswrapper[4612]: I0227 08:26:41.085099 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g8qlt\" (UniqueName: \"kubernetes.io/projected/f4308354-0bda-4fb3-976f-cc2fa471dcb4-kube-api-access-g8qlt\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-27c56\" (UID: \"f4308354-0bda-4fb3-976f-cc2fa471dcb4\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-27c56" Feb 27 08:26:41 crc kubenswrapper[4612]: I0227 08:26:41.127585 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-27c56" Feb 27 08:26:41 crc kubenswrapper[4612]: I0227 08:26:41.652004 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-27c56"] Feb 27 08:26:41 crc kubenswrapper[4612]: I0227 08:26:41.714907 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-27c56" event={"ID":"f4308354-0bda-4fb3-976f-cc2fa471dcb4","Type":"ContainerStarted","Data":"12344af5441dc29acf62d0da60c4a7b403628677e00c02b5a28f40fef7a60c6b"} Feb 27 08:26:43 crc kubenswrapper[4612]: I0227 08:26:43.733872 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-27c56" event={"ID":"f4308354-0bda-4fb3-976f-cc2fa471dcb4","Type":"ContainerStarted","Data":"8e6f6e66e01c1e16a3b54d7d03f2cfdf920a64f28ebd1cf42df58a1e0143a429"} Feb 27 08:26:43 crc kubenswrapper[4612]: I0227 08:26:43.752878 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-27c56" podStartSLOduration=3.022443081 podStartE2EDuration="3.752853264s" podCreationTimestamp="2026-02-27 08:26:40 +0000 UTC" firstStartedPulling="2026-02-27 08:26:41.659218868 +0000 UTC m=+2259.513148866" lastFinishedPulling="2026-02-27 08:26:42.389629051 +0000 UTC m=+2260.243559049" observedRunningTime="2026-02-27 08:26:43.748353364 +0000 UTC m=+2261.602283372" watchObservedRunningTime="2026-02-27 08:26:43.752853264 +0000 UTC m=+2261.606783262" Feb 27 08:26:46 crc kubenswrapper[4612]: I0227 08:26:46.027393 4612 patch_prober.go:28] interesting pod/machine-config-daemon-924vb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 27 08:26:46 crc kubenswrapper[4612]: I0227 08:26:46.028021 4612 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 27 08:26:46 crc kubenswrapper[4612]: I0227 08:26:46.028084 4612 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-924vb" Feb 27 08:26:46 crc kubenswrapper[4612]: I0227 08:26:46.028916 4612 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"76c2de0055c53fc536f48b64de557ea4f7b56a2de457f8f1b8708b99c7c612cb"} pod="openshift-machine-config-operator/machine-config-daemon-924vb" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 27 08:26:46 crc kubenswrapper[4612]: I0227 08:26:46.028986 4612 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" containerName="machine-config-daemon" containerID="cri-o://76c2de0055c53fc536f48b64de557ea4f7b56a2de457f8f1b8708b99c7c612cb" gracePeriod=600 Feb 27 08:26:46 crc kubenswrapper[4612]: I0227 08:26:46.778656 4612 generic.go:334] "Generic (PLEG): container finished" podID="24599373-7adc-4a1b-8bb4-797bf726f43d" containerID="76c2de0055c53fc536f48b64de557ea4f7b56a2de457f8f1b8708b99c7c612cb" exitCode=0 Feb 27 08:26:46 crc kubenswrapper[4612]: I0227 08:26:46.778747 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-924vb" event={"ID":"24599373-7adc-4a1b-8bb4-797bf726f43d","Type":"ContainerDied","Data":"76c2de0055c53fc536f48b64de557ea4f7b56a2de457f8f1b8708b99c7c612cb"} Feb 27 08:26:46 crc kubenswrapper[4612]: I0227 08:26:46.778818 4612 scope.go:117] "RemoveContainer" containerID="f80ae3b7ec2dce55e073a92197c42ea49b624e8fd3575dd785e408adfa1598e5" Feb 27 08:26:46 crc kubenswrapper[4612]: E0227 08:26:46.834533 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-924vb_openshift-machine-config-operator(24599373-7adc-4a1b-8bb4-797bf726f43d)\"" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" Feb 27 08:26:47 crc kubenswrapper[4612]: I0227 08:26:47.788862 4612 scope.go:117] "RemoveContainer" containerID="76c2de0055c53fc536f48b64de557ea4f7b56a2de457f8f1b8708b99c7c612cb" Feb 27 08:26:47 crc kubenswrapper[4612]: E0227 08:26:47.789299 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-924vb_openshift-machine-config-operator(24599373-7adc-4a1b-8bb4-797bf726f43d)\"" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" Feb 27 08:27:00 crc kubenswrapper[4612]: I0227 08:27:00.853228 4612 scope.go:117] "RemoveContainer" containerID="76c2de0055c53fc536f48b64de557ea4f7b56a2de457f8f1b8708b99c7c612cb" Feb 27 08:27:00 crc kubenswrapper[4612]: E0227 08:27:00.854095 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-924vb_openshift-machine-config-operator(24599373-7adc-4a1b-8bb4-797bf726f43d)\"" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" Feb 27 08:27:14 crc kubenswrapper[4612]: I0227 08:27:14.853897 4612 scope.go:117] "RemoveContainer" containerID="76c2de0055c53fc536f48b64de557ea4f7b56a2de457f8f1b8708b99c7c612cb" Feb 27 08:27:14 crc kubenswrapper[4612]: E0227 08:27:14.854762 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-924vb_openshift-machine-config-operator(24599373-7adc-4a1b-8bb4-797bf726f43d)\"" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" Feb 27 08:27:28 crc kubenswrapper[4612]: I0227 08:27:28.853047 4612 scope.go:117] "RemoveContainer" containerID="76c2de0055c53fc536f48b64de557ea4f7b56a2de457f8f1b8708b99c7c612cb" Feb 27 08:27:28 crc kubenswrapper[4612]: E0227 08:27:28.853723 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-924vb_openshift-machine-config-operator(24599373-7adc-4a1b-8bb4-797bf726f43d)\"" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" Feb 27 08:27:41 crc kubenswrapper[4612]: I0227 08:27:41.853042 4612 scope.go:117] "RemoveContainer" containerID="76c2de0055c53fc536f48b64de557ea4f7b56a2de457f8f1b8708b99c7c612cb" Feb 27 08:27:41 crc kubenswrapper[4612]: E0227 08:27:41.853785 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-924vb_openshift-machine-config-operator(24599373-7adc-4a1b-8bb4-797bf726f43d)\"" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" Feb 27 08:27:54 crc kubenswrapper[4612]: I0227 08:27:54.853879 4612 scope.go:117] "RemoveContainer" containerID="76c2de0055c53fc536f48b64de557ea4f7b56a2de457f8f1b8708b99c7c612cb" Feb 27 08:27:54 crc kubenswrapper[4612]: E0227 08:27:54.855054 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-924vb_openshift-machine-config-operator(24599373-7adc-4a1b-8bb4-797bf726f43d)\"" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" Feb 27 08:28:00 crc kubenswrapper[4612]: I0227 08:28:00.160807 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29536348-gmk9c"] Feb 27 08:28:00 crc kubenswrapper[4612]: I0227 08:28:00.162793 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536348-gmk9c" Feb 27 08:28:00 crc kubenswrapper[4612]: I0227 08:28:00.165782 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-5zwfx" Feb 27 08:28:00 crc kubenswrapper[4612]: I0227 08:28:00.166528 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 27 08:28:00 crc kubenswrapper[4612]: I0227 08:28:00.171334 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536348-gmk9c"] Feb 27 08:28:00 crc kubenswrapper[4612]: I0227 08:28:00.174318 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 27 08:28:00 crc kubenswrapper[4612]: I0227 08:28:00.274253 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w247k\" (UniqueName: \"kubernetes.io/projected/28295adb-7038-43da-a687-744152692bd2-kube-api-access-w247k\") pod \"auto-csr-approver-29536348-gmk9c\" (UID: \"28295adb-7038-43da-a687-744152692bd2\") " pod="openshift-infra/auto-csr-approver-29536348-gmk9c" Feb 27 08:28:00 crc kubenswrapper[4612]: I0227 08:28:00.376316 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w247k\" (UniqueName: \"kubernetes.io/projected/28295adb-7038-43da-a687-744152692bd2-kube-api-access-w247k\") pod \"auto-csr-approver-29536348-gmk9c\" (UID: \"28295adb-7038-43da-a687-744152692bd2\") " pod="openshift-infra/auto-csr-approver-29536348-gmk9c" Feb 27 08:28:00 crc kubenswrapper[4612]: I0227 08:28:00.408625 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w247k\" (UniqueName: \"kubernetes.io/projected/28295adb-7038-43da-a687-744152692bd2-kube-api-access-w247k\") pod \"auto-csr-approver-29536348-gmk9c\" (UID: \"28295adb-7038-43da-a687-744152692bd2\") " pod="openshift-infra/auto-csr-approver-29536348-gmk9c" Feb 27 08:28:00 crc kubenswrapper[4612]: I0227 08:28:00.486751 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536348-gmk9c" Feb 27 08:28:00 crc kubenswrapper[4612]: I0227 08:28:00.983531 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536348-gmk9c"] Feb 27 08:28:01 crc kubenswrapper[4612]: I0227 08:28:01.463253 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536348-gmk9c" event={"ID":"28295adb-7038-43da-a687-744152692bd2","Type":"ContainerStarted","Data":"acbd72e6fc0a3221c89107d60bf116ebce9c80db4efff71fd165d453e14809ee"} Feb 27 08:28:03 crc kubenswrapper[4612]: I0227 08:28:03.486173 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536348-gmk9c" event={"ID":"28295adb-7038-43da-a687-744152692bd2","Type":"ContainerStarted","Data":"3521c44e78318294619f5e8a6c30f85dfb9c42064e432e1594c779a464a39199"} Feb 27 08:28:03 crc kubenswrapper[4612]: I0227 08:28:03.503926 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29536348-gmk9c" podStartSLOduration=1.724171887 podStartE2EDuration="3.503904896s" podCreationTimestamp="2026-02-27 08:28:00 +0000 UTC" firstStartedPulling="2026-02-27 08:28:00.991232301 +0000 UTC m=+2338.845162299" lastFinishedPulling="2026-02-27 08:28:02.77096527 +0000 UTC m=+2340.624895308" observedRunningTime="2026-02-27 08:28:03.50126729 +0000 UTC m=+2341.355197298" watchObservedRunningTime="2026-02-27 08:28:03.503904896 +0000 UTC m=+2341.357834904" Feb 27 08:28:04 crc kubenswrapper[4612]: I0227 08:28:04.497777 4612 generic.go:334] "Generic (PLEG): container finished" podID="28295adb-7038-43da-a687-744152692bd2" containerID="3521c44e78318294619f5e8a6c30f85dfb9c42064e432e1594c779a464a39199" exitCode=0 Feb 27 08:28:04 crc kubenswrapper[4612]: I0227 08:28:04.497970 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536348-gmk9c" event={"ID":"28295adb-7038-43da-a687-744152692bd2","Type":"ContainerDied","Data":"3521c44e78318294619f5e8a6c30f85dfb9c42064e432e1594c779a464a39199"} Feb 27 08:28:05 crc kubenswrapper[4612]: I0227 08:28:05.875310 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536348-gmk9c" Feb 27 08:28:05 crc kubenswrapper[4612]: I0227 08:28:05.934845 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w247k\" (UniqueName: \"kubernetes.io/projected/28295adb-7038-43da-a687-744152692bd2-kube-api-access-w247k\") pod \"28295adb-7038-43da-a687-744152692bd2\" (UID: \"28295adb-7038-43da-a687-744152692bd2\") " Feb 27 08:28:05 crc kubenswrapper[4612]: I0227 08:28:05.969180 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/28295adb-7038-43da-a687-744152692bd2-kube-api-access-w247k" (OuterVolumeSpecName: "kube-api-access-w247k") pod "28295adb-7038-43da-a687-744152692bd2" (UID: "28295adb-7038-43da-a687-744152692bd2"). InnerVolumeSpecName "kube-api-access-w247k". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 08:28:05 crc kubenswrapper[4612]: I0227 08:28:05.983637 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29536342-c9hbc"] Feb 27 08:28:05 crc kubenswrapper[4612]: I0227 08:28:05.993865 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29536342-c9hbc"] Feb 27 08:28:06 crc kubenswrapper[4612]: I0227 08:28:06.038483 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w247k\" (UniqueName: \"kubernetes.io/projected/28295adb-7038-43da-a687-744152692bd2-kube-api-access-w247k\") on node \"crc\" DevicePath \"\"" Feb 27 08:28:06 crc kubenswrapper[4612]: I0227 08:28:06.515583 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536348-gmk9c" event={"ID":"28295adb-7038-43da-a687-744152692bd2","Type":"ContainerDied","Data":"acbd72e6fc0a3221c89107d60bf116ebce9c80db4efff71fd165d453e14809ee"} Feb 27 08:28:06 crc kubenswrapper[4612]: I0227 08:28:06.515619 4612 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="acbd72e6fc0a3221c89107d60bf116ebce9c80db4efff71fd165d453e14809ee" Feb 27 08:28:06 crc kubenswrapper[4612]: I0227 08:28:06.515635 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536348-gmk9c" Feb 27 08:28:06 crc kubenswrapper[4612]: I0227 08:28:06.853923 4612 scope.go:117] "RemoveContainer" containerID="76c2de0055c53fc536f48b64de557ea4f7b56a2de457f8f1b8708b99c7c612cb" Feb 27 08:28:06 crc kubenswrapper[4612]: E0227 08:28:06.854430 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-924vb_openshift-machine-config-operator(24599373-7adc-4a1b-8bb4-797bf726f43d)\"" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" Feb 27 08:28:06 crc kubenswrapper[4612]: I0227 08:28:06.864926 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ab70fa78-06b2-4c10-81ea-18287a1d14b3" path="/var/lib/kubelet/pods/ab70fa78-06b2-4c10-81ea-18287a1d14b3/volumes" Feb 27 08:28:17 crc kubenswrapper[4612]: I0227 08:28:17.598198 4612 scope.go:117] "RemoveContainer" containerID="9f368ab9892602f19068a88bb54bed7fbc6b535ce9a9f7df140682d7180e0498" Feb 27 08:28:19 crc kubenswrapper[4612]: I0227 08:28:19.853315 4612 scope.go:117] "RemoveContainer" containerID="76c2de0055c53fc536f48b64de557ea4f7b56a2de457f8f1b8708b99c7c612cb" Feb 27 08:28:19 crc kubenswrapper[4612]: E0227 08:28:19.854021 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-924vb_openshift-machine-config-operator(24599373-7adc-4a1b-8bb4-797bf726f43d)\"" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" Feb 27 08:28:33 crc kubenswrapper[4612]: I0227 08:28:33.853265 4612 scope.go:117] "RemoveContainer" containerID="76c2de0055c53fc536f48b64de557ea4f7b56a2de457f8f1b8708b99c7c612cb" Feb 27 08:28:33 crc kubenswrapper[4612]: E0227 08:28:33.854319 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-924vb_openshift-machine-config-operator(24599373-7adc-4a1b-8bb4-797bf726f43d)\"" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" Feb 27 08:28:46 crc kubenswrapper[4612]: I0227 08:28:46.854072 4612 scope.go:117] "RemoveContainer" containerID="76c2de0055c53fc536f48b64de557ea4f7b56a2de457f8f1b8708b99c7c612cb" Feb 27 08:28:46 crc kubenswrapper[4612]: E0227 08:28:46.855206 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-924vb_openshift-machine-config-operator(24599373-7adc-4a1b-8bb4-797bf726f43d)\"" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" Feb 27 08:29:01 crc kubenswrapper[4612]: I0227 08:29:01.854016 4612 scope.go:117] "RemoveContainer" containerID="76c2de0055c53fc536f48b64de557ea4f7b56a2de457f8f1b8708b99c7c612cb" Feb 27 08:29:01 crc kubenswrapper[4612]: E0227 08:29:01.854953 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-924vb_openshift-machine-config-operator(24599373-7adc-4a1b-8bb4-797bf726f43d)\"" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" Feb 27 08:29:15 crc kubenswrapper[4612]: I0227 08:29:15.853621 4612 scope.go:117] "RemoveContainer" containerID="76c2de0055c53fc536f48b64de557ea4f7b56a2de457f8f1b8708b99c7c612cb" Feb 27 08:29:15 crc kubenswrapper[4612]: E0227 08:29:15.854520 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-924vb_openshift-machine-config-operator(24599373-7adc-4a1b-8bb4-797bf726f43d)\"" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" Feb 27 08:29:30 crc kubenswrapper[4612]: I0227 08:29:30.853783 4612 scope.go:117] "RemoveContainer" containerID="76c2de0055c53fc536f48b64de557ea4f7b56a2de457f8f1b8708b99c7c612cb" Feb 27 08:29:30 crc kubenswrapper[4612]: E0227 08:29:30.855526 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-924vb_openshift-machine-config-operator(24599373-7adc-4a1b-8bb4-797bf726f43d)\"" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" Feb 27 08:29:44 crc kubenswrapper[4612]: I0227 08:29:44.857477 4612 scope.go:117] "RemoveContainer" containerID="76c2de0055c53fc536f48b64de557ea4f7b56a2de457f8f1b8708b99c7c612cb" Feb 27 08:29:44 crc kubenswrapper[4612]: E0227 08:29:44.859873 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-924vb_openshift-machine-config-operator(24599373-7adc-4a1b-8bb4-797bf726f43d)\"" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" Feb 27 08:29:57 crc kubenswrapper[4612]: I0227 08:29:57.853053 4612 scope.go:117] "RemoveContainer" containerID="76c2de0055c53fc536f48b64de557ea4f7b56a2de457f8f1b8708b99c7c612cb" Feb 27 08:29:57 crc kubenswrapper[4612]: E0227 08:29:57.853847 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-924vb_openshift-machine-config-operator(24599373-7adc-4a1b-8bb4-797bf726f43d)\"" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" Feb 27 08:30:00 crc kubenswrapper[4612]: I0227 08:30:00.155999 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29536350-4xptg"] Feb 27 08:30:00 crc kubenswrapper[4612]: E0227 08:30:00.156855 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="28295adb-7038-43da-a687-744152692bd2" containerName="oc" Feb 27 08:30:00 crc kubenswrapper[4612]: I0227 08:30:00.156888 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="28295adb-7038-43da-a687-744152692bd2" containerName="oc" Feb 27 08:30:00 crc kubenswrapper[4612]: I0227 08:30:00.157173 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="28295adb-7038-43da-a687-744152692bd2" containerName="oc" Feb 27 08:30:00 crc kubenswrapper[4612]: I0227 08:30:00.158138 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29536350-4xptg" Feb 27 08:30:00 crc kubenswrapper[4612]: I0227 08:30:00.160193 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Feb 27 08:30:00 crc kubenswrapper[4612]: I0227 08:30:00.161936 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Feb 27 08:30:00 crc kubenswrapper[4612]: I0227 08:30:00.169891 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6ac60cd4-1e56-4ad4-8ba9-50bc36f8f211-config-volume\") pod \"collect-profiles-29536350-4xptg\" (UID: \"6ac60cd4-1e56-4ad4-8ba9-50bc36f8f211\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29536350-4xptg" Feb 27 08:30:00 crc kubenswrapper[4612]: I0227 08:30:00.182116 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29536350-jwm78"] Feb 27 08:30:00 crc kubenswrapper[4612]: I0227 08:30:00.183987 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536350-jwm78" Feb 27 08:30:00 crc kubenswrapper[4612]: I0227 08:30:00.185963 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-5zwfx" Feb 27 08:30:00 crc kubenswrapper[4612]: I0227 08:30:00.186483 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 27 08:30:00 crc kubenswrapper[4612]: I0227 08:30:00.188982 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 27 08:30:00 crc kubenswrapper[4612]: I0227 08:30:00.202243 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536350-jwm78"] Feb 27 08:30:00 crc kubenswrapper[4612]: I0227 08:30:00.216893 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29536350-4xptg"] Feb 27 08:30:00 crc kubenswrapper[4612]: I0227 08:30:00.271949 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lk4js\" (UniqueName: \"kubernetes.io/projected/6ac60cd4-1e56-4ad4-8ba9-50bc36f8f211-kube-api-access-lk4js\") pod \"collect-profiles-29536350-4xptg\" (UID: \"6ac60cd4-1e56-4ad4-8ba9-50bc36f8f211\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29536350-4xptg" Feb 27 08:30:00 crc kubenswrapper[4612]: I0227 08:30:00.272158 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6ac60cd4-1e56-4ad4-8ba9-50bc36f8f211-secret-volume\") pod \"collect-profiles-29536350-4xptg\" (UID: \"6ac60cd4-1e56-4ad4-8ba9-50bc36f8f211\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29536350-4xptg" Feb 27 08:30:00 crc kubenswrapper[4612]: I0227 08:30:00.272232 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6ac60cd4-1e56-4ad4-8ba9-50bc36f8f211-config-volume\") pod \"collect-profiles-29536350-4xptg\" (UID: \"6ac60cd4-1e56-4ad4-8ba9-50bc36f8f211\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29536350-4xptg" Feb 27 08:30:00 crc kubenswrapper[4612]: I0227 08:30:00.273145 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6ac60cd4-1e56-4ad4-8ba9-50bc36f8f211-config-volume\") pod \"collect-profiles-29536350-4xptg\" (UID: \"6ac60cd4-1e56-4ad4-8ba9-50bc36f8f211\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29536350-4xptg" Feb 27 08:30:00 crc kubenswrapper[4612]: I0227 08:30:00.373948 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qdt7t\" (UniqueName: \"kubernetes.io/projected/98ea71d9-1b11-4cac-8281-c1220056a8ec-kube-api-access-qdt7t\") pod \"auto-csr-approver-29536350-jwm78\" (UID: \"98ea71d9-1b11-4cac-8281-c1220056a8ec\") " pod="openshift-infra/auto-csr-approver-29536350-jwm78" Feb 27 08:30:00 crc kubenswrapper[4612]: I0227 08:30:00.374011 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lk4js\" (UniqueName: \"kubernetes.io/projected/6ac60cd4-1e56-4ad4-8ba9-50bc36f8f211-kube-api-access-lk4js\") pod \"collect-profiles-29536350-4xptg\" (UID: \"6ac60cd4-1e56-4ad4-8ba9-50bc36f8f211\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29536350-4xptg" Feb 27 08:30:00 crc kubenswrapper[4612]: I0227 08:30:00.374038 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6ac60cd4-1e56-4ad4-8ba9-50bc36f8f211-secret-volume\") pod \"collect-profiles-29536350-4xptg\" (UID: \"6ac60cd4-1e56-4ad4-8ba9-50bc36f8f211\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29536350-4xptg" Feb 27 08:30:00 crc kubenswrapper[4612]: I0227 08:30:00.386948 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6ac60cd4-1e56-4ad4-8ba9-50bc36f8f211-secret-volume\") pod \"collect-profiles-29536350-4xptg\" (UID: \"6ac60cd4-1e56-4ad4-8ba9-50bc36f8f211\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29536350-4xptg" Feb 27 08:30:00 crc kubenswrapper[4612]: I0227 08:30:00.397820 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lk4js\" (UniqueName: \"kubernetes.io/projected/6ac60cd4-1e56-4ad4-8ba9-50bc36f8f211-kube-api-access-lk4js\") pod \"collect-profiles-29536350-4xptg\" (UID: \"6ac60cd4-1e56-4ad4-8ba9-50bc36f8f211\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29536350-4xptg" Feb 27 08:30:00 crc kubenswrapper[4612]: I0227 08:30:00.476190 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qdt7t\" (UniqueName: \"kubernetes.io/projected/98ea71d9-1b11-4cac-8281-c1220056a8ec-kube-api-access-qdt7t\") pod \"auto-csr-approver-29536350-jwm78\" (UID: \"98ea71d9-1b11-4cac-8281-c1220056a8ec\") " pod="openshift-infra/auto-csr-approver-29536350-jwm78" Feb 27 08:30:00 crc kubenswrapper[4612]: I0227 08:30:00.476611 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29536350-4xptg" Feb 27 08:30:00 crc kubenswrapper[4612]: I0227 08:30:00.498418 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qdt7t\" (UniqueName: \"kubernetes.io/projected/98ea71d9-1b11-4cac-8281-c1220056a8ec-kube-api-access-qdt7t\") pod \"auto-csr-approver-29536350-jwm78\" (UID: \"98ea71d9-1b11-4cac-8281-c1220056a8ec\") " pod="openshift-infra/auto-csr-approver-29536350-jwm78" Feb 27 08:30:00 crc kubenswrapper[4612]: I0227 08:30:00.499391 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536350-jwm78" Feb 27 08:30:00 crc kubenswrapper[4612]: I0227 08:30:00.982892 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29536350-4xptg"] Feb 27 08:30:01 crc kubenswrapper[4612]: W0227 08:30:01.061650 4612 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod98ea71d9_1b11_4cac_8281_c1220056a8ec.slice/crio-e511d1a352a38acd0286969cb7b39dad43333e0ca98bb5cdba567c96448ebc80 WatchSource:0}: Error finding container e511d1a352a38acd0286969cb7b39dad43333e0ca98bb5cdba567c96448ebc80: Status 404 returned error can't find the container with id e511d1a352a38acd0286969cb7b39dad43333e0ca98bb5cdba567c96448ebc80 Feb 27 08:30:01 crc kubenswrapper[4612]: I0227 08:30:01.064434 4612 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 27 08:30:01 crc kubenswrapper[4612]: I0227 08:30:01.072785 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536350-jwm78"] Feb 27 08:30:01 crc kubenswrapper[4612]: I0227 08:30:01.508060 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536350-jwm78" event={"ID":"98ea71d9-1b11-4cac-8281-c1220056a8ec","Type":"ContainerStarted","Data":"e511d1a352a38acd0286969cb7b39dad43333e0ca98bb5cdba567c96448ebc80"} Feb 27 08:30:01 crc kubenswrapper[4612]: I0227 08:30:01.510060 4612 generic.go:334] "Generic (PLEG): container finished" podID="6ac60cd4-1e56-4ad4-8ba9-50bc36f8f211" containerID="41e4d2dd206394df03ffda8f01431a5bff3160fe5b41c3183e814bcd2d8c7d50" exitCode=0 Feb 27 08:30:01 crc kubenswrapper[4612]: I0227 08:30:01.510110 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29536350-4xptg" event={"ID":"6ac60cd4-1e56-4ad4-8ba9-50bc36f8f211","Type":"ContainerDied","Data":"41e4d2dd206394df03ffda8f01431a5bff3160fe5b41c3183e814bcd2d8c7d50"} Feb 27 08:30:01 crc kubenswrapper[4612]: I0227 08:30:01.510245 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29536350-4xptg" event={"ID":"6ac60cd4-1e56-4ad4-8ba9-50bc36f8f211","Type":"ContainerStarted","Data":"465ab5ba5f0a5a8769549ef4c39fb0b97578ad2a66ee0f21da7c14686930d772"} Feb 27 08:30:02 crc kubenswrapper[4612]: I0227 08:30:02.922857 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29536350-4xptg" Feb 27 08:30:03 crc kubenswrapper[4612]: I0227 08:30:03.029991 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6ac60cd4-1e56-4ad4-8ba9-50bc36f8f211-secret-volume\") pod \"6ac60cd4-1e56-4ad4-8ba9-50bc36f8f211\" (UID: \"6ac60cd4-1e56-4ad4-8ba9-50bc36f8f211\") " Feb 27 08:30:03 crc kubenswrapper[4612]: I0227 08:30:03.030109 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lk4js\" (UniqueName: \"kubernetes.io/projected/6ac60cd4-1e56-4ad4-8ba9-50bc36f8f211-kube-api-access-lk4js\") pod \"6ac60cd4-1e56-4ad4-8ba9-50bc36f8f211\" (UID: \"6ac60cd4-1e56-4ad4-8ba9-50bc36f8f211\") " Feb 27 08:30:03 crc kubenswrapper[4612]: I0227 08:30:03.030186 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6ac60cd4-1e56-4ad4-8ba9-50bc36f8f211-config-volume\") pod \"6ac60cd4-1e56-4ad4-8ba9-50bc36f8f211\" (UID: \"6ac60cd4-1e56-4ad4-8ba9-50bc36f8f211\") " Feb 27 08:30:03 crc kubenswrapper[4612]: I0227 08:30:03.030708 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ac60cd4-1e56-4ad4-8ba9-50bc36f8f211-config-volume" (OuterVolumeSpecName: "config-volume") pod "6ac60cd4-1e56-4ad4-8ba9-50bc36f8f211" (UID: "6ac60cd4-1e56-4ad4-8ba9-50bc36f8f211"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 08:30:03 crc kubenswrapper[4612]: I0227 08:30:03.037380 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ac60cd4-1e56-4ad4-8ba9-50bc36f8f211-kube-api-access-lk4js" (OuterVolumeSpecName: "kube-api-access-lk4js") pod "6ac60cd4-1e56-4ad4-8ba9-50bc36f8f211" (UID: "6ac60cd4-1e56-4ad4-8ba9-50bc36f8f211"). InnerVolumeSpecName "kube-api-access-lk4js". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 08:30:03 crc kubenswrapper[4612]: I0227 08:30:03.038967 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ac60cd4-1e56-4ad4-8ba9-50bc36f8f211-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "6ac60cd4-1e56-4ad4-8ba9-50bc36f8f211" (UID: "6ac60cd4-1e56-4ad4-8ba9-50bc36f8f211"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:30:03 crc kubenswrapper[4612]: I0227 08:30:03.132767 4612 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6ac60cd4-1e56-4ad4-8ba9-50bc36f8f211-secret-volume\") on node \"crc\" DevicePath \"\"" Feb 27 08:30:03 crc kubenswrapper[4612]: I0227 08:30:03.132804 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lk4js\" (UniqueName: \"kubernetes.io/projected/6ac60cd4-1e56-4ad4-8ba9-50bc36f8f211-kube-api-access-lk4js\") on node \"crc\" DevicePath \"\"" Feb 27 08:30:03 crc kubenswrapper[4612]: I0227 08:30:03.132816 4612 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6ac60cd4-1e56-4ad4-8ba9-50bc36f8f211-config-volume\") on node \"crc\" DevicePath \"\"" Feb 27 08:30:03 crc kubenswrapper[4612]: I0227 08:30:03.531861 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29536350-4xptg" Feb 27 08:30:03 crc kubenswrapper[4612]: I0227 08:30:03.531854 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29536350-4xptg" event={"ID":"6ac60cd4-1e56-4ad4-8ba9-50bc36f8f211","Type":"ContainerDied","Data":"465ab5ba5f0a5a8769549ef4c39fb0b97578ad2a66ee0f21da7c14686930d772"} Feb 27 08:30:03 crc kubenswrapper[4612]: I0227 08:30:03.532411 4612 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="465ab5ba5f0a5a8769549ef4c39fb0b97578ad2a66ee0f21da7c14686930d772" Feb 27 08:30:03 crc kubenswrapper[4612]: I0227 08:30:03.533127 4612 generic.go:334] "Generic (PLEG): container finished" podID="98ea71d9-1b11-4cac-8281-c1220056a8ec" containerID="8f940b943ef90992f45e40e478a0bc75fb1875a96059db043b0b71454915012f" exitCode=0 Feb 27 08:30:03 crc kubenswrapper[4612]: I0227 08:30:03.533156 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536350-jwm78" event={"ID":"98ea71d9-1b11-4cac-8281-c1220056a8ec","Type":"ContainerDied","Data":"8f940b943ef90992f45e40e478a0bc75fb1875a96059db043b0b71454915012f"} Feb 27 08:30:04 crc kubenswrapper[4612]: I0227 08:30:04.006460 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29536305-vmtnx"] Feb 27 08:30:04 crc kubenswrapper[4612]: I0227 08:30:04.017651 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29536305-vmtnx"] Feb 27 08:30:04 crc kubenswrapper[4612]: I0227 08:30:04.865918 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ea3d0df0-517f-4865-964d-3e12a313696b" path="/var/lib/kubelet/pods/ea3d0df0-517f-4865-964d-3e12a313696b/volumes" Feb 27 08:30:04 crc kubenswrapper[4612]: I0227 08:30:04.865998 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536350-jwm78" Feb 27 08:30:05 crc kubenswrapper[4612]: I0227 08:30:05.069625 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qdt7t\" (UniqueName: \"kubernetes.io/projected/98ea71d9-1b11-4cac-8281-c1220056a8ec-kube-api-access-qdt7t\") pod \"98ea71d9-1b11-4cac-8281-c1220056a8ec\" (UID: \"98ea71d9-1b11-4cac-8281-c1220056a8ec\") " Feb 27 08:30:05 crc kubenswrapper[4612]: I0227 08:30:05.076637 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/98ea71d9-1b11-4cac-8281-c1220056a8ec-kube-api-access-qdt7t" (OuterVolumeSpecName: "kube-api-access-qdt7t") pod "98ea71d9-1b11-4cac-8281-c1220056a8ec" (UID: "98ea71d9-1b11-4cac-8281-c1220056a8ec"). InnerVolumeSpecName "kube-api-access-qdt7t". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 08:30:05 crc kubenswrapper[4612]: I0227 08:30:05.171659 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qdt7t\" (UniqueName: \"kubernetes.io/projected/98ea71d9-1b11-4cac-8281-c1220056a8ec-kube-api-access-qdt7t\") on node \"crc\" DevicePath \"\"" Feb 27 08:30:05 crc kubenswrapper[4612]: I0227 08:30:05.553255 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536350-jwm78" event={"ID":"98ea71d9-1b11-4cac-8281-c1220056a8ec","Type":"ContainerDied","Data":"e511d1a352a38acd0286969cb7b39dad43333e0ca98bb5cdba567c96448ebc80"} Feb 27 08:30:05 crc kubenswrapper[4612]: I0227 08:30:05.553850 4612 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e511d1a352a38acd0286969cb7b39dad43333e0ca98bb5cdba567c96448ebc80" Feb 27 08:30:05 crc kubenswrapper[4612]: I0227 08:30:05.553323 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536350-jwm78" Feb 27 08:30:05 crc kubenswrapper[4612]: I0227 08:30:05.938168 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29536344-pdbxv"] Feb 27 08:30:05 crc kubenswrapper[4612]: I0227 08:30:05.946398 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29536344-pdbxv"] Feb 27 08:30:06 crc kubenswrapper[4612]: I0227 08:30:06.868606 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b8b16b31-ecd0-4fb2-b827-ec0ea2f0ef03" path="/var/lib/kubelet/pods/b8b16b31-ecd0-4fb2-b827-ec0ea2f0ef03/volumes" Feb 27 08:30:10 crc kubenswrapper[4612]: I0227 08:30:10.853568 4612 scope.go:117] "RemoveContainer" containerID="76c2de0055c53fc536f48b64de557ea4f7b56a2de457f8f1b8708b99c7c612cb" Feb 27 08:30:10 crc kubenswrapper[4612]: E0227 08:30:10.870049 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-924vb_openshift-machine-config-operator(24599373-7adc-4a1b-8bb4-797bf726f43d)\"" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" Feb 27 08:30:17 crc kubenswrapper[4612]: I0227 08:30:17.728249 4612 scope.go:117] "RemoveContainer" containerID="9f4a63ddb5dc79f146a7585e5535f6c947de9edebadc177ca8b4362f7aa13188" Feb 27 08:30:17 crc kubenswrapper[4612]: I0227 08:30:17.801973 4612 scope.go:117] "RemoveContainer" containerID="0023d505698c1042b5a378107945cbb81087ed8de071109ea9bcf9910711f71e" Feb 27 08:30:22 crc kubenswrapper[4612]: I0227 08:30:22.858910 4612 scope.go:117] "RemoveContainer" containerID="76c2de0055c53fc536f48b64de557ea4f7b56a2de457f8f1b8708b99c7c612cb" Feb 27 08:30:22 crc kubenswrapper[4612]: E0227 08:30:22.859672 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-924vb_openshift-machine-config-operator(24599373-7adc-4a1b-8bb4-797bf726f43d)\"" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" Feb 27 08:30:34 crc kubenswrapper[4612]: I0227 08:30:34.856512 4612 scope.go:117] "RemoveContainer" containerID="76c2de0055c53fc536f48b64de557ea4f7b56a2de457f8f1b8708b99c7c612cb" Feb 27 08:30:34 crc kubenswrapper[4612]: E0227 08:30:34.857648 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-924vb_openshift-machine-config-operator(24599373-7adc-4a1b-8bb4-797bf726f43d)\"" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" Feb 27 08:30:37 crc kubenswrapper[4612]: I0227 08:30:37.913192 4612 generic.go:334] "Generic (PLEG): container finished" podID="f4308354-0bda-4fb3-976f-cc2fa471dcb4" containerID="8e6f6e66e01c1e16a3b54d7d03f2cfdf920a64f28ebd1cf42df58a1e0143a429" exitCode=0 Feb 27 08:30:37 crc kubenswrapper[4612]: I0227 08:30:37.913281 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-27c56" event={"ID":"f4308354-0bda-4fb3-976f-cc2fa471dcb4","Type":"ContainerDied","Data":"8e6f6e66e01c1e16a3b54d7d03f2cfdf920a64f28ebd1cf42df58a1e0143a429"} Feb 27 08:30:39 crc kubenswrapper[4612]: I0227 08:30:39.366930 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-27c56" Feb 27 08:30:39 crc kubenswrapper[4612]: I0227 08:30:39.423222 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/f4308354-0bda-4fb3-976f-cc2fa471dcb4-ssh-key-openstack-edpm-ipam\") pod \"f4308354-0bda-4fb3-976f-cc2fa471dcb4\" (UID: \"f4308354-0bda-4fb3-976f-cc2fa471dcb4\") " Feb 27 08:30:39 crc kubenswrapper[4612]: I0227 08:30:39.423295 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/f4308354-0bda-4fb3-976f-cc2fa471dcb4-libvirt-secret-0\") pod \"f4308354-0bda-4fb3-976f-cc2fa471dcb4\" (UID: \"f4308354-0bda-4fb3-976f-cc2fa471dcb4\") " Feb 27 08:30:39 crc kubenswrapper[4612]: I0227 08:30:39.423361 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f4308354-0bda-4fb3-976f-cc2fa471dcb4-libvirt-combined-ca-bundle\") pod \"f4308354-0bda-4fb3-976f-cc2fa471dcb4\" (UID: \"f4308354-0bda-4fb3-976f-cc2fa471dcb4\") " Feb 27 08:30:39 crc kubenswrapper[4612]: I0227 08:30:39.423491 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g8qlt\" (UniqueName: \"kubernetes.io/projected/f4308354-0bda-4fb3-976f-cc2fa471dcb4-kube-api-access-g8qlt\") pod \"f4308354-0bda-4fb3-976f-cc2fa471dcb4\" (UID: \"f4308354-0bda-4fb3-976f-cc2fa471dcb4\") " Feb 27 08:30:39 crc kubenswrapper[4612]: I0227 08:30:39.423570 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f4308354-0bda-4fb3-976f-cc2fa471dcb4-inventory\") pod \"f4308354-0bda-4fb3-976f-cc2fa471dcb4\" (UID: \"f4308354-0bda-4fb3-976f-cc2fa471dcb4\") " Feb 27 08:30:39 crc kubenswrapper[4612]: I0227 08:30:39.429754 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f4308354-0bda-4fb3-976f-cc2fa471dcb4-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "f4308354-0bda-4fb3-976f-cc2fa471dcb4" (UID: "f4308354-0bda-4fb3-976f-cc2fa471dcb4"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:30:39 crc kubenswrapper[4612]: I0227 08:30:39.430011 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f4308354-0bda-4fb3-976f-cc2fa471dcb4-kube-api-access-g8qlt" (OuterVolumeSpecName: "kube-api-access-g8qlt") pod "f4308354-0bda-4fb3-976f-cc2fa471dcb4" (UID: "f4308354-0bda-4fb3-976f-cc2fa471dcb4"). InnerVolumeSpecName "kube-api-access-g8qlt". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 08:30:39 crc kubenswrapper[4612]: I0227 08:30:39.453537 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f4308354-0bda-4fb3-976f-cc2fa471dcb4-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "f4308354-0bda-4fb3-976f-cc2fa471dcb4" (UID: "f4308354-0bda-4fb3-976f-cc2fa471dcb4"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:30:39 crc kubenswrapper[4612]: I0227 08:30:39.453886 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f4308354-0bda-4fb3-976f-cc2fa471dcb4-libvirt-secret-0" (OuterVolumeSpecName: "libvirt-secret-0") pod "f4308354-0bda-4fb3-976f-cc2fa471dcb4" (UID: "f4308354-0bda-4fb3-976f-cc2fa471dcb4"). InnerVolumeSpecName "libvirt-secret-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:30:39 crc kubenswrapper[4612]: I0227 08:30:39.465996 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f4308354-0bda-4fb3-976f-cc2fa471dcb4-inventory" (OuterVolumeSpecName: "inventory") pod "f4308354-0bda-4fb3-976f-cc2fa471dcb4" (UID: "f4308354-0bda-4fb3-976f-cc2fa471dcb4"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:30:39 crc kubenswrapper[4612]: I0227 08:30:39.525581 4612 reconciler_common.go:293] "Volume detached for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/f4308354-0bda-4fb3-976f-cc2fa471dcb4-libvirt-secret-0\") on node \"crc\" DevicePath \"\"" Feb 27 08:30:39 crc kubenswrapper[4612]: I0227 08:30:39.525613 4612 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f4308354-0bda-4fb3-976f-cc2fa471dcb4-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 08:30:39 crc kubenswrapper[4612]: I0227 08:30:39.525630 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g8qlt\" (UniqueName: \"kubernetes.io/projected/f4308354-0bda-4fb3-976f-cc2fa471dcb4-kube-api-access-g8qlt\") on node \"crc\" DevicePath \"\"" Feb 27 08:30:39 crc kubenswrapper[4612]: I0227 08:30:39.525641 4612 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f4308354-0bda-4fb3-976f-cc2fa471dcb4-inventory\") on node \"crc\" DevicePath \"\"" Feb 27 08:30:39 crc kubenswrapper[4612]: I0227 08:30:39.525651 4612 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/f4308354-0bda-4fb3-976f-cc2fa471dcb4-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 27 08:30:39 crc kubenswrapper[4612]: I0227 08:30:39.934937 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-27c56" event={"ID":"f4308354-0bda-4fb3-976f-cc2fa471dcb4","Type":"ContainerDied","Data":"12344af5441dc29acf62d0da60c4a7b403628677e00c02b5a28f40fef7a60c6b"} Feb 27 08:30:39 crc kubenswrapper[4612]: I0227 08:30:39.935780 4612 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="12344af5441dc29acf62d0da60c4a7b403628677e00c02b5a28f40fef7a60c6b" Feb 27 08:30:39 crc kubenswrapper[4612]: I0227 08:30:39.935244 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-27c56" Feb 27 08:30:40 crc kubenswrapper[4612]: I0227 08:30:40.021476 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-bnlnx"] Feb 27 08:30:40 crc kubenswrapper[4612]: E0227 08:30:40.021846 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4308354-0bda-4fb3-976f-cc2fa471dcb4" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Feb 27 08:30:40 crc kubenswrapper[4612]: I0227 08:30:40.021862 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4308354-0bda-4fb3-976f-cc2fa471dcb4" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Feb 27 08:30:40 crc kubenswrapper[4612]: E0227 08:30:40.021881 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="98ea71d9-1b11-4cac-8281-c1220056a8ec" containerName="oc" Feb 27 08:30:40 crc kubenswrapper[4612]: I0227 08:30:40.021888 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="98ea71d9-1b11-4cac-8281-c1220056a8ec" containerName="oc" Feb 27 08:30:40 crc kubenswrapper[4612]: E0227 08:30:40.021911 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ac60cd4-1e56-4ad4-8ba9-50bc36f8f211" containerName="collect-profiles" Feb 27 08:30:40 crc kubenswrapper[4612]: I0227 08:30:40.021917 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ac60cd4-1e56-4ad4-8ba9-50bc36f8f211" containerName="collect-profiles" Feb 27 08:30:40 crc kubenswrapper[4612]: I0227 08:30:40.022068 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="6ac60cd4-1e56-4ad4-8ba9-50bc36f8f211" containerName="collect-profiles" Feb 27 08:30:40 crc kubenswrapper[4612]: I0227 08:30:40.022084 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4308354-0bda-4fb3-976f-cc2fa471dcb4" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Feb 27 08:30:40 crc kubenswrapper[4612]: I0227 08:30:40.022096 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="98ea71d9-1b11-4cac-8281-c1220056a8ec" containerName="oc" Feb 27 08:30:40 crc kubenswrapper[4612]: I0227 08:30:40.022666 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-bnlnx" Feb 27 08:30:40 crc kubenswrapper[4612]: I0227 08:30:40.025365 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 27 08:30:40 crc kubenswrapper[4612]: I0227 08:30:40.025883 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 27 08:30:40 crc kubenswrapper[4612]: I0227 08:30:40.026375 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Feb 27 08:30:40 crc kubenswrapper[4612]: I0227 08:30:40.026936 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 27 08:30:40 crc kubenswrapper[4612]: I0227 08:30:40.027330 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Feb 27 08:30:40 crc kubenswrapper[4612]: I0227 08:30:40.027682 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-extra-config" Feb 27 08:30:40 crc kubenswrapper[4612]: I0227 08:30:40.036308 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-2pwrb" Feb 27 08:30:40 crc kubenswrapper[4612]: I0227 08:30:40.075450 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-bnlnx"] Feb 27 08:30:40 crc kubenswrapper[4612]: I0227 08:30:40.140604 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-2\" (UniqueName: \"kubernetes.io/secret/24758178-1be1-49f2-a15e-196426ec0dcd-nova-cell1-compute-config-2\") pod \"nova-edpm-deployment-openstack-edpm-ipam-bnlnx\" (UID: \"24758178-1be1-49f2-a15e-196426ec0dcd\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-bnlnx" Feb 27 08:30:40 crc kubenswrapper[4612]: I0227 08:30:40.140681 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/24758178-1be1-49f2-a15e-196426ec0dcd-ssh-key-openstack-edpm-ipam\") pod \"nova-edpm-deployment-openstack-edpm-ipam-bnlnx\" (UID: \"24758178-1be1-49f2-a15e-196426ec0dcd\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-bnlnx" Feb 27 08:30:40 crc kubenswrapper[4612]: I0227 08:30:40.140800 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/24758178-1be1-49f2-a15e-196426ec0dcd-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-bnlnx\" (UID: \"24758178-1be1-49f2-a15e-196426ec0dcd\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-bnlnx" Feb 27 08:30:40 crc kubenswrapper[4612]: I0227 08:30:40.140830 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/24758178-1be1-49f2-a15e-196426ec0dcd-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-bnlnx\" (UID: \"24758178-1be1-49f2-a15e-196426ec0dcd\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-bnlnx" Feb 27 08:30:40 crc kubenswrapper[4612]: I0227 08:30:40.140851 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24758178-1be1-49f2-a15e-196426ec0dcd-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-bnlnx\" (UID: \"24758178-1be1-49f2-a15e-196426ec0dcd\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-bnlnx" Feb 27 08:30:40 crc kubenswrapper[4612]: I0227 08:30:40.140881 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tfdnl\" (UniqueName: \"kubernetes.io/projected/24758178-1be1-49f2-a15e-196426ec0dcd-kube-api-access-tfdnl\") pod \"nova-edpm-deployment-openstack-edpm-ipam-bnlnx\" (UID: \"24758178-1be1-49f2-a15e-196426ec0dcd\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-bnlnx" Feb 27 08:30:40 crc kubenswrapper[4612]: I0227 08:30:40.140940 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/24758178-1be1-49f2-a15e-196426ec0dcd-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-bnlnx\" (UID: \"24758178-1be1-49f2-a15e-196426ec0dcd\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-bnlnx" Feb 27 08:30:40 crc kubenswrapper[4612]: I0227 08:30:40.141000 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/24758178-1be1-49f2-a15e-196426ec0dcd-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-bnlnx\" (UID: \"24758178-1be1-49f2-a15e-196426ec0dcd\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-bnlnx" Feb 27 08:30:40 crc kubenswrapper[4612]: I0227 08:30:40.141043 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-3\" (UniqueName: \"kubernetes.io/secret/24758178-1be1-49f2-a15e-196426ec0dcd-nova-cell1-compute-config-3\") pod \"nova-edpm-deployment-openstack-edpm-ipam-bnlnx\" (UID: \"24758178-1be1-49f2-a15e-196426ec0dcd\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-bnlnx" Feb 27 08:30:40 crc kubenswrapper[4612]: I0227 08:30:40.141101 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/24758178-1be1-49f2-a15e-196426ec0dcd-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-bnlnx\" (UID: \"24758178-1be1-49f2-a15e-196426ec0dcd\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-bnlnx" Feb 27 08:30:40 crc kubenswrapper[4612]: I0227 08:30:40.141138 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/24758178-1be1-49f2-a15e-196426ec0dcd-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-bnlnx\" (UID: \"24758178-1be1-49f2-a15e-196426ec0dcd\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-bnlnx" Feb 27 08:30:40 crc kubenswrapper[4612]: I0227 08:30:40.242294 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-3\" (UniqueName: \"kubernetes.io/secret/24758178-1be1-49f2-a15e-196426ec0dcd-nova-cell1-compute-config-3\") pod \"nova-edpm-deployment-openstack-edpm-ipam-bnlnx\" (UID: \"24758178-1be1-49f2-a15e-196426ec0dcd\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-bnlnx" Feb 27 08:30:40 crc kubenswrapper[4612]: I0227 08:30:40.242357 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/24758178-1be1-49f2-a15e-196426ec0dcd-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-bnlnx\" (UID: \"24758178-1be1-49f2-a15e-196426ec0dcd\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-bnlnx" Feb 27 08:30:40 crc kubenswrapper[4612]: I0227 08:30:40.242377 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/24758178-1be1-49f2-a15e-196426ec0dcd-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-bnlnx\" (UID: \"24758178-1be1-49f2-a15e-196426ec0dcd\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-bnlnx" Feb 27 08:30:40 crc kubenswrapper[4612]: I0227 08:30:40.242400 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-2\" (UniqueName: \"kubernetes.io/secret/24758178-1be1-49f2-a15e-196426ec0dcd-nova-cell1-compute-config-2\") pod \"nova-edpm-deployment-openstack-edpm-ipam-bnlnx\" (UID: \"24758178-1be1-49f2-a15e-196426ec0dcd\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-bnlnx" Feb 27 08:30:40 crc kubenswrapper[4612]: I0227 08:30:40.242444 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/24758178-1be1-49f2-a15e-196426ec0dcd-ssh-key-openstack-edpm-ipam\") pod \"nova-edpm-deployment-openstack-edpm-ipam-bnlnx\" (UID: \"24758178-1be1-49f2-a15e-196426ec0dcd\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-bnlnx" Feb 27 08:30:40 crc kubenswrapper[4612]: I0227 08:30:40.242499 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/24758178-1be1-49f2-a15e-196426ec0dcd-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-bnlnx\" (UID: \"24758178-1be1-49f2-a15e-196426ec0dcd\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-bnlnx" Feb 27 08:30:40 crc kubenswrapper[4612]: I0227 08:30:40.242519 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/24758178-1be1-49f2-a15e-196426ec0dcd-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-bnlnx\" (UID: \"24758178-1be1-49f2-a15e-196426ec0dcd\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-bnlnx" Feb 27 08:30:40 crc kubenswrapper[4612]: I0227 08:30:40.242535 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24758178-1be1-49f2-a15e-196426ec0dcd-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-bnlnx\" (UID: \"24758178-1be1-49f2-a15e-196426ec0dcd\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-bnlnx" Feb 27 08:30:40 crc kubenswrapper[4612]: I0227 08:30:40.242556 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tfdnl\" (UniqueName: \"kubernetes.io/projected/24758178-1be1-49f2-a15e-196426ec0dcd-kube-api-access-tfdnl\") pod \"nova-edpm-deployment-openstack-edpm-ipam-bnlnx\" (UID: \"24758178-1be1-49f2-a15e-196426ec0dcd\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-bnlnx" Feb 27 08:30:40 crc kubenswrapper[4612]: I0227 08:30:40.242579 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/24758178-1be1-49f2-a15e-196426ec0dcd-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-bnlnx\" (UID: \"24758178-1be1-49f2-a15e-196426ec0dcd\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-bnlnx" Feb 27 08:30:40 crc kubenswrapper[4612]: I0227 08:30:40.242618 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/24758178-1be1-49f2-a15e-196426ec0dcd-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-bnlnx\" (UID: \"24758178-1be1-49f2-a15e-196426ec0dcd\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-bnlnx" Feb 27 08:30:40 crc kubenswrapper[4612]: I0227 08:30:40.244611 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/24758178-1be1-49f2-a15e-196426ec0dcd-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-bnlnx\" (UID: \"24758178-1be1-49f2-a15e-196426ec0dcd\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-bnlnx" Feb 27 08:30:40 crc kubenswrapper[4612]: I0227 08:30:40.247382 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/24758178-1be1-49f2-a15e-196426ec0dcd-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-bnlnx\" (UID: \"24758178-1be1-49f2-a15e-196426ec0dcd\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-bnlnx" Feb 27 08:30:40 crc kubenswrapper[4612]: I0227 08:30:40.248258 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/24758178-1be1-49f2-a15e-196426ec0dcd-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-bnlnx\" (UID: \"24758178-1be1-49f2-a15e-196426ec0dcd\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-bnlnx" Feb 27 08:30:40 crc kubenswrapper[4612]: I0227 08:30:40.249024 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/24758178-1be1-49f2-a15e-196426ec0dcd-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-bnlnx\" (UID: \"24758178-1be1-49f2-a15e-196426ec0dcd\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-bnlnx" Feb 27 08:30:40 crc kubenswrapper[4612]: I0227 08:30:40.249652 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/24758178-1be1-49f2-a15e-196426ec0dcd-ssh-key-openstack-edpm-ipam\") pod \"nova-edpm-deployment-openstack-edpm-ipam-bnlnx\" (UID: \"24758178-1be1-49f2-a15e-196426ec0dcd\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-bnlnx" Feb 27 08:30:40 crc kubenswrapper[4612]: I0227 08:30:40.249725 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24758178-1be1-49f2-a15e-196426ec0dcd-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-bnlnx\" (UID: \"24758178-1be1-49f2-a15e-196426ec0dcd\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-bnlnx" Feb 27 08:30:40 crc kubenswrapper[4612]: I0227 08:30:40.249968 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/24758178-1be1-49f2-a15e-196426ec0dcd-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-bnlnx\" (UID: \"24758178-1be1-49f2-a15e-196426ec0dcd\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-bnlnx" Feb 27 08:30:40 crc kubenswrapper[4612]: I0227 08:30:40.250724 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-3\" (UniqueName: \"kubernetes.io/secret/24758178-1be1-49f2-a15e-196426ec0dcd-nova-cell1-compute-config-3\") pod \"nova-edpm-deployment-openstack-edpm-ipam-bnlnx\" (UID: \"24758178-1be1-49f2-a15e-196426ec0dcd\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-bnlnx" Feb 27 08:30:40 crc kubenswrapper[4612]: I0227 08:30:40.252209 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-2\" (UniqueName: \"kubernetes.io/secret/24758178-1be1-49f2-a15e-196426ec0dcd-nova-cell1-compute-config-2\") pod \"nova-edpm-deployment-openstack-edpm-ipam-bnlnx\" (UID: \"24758178-1be1-49f2-a15e-196426ec0dcd\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-bnlnx" Feb 27 08:30:40 crc kubenswrapper[4612]: I0227 08:30:40.253360 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/24758178-1be1-49f2-a15e-196426ec0dcd-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-bnlnx\" (UID: \"24758178-1be1-49f2-a15e-196426ec0dcd\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-bnlnx" Feb 27 08:30:40 crc kubenswrapper[4612]: I0227 08:30:40.267661 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tfdnl\" (UniqueName: \"kubernetes.io/projected/24758178-1be1-49f2-a15e-196426ec0dcd-kube-api-access-tfdnl\") pod \"nova-edpm-deployment-openstack-edpm-ipam-bnlnx\" (UID: \"24758178-1be1-49f2-a15e-196426ec0dcd\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-bnlnx" Feb 27 08:30:40 crc kubenswrapper[4612]: I0227 08:30:40.352724 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-bnlnx" Feb 27 08:30:40 crc kubenswrapper[4612]: I0227 08:30:40.930252 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-bnlnx"] Feb 27 08:30:40 crc kubenswrapper[4612]: W0227 08:30:40.931854 4612 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod24758178_1be1_49f2_a15e_196426ec0dcd.slice/crio-1bf0f54131813b7e9659d0172113a5e289d8d66b70547296e6ba2a72e414ee9f WatchSource:0}: Error finding container 1bf0f54131813b7e9659d0172113a5e289d8d66b70547296e6ba2a72e414ee9f: Status 404 returned error can't find the container with id 1bf0f54131813b7e9659d0172113a5e289d8d66b70547296e6ba2a72e414ee9f Feb 27 08:30:40 crc kubenswrapper[4612]: I0227 08:30:40.944811 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-bnlnx" event={"ID":"24758178-1be1-49f2-a15e-196426ec0dcd","Type":"ContainerStarted","Data":"1bf0f54131813b7e9659d0172113a5e289d8d66b70547296e6ba2a72e414ee9f"} Feb 27 08:30:41 crc kubenswrapper[4612]: I0227 08:30:41.985157 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-bnlnx" event={"ID":"24758178-1be1-49f2-a15e-196426ec0dcd","Type":"ContainerStarted","Data":"fb5b0f9df995b9b980e1ebbd6a95bdd3747fbda8692f05813bc16c3494e152fc"} Feb 27 08:30:42 crc kubenswrapper[4612]: I0227 08:30:42.011664 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-bnlnx" podStartSLOduration=1.378711129 podStartE2EDuration="2.01164715s" podCreationTimestamp="2026-02-27 08:30:40 +0000 UTC" firstStartedPulling="2026-02-27 08:30:40.933808275 +0000 UTC m=+2498.787738273" lastFinishedPulling="2026-02-27 08:30:41.566744296 +0000 UTC m=+2499.420674294" observedRunningTime="2026-02-27 08:30:42.007562013 +0000 UTC m=+2499.861492011" watchObservedRunningTime="2026-02-27 08:30:42.01164715 +0000 UTC m=+2499.865577148" Feb 27 08:30:45 crc kubenswrapper[4612]: I0227 08:30:45.853096 4612 scope.go:117] "RemoveContainer" containerID="76c2de0055c53fc536f48b64de557ea4f7b56a2de457f8f1b8708b99c7c612cb" Feb 27 08:30:45 crc kubenswrapper[4612]: E0227 08:30:45.853817 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-924vb_openshift-machine-config-operator(24599373-7adc-4a1b-8bb4-797bf726f43d)\"" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" Feb 27 08:30:56 crc kubenswrapper[4612]: I0227 08:30:56.853065 4612 scope.go:117] "RemoveContainer" containerID="76c2de0055c53fc536f48b64de557ea4f7b56a2de457f8f1b8708b99c7c612cb" Feb 27 08:30:56 crc kubenswrapper[4612]: E0227 08:30:56.854662 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-924vb_openshift-machine-config-operator(24599373-7adc-4a1b-8bb4-797bf726f43d)\"" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" Feb 27 08:31:09 crc kubenswrapper[4612]: I0227 08:31:09.853507 4612 scope.go:117] "RemoveContainer" containerID="76c2de0055c53fc536f48b64de557ea4f7b56a2de457f8f1b8708b99c7c612cb" Feb 27 08:31:09 crc kubenswrapper[4612]: E0227 08:31:09.854300 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-924vb_openshift-machine-config-operator(24599373-7adc-4a1b-8bb4-797bf726f43d)\"" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" Feb 27 08:31:22 crc kubenswrapper[4612]: I0227 08:31:22.861270 4612 scope.go:117] "RemoveContainer" containerID="76c2de0055c53fc536f48b64de557ea4f7b56a2de457f8f1b8708b99c7c612cb" Feb 27 08:31:22 crc kubenswrapper[4612]: E0227 08:31:22.862348 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-924vb_openshift-machine-config-operator(24599373-7adc-4a1b-8bb4-797bf726f43d)\"" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" Feb 27 08:31:33 crc kubenswrapper[4612]: I0227 08:31:33.852982 4612 scope.go:117] "RemoveContainer" containerID="76c2de0055c53fc536f48b64de557ea4f7b56a2de457f8f1b8708b99c7c612cb" Feb 27 08:31:33 crc kubenswrapper[4612]: E0227 08:31:33.853673 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-924vb_openshift-machine-config-operator(24599373-7adc-4a1b-8bb4-797bf726f43d)\"" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" Feb 27 08:31:47 crc kubenswrapper[4612]: I0227 08:31:47.852868 4612 scope.go:117] "RemoveContainer" containerID="76c2de0055c53fc536f48b64de557ea4f7b56a2de457f8f1b8708b99c7c612cb" Feb 27 08:31:48 crc kubenswrapper[4612]: I0227 08:31:48.614398 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-924vb" event={"ID":"24599373-7adc-4a1b-8bb4-797bf726f43d","Type":"ContainerStarted","Data":"542804ac84a667bc92e9df5b1d69c8ad7e48579a796d2b39142a8890f0acd72f"} Feb 27 08:32:00 crc kubenswrapper[4612]: I0227 08:32:00.151430 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29536352-brq8t"] Feb 27 08:32:00 crc kubenswrapper[4612]: I0227 08:32:00.158001 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536352-brq8t" Feb 27 08:32:00 crc kubenswrapper[4612]: I0227 08:32:00.161263 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-5zwfx" Feb 27 08:32:00 crc kubenswrapper[4612]: I0227 08:32:00.161281 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 27 08:32:00 crc kubenswrapper[4612]: I0227 08:32:00.162834 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 27 08:32:00 crc kubenswrapper[4612]: I0227 08:32:00.165449 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536352-brq8t"] Feb 27 08:32:00 crc kubenswrapper[4612]: I0227 08:32:00.312062 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gkp5x\" (UniqueName: \"kubernetes.io/projected/318ef53b-57e4-4954-867e-d845fc77a484-kube-api-access-gkp5x\") pod \"auto-csr-approver-29536352-brq8t\" (UID: \"318ef53b-57e4-4954-867e-d845fc77a484\") " pod="openshift-infra/auto-csr-approver-29536352-brq8t" Feb 27 08:32:00 crc kubenswrapper[4612]: I0227 08:32:00.414203 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gkp5x\" (UniqueName: \"kubernetes.io/projected/318ef53b-57e4-4954-867e-d845fc77a484-kube-api-access-gkp5x\") pod \"auto-csr-approver-29536352-brq8t\" (UID: \"318ef53b-57e4-4954-867e-d845fc77a484\") " pod="openshift-infra/auto-csr-approver-29536352-brq8t" Feb 27 08:32:00 crc kubenswrapper[4612]: I0227 08:32:00.441014 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gkp5x\" (UniqueName: \"kubernetes.io/projected/318ef53b-57e4-4954-867e-d845fc77a484-kube-api-access-gkp5x\") pod \"auto-csr-approver-29536352-brq8t\" (UID: \"318ef53b-57e4-4954-867e-d845fc77a484\") " pod="openshift-infra/auto-csr-approver-29536352-brq8t" Feb 27 08:32:00 crc kubenswrapper[4612]: I0227 08:32:00.486349 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536352-brq8t" Feb 27 08:32:01 crc kubenswrapper[4612]: I0227 08:32:01.002707 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536352-brq8t"] Feb 27 08:32:01 crc kubenswrapper[4612]: I0227 08:32:01.746620 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536352-brq8t" event={"ID":"318ef53b-57e4-4954-867e-d845fc77a484","Type":"ContainerStarted","Data":"4782d427f65fbc4e17171330352a2814108428f3126918535bcad514de307b37"} Feb 27 08:32:02 crc kubenswrapper[4612]: I0227 08:32:02.757478 4612 generic.go:334] "Generic (PLEG): container finished" podID="318ef53b-57e4-4954-867e-d845fc77a484" containerID="126c9df520238578841820ddbaf8bb9375205640276c7baf7814f85751306ecb" exitCode=0 Feb 27 08:32:02 crc kubenswrapper[4612]: I0227 08:32:02.757551 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536352-brq8t" event={"ID":"318ef53b-57e4-4954-867e-d845fc77a484","Type":"ContainerDied","Data":"126c9df520238578841820ddbaf8bb9375205640276c7baf7814f85751306ecb"} Feb 27 08:32:04 crc kubenswrapper[4612]: I0227 08:32:04.122494 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536352-brq8t" Feb 27 08:32:04 crc kubenswrapper[4612]: I0227 08:32:04.290988 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gkp5x\" (UniqueName: \"kubernetes.io/projected/318ef53b-57e4-4954-867e-d845fc77a484-kube-api-access-gkp5x\") pod \"318ef53b-57e4-4954-867e-d845fc77a484\" (UID: \"318ef53b-57e4-4954-867e-d845fc77a484\") " Feb 27 08:32:04 crc kubenswrapper[4612]: I0227 08:32:04.301653 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/318ef53b-57e4-4954-867e-d845fc77a484-kube-api-access-gkp5x" (OuterVolumeSpecName: "kube-api-access-gkp5x") pod "318ef53b-57e4-4954-867e-d845fc77a484" (UID: "318ef53b-57e4-4954-867e-d845fc77a484"). InnerVolumeSpecName "kube-api-access-gkp5x". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 08:32:04 crc kubenswrapper[4612]: I0227 08:32:04.393263 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gkp5x\" (UniqueName: \"kubernetes.io/projected/318ef53b-57e4-4954-867e-d845fc77a484-kube-api-access-gkp5x\") on node \"crc\" DevicePath \"\"" Feb 27 08:32:04 crc kubenswrapper[4612]: I0227 08:32:04.777744 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536352-brq8t" event={"ID":"318ef53b-57e4-4954-867e-d845fc77a484","Type":"ContainerDied","Data":"4782d427f65fbc4e17171330352a2814108428f3126918535bcad514de307b37"} Feb 27 08:32:04 crc kubenswrapper[4612]: I0227 08:32:04.778247 4612 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4782d427f65fbc4e17171330352a2814108428f3126918535bcad514de307b37" Feb 27 08:32:04 crc kubenswrapper[4612]: I0227 08:32:04.777796 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536352-brq8t" Feb 27 08:32:05 crc kubenswrapper[4612]: I0227 08:32:05.191099 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29536346-c26b4"] Feb 27 08:32:05 crc kubenswrapper[4612]: I0227 08:32:05.202295 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29536346-c26b4"] Feb 27 08:32:06 crc kubenswrapper[4612]: I0227 08:32:06.864758 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fba772a3-9097-4773-bff1-98ee5ebaa81c" path="/var/lib/kubelet/pods/fba772a3-9097-4773-bff1-98ee5ebaa81c/volumes" Feb 27 08:32:17 crc kubenswrapper[4612]: I0227 08:32:17.975947 4612 scope.go:117] "RemoveContainer" containerID="455208a77a54fb810ab226d17541bdb62481523d5d2a486b7aff76c2d10e2427" Feb 27 08:33:16 crc kubenswrapper[4612]: I0227 08:33:16.612922 4612 generic.go:334] "Generic (PLEG): container finished" podID="24758178-1be1-49f2-a15e-196426ec0dcd" containerID="fb5b0f9df995b9b980e1ebbd6a95bdd3747fbda8692f05813bc16c3494e152fc" exitCode=0 Feb 27 08:33:16 crc kubenswrapper[4612]: I0227 08:33:16.612983 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-bnlnx" event={"ID":"24758178-1be1-49f2-a15e-196426ec0dcd","Type":"ContainerDied","Data":"fb5b0f9df995b9b980e1ebbd6a95bdd3747fbda8692f05813bc16c3494e152fc"} Feb 27 08:33:18 crc kubenswrapper[4612]: I0227 08:33:18.053906 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-bnlnx" Feb 27 08:33:18 crc kubenswrapper[4612]: I0227 08:33:18.232579 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24758178-1be1-49f2-a15e-196426ec0dcd-nova-combined-ca-bundle\") pod \"24758178-1be1-49f2-a15e-196426ec0dcd\" (UID: \"24758178-1be1-49f2-a15e-196426ec0dcd\") " Feb 27 08:33:18 crc kubenswrapper[4612]: I0227 08:33:18.232638 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/24758178-1be1-49f2-a15e-196426ec0dcd-nova-cell1-compute-config-0\") pod \"24758178-1be1-49f2-a15e-196426ec0dcd\" (UID: \"24758178-1be1-49f2-a15e-196426ec0dcd\") " Feb 27 08:33:18 crc kubenswrapper[4612]: I0227 08:33:18.232669 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/24758178-1be1-49f2-a15e-196426ec0dcd-nova-cell1-compute-config-1\") pod \"24758178-1be1-49f2-a15e-196426ec0dcd\" (UID: \"24758178-1be1-49f2-a15e-196426ec0dcd\") " Feb 27 08:33:18 crc kubenswrapper[4612]: I0227 08:33:18.232710 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/24758178-1be1-49f2-a15e-196426ec0dcd-ssh-key-openstack-edpm-ipam\") pod \"24758178-1be1-49f2-a15e-196426ec0dcd\" (UID: \"24758178-1be1-49f2-a15e-196426ec0dcd\") " Feb 27 08:33:18 crc kubenswrapper[4612]: I0227 08:33:18.232737 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/24758178-1be1-49f2-a15e-196426ec0dcd-nova-extra-config-0\") pod \"24758178-1be1-49f2-a15e-196426ec0dcd\" (UID: \"24758178-1be1-49f2-a15e-196426ec0dcd\") " Feb 27 08:33:18 crc kubenswrapper[4612]: I0227 08:33:18.233571 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-3\" (UniqueName: \"kubernetes.io/secret/24758178-1be1-49f2-a15e-196426ec0dcd-nova-cell1-compute-config-3\") pod \"24758178-1be1-49f2-a15e-196426ec0dcd\" (UID: \"24758178-1be1-49f2-a15e-196426ec0dcd\") " Feb 27 08:33:18 crc kubenswrapper[4612]: I0227 08:33:18.233616 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/24758178-1be1-49f2-a15e-196426ec0dcd-nova-migration-ssh-key-1\") pod \"24758178-1be1-49f2-a15e-196426ec0dcd\" (UID: \"24758178-1be1-49f2-a15e-196426ec0dcd\") " Feb 27 08:33:18 crc kubenswrapper[4612]: I0227 08:33:18.233721 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-2\" (UniqueName: \"kubernetes.io/secret/24758178-1be1-49f2-a15e-196426ec0dcd-nova-cell1-compute-config-2\") pod \"24758178-1be1-49f2-a15e-196426ec0dcd\" (UID: \"24758178-1be1-49f2-a15e-196426ec0dcd\") " Feb 27 08:33:18 crc kubenswrapper[4612]: I0227 08:33:18.233744 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/24758178-1be1-49f2-a15e-196426ec0dcd-nova-migration-ssh-key-0\") pod \"24758178-1be1-49f2-a15e-196426ec0dcd\" (UID: \"24758178-1be1-49f2-a15e-196426ec0dcd\") " Feb 27 08:33:18 crc kubenswrapper[4612]: I0227 08:33:18.233783 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/24758178-1be1-49f2-a15e-196426ec0dcd-inventory\") pod \"24758178-1be1-49f2-a15e-196426ec0dcd\" (UID: \"24758178-1be1-49f2-a15e-196426ec0dcd\") " Feb 27 08:33:18 crc kubenswrapper[4612]: I0227 08:33:18.233844 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tfdnl\" (UniqueName: \"kubernetes.io/projected/24758178-1be1-49f2-a15e-196426ec0dcd-kube-api-access-tfdnl\") pod \"24758178-1be1-49f2-a15e-196426ec0dcd\" (UID: \"24758178-1be1-49f2-a15e-196426ec0dcd\") " Feb 27 08:33:18 crc kubenswrapper[4612]: I0227 08:33:18.259018 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/24758178-1be1-49f2-a15e-196426ec0dcd-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "24758178-1be1-49f2-a15e-196426ec0dcd" (UID: "24758178-1be1-49f2-a15e-196426ec0dcd"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:33:18 crc kubenswrapper[4612]: I0227 08:33:18.260074 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/24758178-1be1-49f2-a15e-196426ec0dcd-kube-api-access-tfdnl" (OuterVolumeSpecName: "kube-api-access-tfdnl") pod "24758178-1be1-49f2-a15e-196426ec0dcd" (UID: "24758178-1be1-49f2-a15e-196426ec0dcd"). InnerVolumeSpecName "kube-api-access-tfdnl". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 08:33:18 crc kubenswrapper[4612]: I0227 08:33:18.270579 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/24758178-1be1-49f2-a15e-196426ec0dcd-nova-extra-config-0" (OuterVolumeSpecName: "nova-extra-config-0") pod "24758178-1be1-49f2-a15e-196426ec0dcd" (UID: "24758178-1be1-49f2-a15e-196426ec0dcd"). InnerVolumeSpecName "nova-extra-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 08:33:18 crc kubenswrapper[4612]: I0227 08:33:18.276995 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/24758178-1be1-49f2-a15e-196426ec0dcd-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "24758178-1be1-49f2-a15e-196426ec0dcd" (UID: "24758178-1be1-49f2-a15e-196426ec0dcd"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:33:18 crc kubenswrapper[4612]: I0227 08:33:18.293566 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/24758178-1be1-49f2-a15e-196426ec0dcd-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "24758178-1be1-49f2-a15e-196426ec0dcd" (UID: "24758178-1be1-49f2-a15e-196426ec0dcd"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:33:18 crc kubenswrapper[4612]: I0227 08:33:18.307304 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/24758178-1be1-49f2-a15e-196426ec0dcd-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "24758178-1be1-49f2-a15e-196426ec0dcd" (UID: "24758178-1be1-49f2-a15e-196426ec0dcd"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:33:18 crc kubenswrapper[4612]: I0227 08:33:18.309961 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/24758178-1be1-49f2-a15e-196426ec0dcd-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "24758178-1be1-49f2-a15e-196426ec0dcd" (UID: "24758178-1be1-49f2-a15e-196426ec0dcd"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:33:18 crc kubenswrapper[4612]: I0227 08:33:18.315054 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/24758178-1be1-49f2-a15e-196426ec0dcd-nova-cell1-compute-config-2" (OuterVolumeSpecName: "nova-cell1-compute-config-2") pod "24758178-1be1-49f2-a15e-196426ec0dcd" (UID: "24758178-1be1-49f2-a15e-196426ec0dcd"). InnerVolumeSpecName "nova-cell1-compute-config-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:33:18 crc kubenswrapper[4612]: I0227 08:33:18.319146 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/24758178-1be1-49f2-a15e-196426ec0dcd-nova-cell1-compute-config-3" (OuterVolumeSpecName: "nova-cell1-compute-config-3") pod "24758178-1be1-49f2-a15e-196426ec0dcd" (UID: "24758178-1be1-49f2-a15e-196426ec0dcd"). InnerVolumeSpecName "nova-cell1-compute-config-3". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:33:18 crc kubenswrapper[4612]: I0227 08:33:18.326351 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/24758178-1be1-49f2-a15e-196426ec0dcd-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "24758178-1be1-49f2-a15e-196426ec0dcd" (UID: "24758178-1be1-49f2-a15e-196426ec0dcd"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:33:18 crc kubenswrapper[4612]: I0227 08:33:18.333840 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/24758178-1be1-49f2-a15e-196426ec0dcd-inventory" (OuterVolumeSpecName: "inventory") pod "24758178-1be1-49f2-a15e-196426ec0dcd" (UID: "24758178-1be1-49f2-a15e-196426ec0dcd"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:33:18 crc kubenswrapper[4612]: I0227 08:33:18.335374 4612 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-2\" (UniqueName: \"kubernetes.io/secret/24758178-1be1-49f2-a15e-196426ec0dcd-nova-cell1-compute-config-2\") on node \"crc\" DevicePath \"\"" Feb 27 08:33:18 crc kubenswrapper[4612]: I0227 08:33:18.335414 4612 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/24758178-1be1-49f2-a15e-196426ec0dcd-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Feb 27 08:33:18 crc kubenswrapper[4612]: I0227 08:33:18.335427 4612 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/24758178-1be1-49f2-a15e-196426ec0dcd-inventory\") on node \"crc\" DevicePath \"\"" Feb 27 08:33:18 crc kubenswrapper[4612]: I0227 08:33:18.335439 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tfdnl\" (UniqueName: \"kubernetes.io/projected/24758178-1be1-49f2-a15e-196426ec0dcd-kube-api-access-tfdnl\") on node \"crc\" DevicePath \"\"" Feb 27 08:33:18 crc kubenswrapper[4612]: I0227 08:33:18.335452 4612 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/24758178-1be1-49f2-a15e-196426ec0dcd-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 08:33:18 crc kubenswrapper[4612]: I0227 08:33:18.335466 4612 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/24758178-1be1-49f2-a15e-196426ec0dcd-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Feb 27 08:33:18 crc kubenswrapper[4612]: I0227 08:33:18.335482 4612 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/24758178-1be1-49f2-a15e-196426ec0dcd-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Feb 27 08:33:18 crc kubenswrapper[4612]: I0227 08:33:18.335497 4612 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/24758178-1be1-49f2-a15e-196426ec0dcd-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 27 08:33:18 crc kubenswrapper[4612]: I0227 08:33:18.335510 4612 reconciler_common.go:293] "Volume detached for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/24758178-1be1-49f2-a15e-196426ec0dcd-nova-extra-config-0\") on node \"crc\" DevicePath \"\"" Feb 27 08:33:18 crc kubenswrapper[4612]: I0227 08:33:18.335521 4612 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-3\" (UniqueName: \"kubernetes.io/secret/24758178-1be1-49f2-a15e-196426ec0dcd-nova-cell1-compute-config-3\") on node \"crc\" DevicePath \"\"" Feb 27 08:33:18 crc kubenswrapper[4612]: I0227 08:33:18.335532 4612 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/24758178-1be1-49f2-a15e-196426ec0dcd-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Feb 27 08:33:18 crc kubenswrapper[4612]: I0227 08:33:18.639480 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-bnlnx" event={"ID":"24758178-1be1-49f2-a15e-196426ec0dcd","Type":"ContainerDied","Data":"1bf0f54131813b7e9659d0172113a5e289d8d66b70547296e6ba2a72e414ee9f"} Feb 27 08:33:18 crc kubenswrapper[4612]: I0227 08:33:18.639754 4612 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1bf0f54131813b7e9659d0172113a5e289d8d66b70547296e6ba2a72e414ee9f" Feb 27 08:33:18 crc kubenswrapper[4612]: I0227 08:33:18.639603 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-bnlnx" Feb 27 08:33:18 crc kubenswrapper[4612]: I0227 08:33:18.749455 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qthr6"] Feb 27 08:33:18 crc kubenswrapper[4612]: E0227 08:33:18.749917 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="24758178-1be1-49f2-a15e-196426ec0dcd" containerName="nova-edpm-deployment-openstack-edpm-ipam" Feb 27 08:33:18 crc kubenswrapper[4612]: I0227 08:33:18.749937 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="24758178-1be1-49f2-a15e-196426ec0dcd" containerName="nova-edpm-deployment-openstack-edpm-ipam" Feb 27 08:33:18 crc kubenswrapper[4612]: E0227 08:33:18.749949 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="318ef53b-57e4-4954-867e-d845fc77a484" containerName="oc" Feb 27 08:33:18 crc kubenswrapper[4612]: I0227 08:33:18.749958 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="318ef53b-57e4-4954-867e-d845fc77a484" containerName="oc" Feb 27 08:33:18 crc kubenswrapper[4612]: I0227 08:33:18.750226 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="24758178-1be1-49f2-a15e-196426ec0dcd" containerName="nova-edpm-deployment-openstack-edpm-ipam" Feb 27 08:33:18 crc kubenswrapper[4612]: I0227 08:33:18.750265 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="318ef53b-57e4-4954-867e-d845fc77a484" containerName="oc" Feb 27 08:33:18 crc kubenswrapper[4612]: I0227 08:33:18.750921 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qthr6" Feb 27 08:33:18 crc kubenswrapper[4612]: I0227 08:33:18.761532 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-compute-config-data" Feb 27 08:33:18 crc kubenswrapper[4612]: I0227 08:33:18.761893 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Feb 27 08:33:18 crc kubenswrapper[4612]: I0227 08:33:18.762062 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-2pwrb" Feb 27 08:33:18 crc kubenswrapper[4612]: I0227 08:33:18.762057 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Feb 27 08:33:18 crc kubenswrapper[4612]: I0227 08:33:18.762583 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jl2bw\" (UniqueName: \"kubernetes.io/projected/f69e6377-1d38-4128-b6ad-af91cf8f7793-kube-api-access-jl2bw\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-qthr6\" (UID: \"f69e6377-1d38-4128-b6ad-af91cf8f7793\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qthr6" Feb 27 08:33:18 crc kubenswrapper[4612]: I0227 08:33:18.762622 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/f69e6377-1d38-4128-b6ad-af91cf8f7793-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-qthr6\" (UID: \"f69e6377-1d38-4128-b6ad-af91cf8f7793\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qthr6" Feb 27 08:33:18 crc kubenswrapper[4612]: I0227 08:33:18.762641 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/f69e6377-1d38-4128-b6ad-af91cf8f7793-ssh-key-openstack-edpm-ipam\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-qthr6\" (UID: \"f69e6377-1d38-4128-b6ad-af91cf8f7793\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qthr6" Feb 27 08:33:18 crc kubenswrapper[4612]: I0227 08:33:18.762660 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f69e6377-1d38-4128-b6ad-af91cf8f7793-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-qthr6\" (UID: \"f69e6377-1d38-4128-b6ad-af91cf8f7793\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qthr6" Feb 27 08:33:18 crc kubenswrapper[4612]: I0227 08:33:18.762725 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/f69e6377-1d38-4128-b6ad-af91cf8f7793-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-qthr6\" (UID: \"f69e6377-1d38-4128-b6ad-af91cf8f7793\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qthr6" Feb 27 08:33:18 crc kubenswrapper[4612]: I0227 08:33:18.762786 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/f69e6377-1d38-4128-b6ad-af91cf8f7793-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-qthr6\" (UID: \"f69e6377-1d38-4128-b6ad-af91cf8f7793\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qthr6" Feb 27 08:33:18 crc kubenswrapper[4612]: I0227 08:33:18.762820 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f69e6377-1d38-4128-b6ad-af91cf8f7793-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-qthr6\" (UID: \"f69e6377-1d38-4128-b6ad-af91cf8f7793\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qthr6" Feb 27 08:33:18 crc kubenswrapper[4612]: I0227 08:33:18.762954 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Feb 27 08:33:18 crc kubenswrapper[4612]: I0227 08:33:18.792923 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qthr6"] Feb 27 08:33:18 crc kubenswrapper[4612]: I0227 08:33:18.864040 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/f69e6377-1d38-4128-b6ad-af91cf8f7793-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-qthr6\" (UID: \"f69e6377-1d38-4128-b6ad-af91cf8f7793\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qthr6" Feb 27 08:33:18 crc kubenswrapper[4612]: I0227 08:33:18.864108 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f69e6377-1d38-4128-b6ad-af91cf8f7793-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-qthr6\" (UID: \"f69e6377-1d38-4128-b6ad-af91cf8f7793\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qthr6" Feb 27 08:33:18 crc kubenswrapper[4612]: I0227 08:33:18.864189 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jl2bw\" (UniqueName: \"kubernetes.io/projected/f69e6377-1d38-4128-b6ad-af91cf8f7793-kube-api-access-jl2bw\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-qthr6\" (UID: \"f69e6377-1d38-4128-b6ad-af91cf8f7793\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qthr6" Feb 27 08:33:18 crc kubenswrapper[4612]: I0227 08:33:18.864216 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/f69e6377-1d38-4128-b6ad-af91cf8f7793-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-qthr6\" (UID: \"f69e6377-1d38-4128-b6ad-af91cf8f7793\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qthr6" Feb 27 08:33:18 crc kubenswrapper[4612]: I0227 08:33:18.864255 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/f69e6377-1d38-4128-b6ad-af91cf8f7793-ssh-key-openstack-edpm-ipam\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-qthr6\" (UID: \"f69e6377-1d38-4128-b6ad-af91cf8f7793\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qthr6" Feb 27 08:33:18 crc kubenswrapper[4612]: I0227 08:33:18.864294 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f69e6377-1d38-4128-b6ad-af91cf8f7793-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-qthr6\" (UID: \"f69e6377-1d38-4128-b6ad-af91cf8f7793\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qthr6" Feb 27 08:33:18 crc kubenswrapper[4612]: I0227 08:33:18.864371 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/f69e6377-1d38-4128-b6ad-af91cf8f7793-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-qthr6\" (UID: \"f69e6377-1d38-4128-b6ad-af91cf8f7793\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qthr6" Feb 27 08:33:18 crc kubenswrapper[4612]: I0227 08:33:18.868818 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/f69e6377-1d38-4128-b6ad-af91cf8f7793-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-qthr6\" (UID: \"f69e6377-1d38-4128-b6ad-af91cf8f7793\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qthr6" Feb 27 08:33:18 crc kubenswrapper[4612]: I0227 08:33:18.869324 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/f69e6377-1d38-4128-b6ad-af91cf8f7793-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-qthr6\" (UID: \"f69e6377-1d38-4128-b6ad-af91cf8f7793\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qthr6" Feb 27 08:33:18 crc kubenswrapper[4612]: I0227 08:33:18.870453 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f69e6377-1d38-4128-b6ad-af91cf8f7793-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-qthr6\" (UID: \"f69e6377-1d38-4128-b6ad-af91cf8f7793\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qthr6" Feb 27 08:33:18 crc kubenswrapper[4612]: I0227 08:33:18.870494 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/f69e6377-1d38-4128-b6ad-af91cf8f7793-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-qthr6\" (UID: \"f69e6377-1d38-4128-b6ad-af91cf8f7793\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qthr6" Feb 27 08:33:18 crc kubenswrapper[4612]: I0227 08:33:18.871271 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/f69e6377-1d38-4128-b6ad-af91cf8f7793-ssh-key-openstack-edpm-ipam\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-qthr6\" (UID: \"f69e6377-1d38-4128-b6ad-af91cf8f7793\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qthr6" Feb 27 08:33:18 crc kubenswrapper[4612]: I0227 08:33:18.872171 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f69e6377-1d38-4128-b6ad-af91cf8f7793-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-qthr6\" (UID: \"f69e6377-1d38-4128-b6ad-af91cf8f7793\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qthr6" Feb 27 08:33:18 crc kubenswrapper[4612]: I0227 08:33:18.882456 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jl2bw\" (UniqueName: \"kubernetes.io/projected/f69e6377-1d38-4128-b6ad-af91cf8f7793-kube-api-access-jl2bw\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-qthr6\" (UID: \"f69e6377-1d38-4128-b6ad-af91cf8f7793\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qthr6" Feb 27 08:33:19 crc kubenswrapper[4612]: I0227 08:33:19.072878 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qthr6" Feb 27 08:33:19 crc kubenswrapper[4612]: I0227 08:33:19.691363 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qthr6"] Feb 27 08:33:20 crc kubenswrapper[4612]: I0227 08:33:20.665043 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qthr6" event={"ID":"f69e6377-1d38-4128-b6ad-af91cf8f7793","Type":"ContainerStarted","Data":"7c1ba70c93c91474ac80bc9be188fb1d4b6d5703023259f6d87838d2c096101c"} Feb 27 08:33:20 crc kubenswrapper[4612]: I0227 08:33:20.666660 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qthr6" event={"ID":"f69e6377-1d38-4128-b6ad-af91cf8f7793","Type":"ContainerStarted","Data":"1b5ff7e22385037c7f3214c5f5fa630513f8e13ca03388a1932e01441a1b3e9a"} Feb 27 08:33:20 crc kubenswrapper[4612]: I0227 08:33:20.687898 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qthr6" podStartSLOduration=2.269270601 podStartE2EDuration="2.687879154s" podCreationTimestamp="2026-02-27 08:33:18 +0000 UTC" firstStartedPulling="2026-02-27 08:33:19.706582583 +0000 UTC m=+2657.560512581" lastFinishedPulling="2026-02-27 08:33:20.125191136 +0000 UTC m=+2657.979121134" observedRunningTime="2026-02-27 08:33:20.684866469 +0000 UTC m=+2658.538796477" watchObservedRunningTime="2026-02-27 08:33:20.687879154 +0000 UTC m=+2658.541809172" Feb 27 08:34:00 crc kubenswrapper[4612]: I0227 08:34:00.154744 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29536354-tb4sm"] Feb 27 08:34:00 crc kubenswrapper[4612]: I0227 08:34:00.156902 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536354-tb4sm" Feb 27 08:34:00 crc kubenswrapper[4612]: I0227 08:34:00.159476 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 27 08:34:00 crc kubenswrapper[4612]: I0227 08:34:00.160882 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-5zwfx" Feb 27 08:34:00 crc kubenswrapper[4612]: I0227 08:34:00.162956 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 27 08:34:00 crc kubenswrapper[4612]: I0227 08:34:00.180415 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536354-tb4sm"] Feb 27 08:34:00 crc kubenswrapper[4612]: I0227 08:34:00.335505 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dq7sw\" (UniqueName: \"kubernetes.io/projected/6973ea36-6dbd-4c24-bc4d-04b2180d9614-kube-api-access-dq7sw\") pod \"auto-csr-approver-29536354-tb4sm\" (UID: \"6973ea36-6dbd-4c24-bc4d-04b2180d9614\") " pod="openshift-infra/auto-csr-approver-29536354-tb4sm" Feb 27 08:34:00 crc kubenswrapper[4612]: I0227 08:34:00.438558 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dq7sw\" (UniqueName: \"kubernetes.io/projected/6973ea36-6dbd-4c24-bc4d-04b2180d9614-kube-api-access-dq7sw\") pod \"auto-csr-approver-29536354-tb4sm\" (UID: \"6973ea36-6dbd-4c24-bc4d-04b2180d9614\") " pod="openshift-infra/auto-csr-approver-29536354-tb4sm" Feb 27 08:34:00 crc kubenswrapper[4612]: I0227 08:34:00.473941 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dq7sw\" (UniqueName: \"kubernetes.io/projected/6973ea36-6dbd-4c24-bc4d-04b2180d9614-kube-api-access-dq7sw\") pod \"auto-csr-approver-29536354-tb4sm\" (UID: \"6973ea36-6dbd-4c24-bc4d-04b2180d9614\") " pod="openshift-infra/auto-csr-approver-29536354-tb4sm" Feb 27 08:34:00 crc kubenswrapper[4612]: I0227 08:34:00.480816 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536354-tb4sm" Feb 27 08:34:00 crc kubenswrapper[4612]: I0227 08:34:00.965370 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536354-tb4sm"] Feb 27 08:34:01 crc kubenswrapper[4612]: I0227 08:34:01.075470 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536354-tb4sm" event={"ID":"6973ea36-6dbd-4c24-bc4d-04b2180d9614","Type":"ContainerStarted","Data":"a226c71af95ce0ff03c487bfca4dcd72a6152c87077420bac12bed77e22687cf"} Feb 27 08:34:03 crc kubenswrapper[4612]: I0227 08:34:03.101529 4612 generic.go:334] "Generic (PLEG): container finished" podID="6973ea36-6dbd-4c24-bc4d-04b2180d9614" containerID="e537681481b958e073c9bb153dc93414b2a60df2afb30af474d132c00ab6819e" exitCode=0 Feb 27 08:34:03 crc kubenswrapper[4612]: I0227 08:34:03.101619 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536354-tb4sm" event={"ID":"6973ea36-6dbd-4c24-bc4d-04b2180d9614","Type":"ContainerDied","Data":"e537681481b958e073c9bb153dc93414b2a60df2afb30af474d132c00ab6819e"} Feb 27 08:34:04 crc kubenswrapper[4612]: I0227 08:34:04.529757 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536354-tb4sm" Feb 27 08:34:04 crc kubenswrapper[4612]: I0227 08:34:04.639656 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dq7sw\" (UniqueName: \"kubernetes.io/projected/6973ea36-6dbd-4c24-bc4d-04b2180d9614-kube-api-access-dq7sw\") pod \"6973ea36-6dbd-4c24-bc4d-04b2180d9614\" (UID: \"6973ea36-6dbd-4c24-bc4d-04b2180d9614\") " Feb 27 08:34:04 crc kubenswrapper[4612]: I0227 08:34:04.646209 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6973ea36-6dbd-4c24-bc4d-04b2180d9614-kube-api-access-dq7sw" (OuterVolumeSpecName: "kube-api-access-dq7sw") pod "6973ea36-6dbd-4c24-bc4d-04b2180d9614" (UID: "6973ea36-6dbd-4c24-bc4d-04b2180d9614"). InnerVolumeSpecName "kube-api-access-dq7sw". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 08:34:04 crc kubenswrapper[4612]: I0227 08:34:04.742496 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dq7sw\" (UniqueName: \"kubernetes.io/projected/6973ea36-6dbd-4c24-bc4d-04b2180d9614-kube-api-access-dq7sw\") on node \"crc\" DevicePath \"\"" Feb 27 08:34:05 crc kubenswrapper[4612]: I0227 08:34:05.122189 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536354-tb4sm" event={"ID":"6973ea36-6dbd-4c24-bc4d-04b2180d9614","Type":"ContainerDied","Data":"a226c71af95ce0ff03c487bfca4dcd72a6152c87077420bac12bed77e22687cf"} Feb 27 08:34:05 crc kubenswrapper[4612]: I0227 08:34:05.122511 4612 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a226c71af95ce0ff03c487bfca4dcd72a6152c87077420bac12bed77e22687cf" Feb 27 08:34:05 crc kubenswrapper[4612]: I0227 08:34:05.122309 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536354-tb4sm" Feb 27 08:34:05 crc kubenswrapper[4612]: I0227 08:34:05.610088 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29536348-gmk9c"] Feb 27 08:34:05 crc kubenswrapper[4612]: I0227 08:34:05.618718 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29536348-gmk9c"] Feb 27 08:34:06 crc kubenswrapper[4612]: I0227 08:34:06.866610 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="28295adb-7038-43da-a687-744152692bd2" path="/var/lib/kubelet/pods/28295adb-7038-43da-a687-744152692bd2/volumes" Feb 27 08:34:16 crc kubenswrapper[4612]: I0227 08:34:16.027409 4612 patch_prober.go:28] interesting pod/machine-config-daemon-924vb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 27 08:34:16 crc kubenswrapper[4612]: I0227 08:34:16.029600 4612 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 27 08:34:18 crc kubenswrapper[4612]: I0227 08:34:18.072525 4612 scope.go:117] "RemoveContainer" containerID="3521c44e78318294619f5e8a6c30f85dfb9c42064e432e1594c779a464a39199" Feb 27 08:34:46 crc kubenswrapper[4612]: I0227 08:34:46.027570 4612 patch_prober.go:28] interesting pod/machine-config-daemon-924vb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 27 08:34:46 crc kubenswrapper[4612]: I0227 08:34:46.028166 4612 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 27 08:34:54 crc kubenswrapper[4612]: I0227 08:34:54.057601 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-c6mx6"] Feb 27 08:34:54 crc kubenswrapper[4612]: E0227 08:34:54.065716 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6973ea36-6dbd-4c24-bc4d-04b2180d9614" containerName="oc" Feb 27 08:34:54 crc kubenswrapper[4612]: I0227 08:34:54.065936 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="6973ea36-6dbd-4c24-bc4d-04b2180d9614" containerName="oc" Feb 27 08:34:54 crc kubenswrapper[4612]: I0227 08:34:54.066154 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="6973ea36-6dbd-4c24-bc4d-04b2180d9614" containerName="oc" Feb 27 08:34:54 crc kubenswrapper[4612]: I0227 08:34:54.069026 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-c6mx6" Feb 27 08:34:54 crc kubenswrapper[4612]: I0227 08:34:54.076351 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-c6mx6"] Feb 27 08:34:54 crc kubenswrapper[4612]: I0227 08:34:54.171478 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/63a0ed0a-40a9-4250-b0f7-f23bd25675f8-catalog-content\") pod \"redhat-marketplace-c6mx6\" (UID: \"63a0ed0a-40a9-4250-b0f7-f23bd25675f8\") " pod="openshift-marketplace/redhat-marketplace-c6mx6" Feb 27 08:34:54 crc kubenswrapper[4612]: I0227 08:34:54.171662 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/63a0ed0a-40a9-4250-b0f7-f23bd25675f8-utilities\") pod \"redhat-marketplace-c6mx6\" (UID: \"63a0ed0a-40a9-4250-b0f7-f23bd25675f8\") " pod="openshift-marketplace/redhat-marketplace-c6mx6" Feb 27 08:34:54 crc kubenswrapper[4612]: I0227 08:34:54.171788 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sf9nb\" (UniqueName: \"kubernetes.io/projected/63a0ed0a-40a9-4250-b0f7-f23bd25675f8-kube-api-access-sf9nb\") pod \"redhat-marketplace-c6mx6\" (UID: \"63a0ed0a-40a9-4250-b0f7-f23bd25675f8\") " pod="openshift-marketplace/redhat-marketplace-c6mx6" Feb 27 08:34:54 crc kubenswrapper[4612]: I0227 08:34:54.276910 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/63a0ed0a-40a9-4250-b0f7-f23bd25675f8-utilities\") pod \"redhat-marketplace-c6mx6\" (UID: \"63a0ed0a-40a9-4250-b0f7-f23bd25675f8\") " pod="openshift-marketplace/redhat-marketplace-c6mx6" Feb 27 08:34:54 crc kubenswrapper[4612]: I0227 08:34:54.276999 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sf9nb\" (UniqueName: \"kubernetes.io/projected/63a0ed0a-40a9-4250-b0f7-f23bd25675f8-kube-api-access-sf9nb\") pod \"redhat-marketplace-c6mx6\" (UID: \"63a0ed0a-40a9-4250-b0f7-f23bd25675f8\") " pod="openshift-marketplace/redhat-marketplace-c6mx6" Feb 27 08:34:54 crc kubenswrapper[4612]: I0227 08:34:54.277046 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/63a0ed0a-40a9-4250-b0f7-f23bd25675f8-catalog-content\") pod \"redhat-marketplace-c6mx6\" (UID: \"63a0ed0a-40a9-4250-b0f7-f23bd25675f8\") " pod="openshift-marketplace/redhat-marketplace-c6mx6" Feb 27 08:34:54 crc kubenswrapper[4612]: I0227 08:34:54.277400 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/63a0ed0a-40a9-4250-b0f7-f23bd25675f8-utilities\") pod \"redhat-marketplace-c6mx6\" (UID: \"63a0ed0a-40a9-4250-b0f7-f23bd25675f8\") " pod="openshift-marketplace/redhat-marketplace-c6mx6" Feb 27 08:34:54 crc kubenswrapper[4612]: I0227 08:34:54.277536 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/63a0ed0a-40a9-4250-b0f7-f23bd25675f8-catalog-content\") pod \"redhat-marketplace-c6mx6\" (UID: \"63a0ed0a-40a9-4250-b0f7-f23bd25675f8\") " pod="openshift-marketplace/redhat-marketplace-c6mx6" Feb 27 08:34:54 crc kubenswrapper[4612]: I0227 08:34:54.307966 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sf9nb\" (UniqueName: \"kubernetes.io/projected/63a0ed0a-40a9-4250-b0f7-f23bd25675f8-kube-api-access-sf9nb\") pod \"redhat-marketplace-c6mx6\" (UID: \"63a0ed0a-40a9-4250-b0f7-f23bd25675f8\") " pod="openshift-marketplace/redhat-marketplace-c6mx6" Feb 27 08:34:54 crc kubenswrapper[4612]: I0227 08:34:54.393108 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-c6mx6" Feb 27 08:34:54 crc kubenswrapper[4612]: I0227 08:34:54.944441 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-c6mx6"] Feb 27 08:34:55 crc kubenswrapper[4612]: I0227 08:34:55.662515 4612 generic.go:334] "Generic (PLEG): container finished" podID="63a0ed0a-40a9-4250-b0f7-f23bd25675f8" containerID="65cbc37f7c0535cc764f4dba36c2953e2e93691d6d58173c3b59e0c40732f13d" exitCode=0 Feb 27 08:34:55 crc kubenswrapper[4612]: I0227 08:34:55.663002 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c6mx6" event={"ID":"63a0ed0a-40a9-4250-b0f7-f23bd25675f8","Type":"ContainerDied","Data":"65cbc37f7c0535cc764f4dba36c2953e2e93691d6d58173c3b59e0c40732f13d"} Feb 27 08:34:55 crc kubenswrapper[4612]: I0227 08:34:55.663057 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c6mx6" event={"ID":"63a0ed0a-40a9-4250-b0f7-f23bd25675f8","Type":"ContainerStarted","Data":"01bb8ea89d0c377233b264f5901be44739615c4a9a426e0541a1e768fa16f033"} Feb 27 08:34:57 crc kubenswrapper[4612]: I0227 08:34:57.682662 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c6mx6" event={"ID":"63a0ed0a-40a9-4250-b0f7-f23bd25675f8","Type":"ContainerStarted","Data":"38a79b07e0b2f09fa963cffb2514a3ab2b016c8ff6f800b47980693da8c5140e"} Feb 27 08:34:58 crc kubenswrapper[4612]: I0227 08:34:58.697813 4612 generic.go:334] "Generic (PLEG): container finished" podID="63a0ed0a-40a9-4250-b0f7-f23bd25675f8" containerID="38a79b07e0b2f09fa963cffb2514a3ab2b016c8ff6f800b47980693da8c5140e" exitCode=0 Feb 27 08:34:58 crc kubenswrapper[4612]: I0227 08:34:58.698013 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c6mx6" event={"ID":"63a0ed0a-40a9-4250-b0f7-f23bd25675f8","Type":"ContainerDied","Data":"38a79b07e0b2f09fa963cffb2514a3ab2b016c8ff6f800b47980693da8c5140e"} Feb 27 08:34:59 crc kubenswrapper[4612]: I0227 08:34:59.713378 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c6mx6" event={"ID":"63a0ed0a-40a9-4250-b0f7-f23bd25675f8","Type":"ContainerStarted","Data":"a2a6645693d8bcbcd7e27a85b4f6ec35e31795abc2bd356418457a62a8704dcd"} Feb 27 08:34:59 crc kubenswrapper[4612]: I0227 08:34:59.744497 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-c6mx6" podStartSLOduration=2.274356583 podStartE2EDuration="5.744472191s" podCreationTimestamp="2026-02-27 08:34:54 +0000 UTC" firstStartedPulling="2026-02-27 08:34:55.66547697 +0000 UTC m=+2753.519407008" lastFinishedPulling="2026-02-27 08:34:59.135592618 +0000 UTC m=+2756.989522616" observedRunningTime="2026-02-27 08:34:59.736923826 +0000 UTC m=+2757.590853844" watchObservedRunningTime="2026-02-27 08:34:59.744472191 +0000 UTC m=+2757.598402179" Feb 27 08:35:04 crc kubenswrapper[4612]: I0227 08:35:04.394443 4612 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-c6mx6" Feb 27 08:35:04 crc kubenswrapper[4612]: I0227 08:35:04.395071 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-c6mx6" Feb 27 08:35:04 crc kubenswrapper[4612]: I0227 08:35:04.440293 4612 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-c6mx6" Feb 27 08:35:04 crc kubenswrapper[4612]: I0227 08:35:04.828973 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-c6mx6" Feb 27 08:35:04 crc kubenswrapper[4612]: I0227 08:35:04.888016 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-c6mx6"] Feb 27 08:35:06 crc kubenswrapper[4612]: I0227 08:35:06.791715 4612 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-c6mx6" podUID="63a0ed0a-40a9-4250-b0f7-f23bd25675f8" containerName="registry-server" containerID="cri-o://a2a6645693d8bcbcd7e27a85b4f6ec35e31795abc2bd356418457a62a8704dcd" gracePeriod=2 Feb 27 08:35:07 crc kubenswrapper[4612]: I0227 08:35:07.250061 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-c6mx6" Feb 27 08:35:07 crc kubenswrapper[4612]: I0227 08:35:07.421241 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sf9nb\" (UniqueName: \"kubernetes.io/projected/63a0ed0a-40a9-4250-b0f7-f23bd25675f8-kube-api-access-sf9nb\") pod \"63a0ed0a-40a9-4250-b0f7-f23bd25675f8\" (UID: \"63a0ed0a-40a9-4250-b0f7-f23bd25675f8\") " Feb 27 08:35:07 crc kubenswrapper[4612]: I0227 08:35:07.421346 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/63a0ed0a-40a9-4250-b0f7-f23bd25675f8-catalog-content\") pod \"63a0ed0a-40a9-4250-b0f7-f23bd25675f8\" (UID: \"63a0ed0a-40a9-4250-b0f7-f23bd25675f8\") " Feb 27 08:35:07 crc kubenswrapper[4612]: I0227 08:35:07.421392 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/63a0ed0a-40a9-4250-b0f7-f23bd25675f8-utilities\") pod \"63a0ed0a-40a9-4250-b0f7-f23bd25675f8\" (UID: \"63a0ed0a-40a9-4250-b0f7-f23bd25675f8\") " Feb 27 08:35:07 crc kubenswrapper[4612]: I0227 08:35:07.423190 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/63a0ed0a-40a9-4250-b0f7-f23bd25675f8-utilities" (OuterVolumeSpecName: "utilities") pod "63a0ed0a-40a9-4250-b0f7-f23bd25675f8" (UID: "63a0ed0a-40a9-4250-b0f7-f23bd25675f8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 08:35:07 crc kubenswrapper[4612]: I0227 08:35:07.433036 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/63a0ed0a-40a9-4250-b0f7-f23bd25675f8-kube-api-access-sf9nb" (OuterVolumeSpecName: "kube-api-access-sf9nb") pod "63a0ed0a-40a9-4250-b0f7-f23bd25675f8" (UID: "63a0ed0a-40a9-4250-b0f7-f23bd25675f8"). InnerVolumeSpecName "kube-api-access-sf9nb". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 08:35:07 crc kubenswrapper[4612]: I0227 08:35:07.447665 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/63a0ed0a-40a9-4250-b0f7-f23bd25675f8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "63a0ed0a-40a9-4250-b0f7-f23bd25675f8" (UID: "63a0ed0a-40a9-4250-b0f7-f23bd25675f8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 08:35:07 crc kubenswrapper[4612]: I0227 08:35:07.523267 4612 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/63a0ed0a-40a9-4250-b0f7-f23bd25675f8-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 27 08:35:07 crc kubenswrapper[4612]: I0227 08:35:07.523297 4612 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/63a0ed0a-40a9-4250-b0f7-f23bd25675f8-utilities\") on node \"crc\" DevicePath \"\"" Feb 27 08:35:07 crc kubenswrapper[4612]: I0227 08:35:07.523308 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sf9nb\" (UniqueName: \"kubernetes.io/projected/63a0ed0a-40a9-4250-b0f7-f23bd25675f8-kube-api-access-sf9nb\") on node \"crc\" DevicePath \"\"" Feb 27 08:35:07 crc kubenswrapper[4612]: I0227 08:35:07.804752 4612 generic.go:334] "Generic (PLEG): container finished" podID="63a0ed0a-40a9-4250-b0f7-f23bd25675f8" containerID="a2a6645693d8bcbcd7e27a85b4f6ec35e31795abc2bd356418457a62a8704dcd" exitCode=0 Feb 27 08:35:07 crc kubenswrapper[4612]: I0227 08:35:07.804809 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c6mx6" event={"ID":"63a0ed0a-40a9-4250-b0f7-f23bd25675f8","Type":"ContainerDied","Data":"a2a6645693d8bcbcd7e27a85b4f6ec35e31795abc2bd356418457a62a8704dcd"} Feb 27 08:35:07 crc kubenswrapper[4612]: I0227 08:35:07.805154 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c6mx6" event={"ID":"63a0ed0a-40a9-4250-b0f7-f23bd25675f8","Type":"ContainerDied","Data":"01bb8ea89d0c377233b264f5901be44739615c4a9a426e0541a1e768fa16f033"} Feb 27 08:35:07 crc kubenswrapper[4612]: I0227 08:35:07.805180 4612 scope.go:117] "RemoveContainer" containerID="a2a6645693d8bcbcd7e27a85b4f6ec35e31795abc2bd356418457a62a8704dcd" Feb 27 08:35:07 crc kubenswrapper[4612]: I0227 08:35:07.804837 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-c6mx6" Feb 27 08:35:07 crc kubenswrapper[4612]: I0227 08:35:07.831332 4612 scope.go:117] "RemoveContainer" containerID="38a79b07e0b2f09fa963cffb2514a3ab2b016c8ff6f800b47980693da8c5140e" Feb 27 08:35:07 crc kubenswrapper[4612]: I0227 08:35:07.844116 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-c6mx6"] Feb 27 08:35:07 crc kubenswrapper[4612]: I0227 08:35:07.853843 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-c6mx6"] Feb 27 08:35:07 crc kubenswrapper[4612]: I0227 08:35:07.857253 4612 scope.go:117] "RemoveContainer" containerID="65cbc37f7c0535cc764f4dba36c2953e2e93691d6d58173c3b59e0c40732f13d" Feb 27 08:35:07 crc kubenswrapper[4612]: I0227 08:35:07.895485 4612 scope.go:117] "RemoveContainer" containerID="a2a6645693d8bcbcd7e27a85b4f6ec35e31795abc2bd356418457a62a8704dcd" Feb 27 08:35:07 crc kubenswrapper[4612]: E0227 08:35:07.896258 4612 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a2a6645693d8bcbcd7e27a85b4f6ec35e31795abc2bd356418457a62a8704dcd\": container with ID starting with a2a6645693d8bcbcd7e27a85b4f6ec35e31795abc2bd356418457a62a8704dcd not found: ID does not exist" containerID="a2a6645693d8bcbcd7e27a85b4f6ec35e31795abc2bd356418457a62a8704dcd" Feb 27 08:35:07 crc kubenswrapper[4612]: I0227 08:35:07.896302 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a2a6645693d8bcbcd7e27a85b4f6ec35e31795abc2bd356418457a62a8704dcd"} err="failed to get container status \"a2a6645693d8bcbcd7e27a85b4f6ec35e31795abc2bd356418457a62a8704dcd\": rpc error: code = NotFound desc = could not find container \"a2a6645693d8bcbcd7e27a85b4f6ec35e31795abc2bd356418457a62a8704dcd\": container with ID starting with a2a6645693d8bcbcd7e27a85b4f6ec35e31795abc2bd356418457a62a8704dcd not found: ID does not exist" Feb 27 08:35:07 crc kubenswrapper[4612]: I0227 08:35:07.896332 4612 scope.go:117] "RemoveContainer" containerID="38a79b07e0b2f09fa963cffb2514a3ab2b016c8ff6f800b47980693da8c5140e" Feb 27 08:35:07 crc kubenswrapper[4612]: E0227 08:35:07.896660 4612 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"38a79b07e0b2f09fa963cffb2514a3ab2b016c8ff6f800b47980693da8c5140e\": container with ID starting with 38a79b07e0b2f09fa963cffb2514a3ab2b016c8ff6f800b47980693da8c5140e not found: ID does not exist" containerID="38a79b07e0b2f09fa963cffb2514a3ab2b016c8ff6f800b47980693da8c5140e" Feb 27 08:35:07 crc kubenswrapper[4612]: I0227 08:35:07.896681 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"38a79b07e0b2f09fa963cffb2514a3ab2b016c8ff6f800b47980693da8c5140e"} err="failed to get container status \"38a79b07e0b2f09fa963cffb2514a3ab2b016c8ff6f800b47980693da8c5140e\": rpc error: code = NotFound desc = could not find container \"38a79b07e0b2f09fa963cffb2514a3ab2b016c8ff6f800b47980693da8c5140e\": container with ID starting with 38a79b07e0b2f09fa963cffb2514a3ab2b016c8ff6f800b47980693da8c5140e not found: ID does not exist" Feb 27 08:35:07 crc kubenswrapper[4612]: I0227 08:35:07.896719 4612 scope.go:117] "RemoveContainer" containerID="65cbc37f7c0535cc764f4dba36c2953e2e93691d6d58173c3b59e0c40732f13d" Feb 27 08:35:07 crc kubenswrapper[4612]: E0227 08:35:07.897134 4612 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"65cbc37f7c0535cc764f4dba36c2953e2e93691d6d58173c3b59e0c40732f13d\": container with ID starting with 65cbc37f7c0535cc764f4dba36c2953e2e93691d6d58173c3b59e0c40732f13d not found: ID does not exist" containerID="65cbc37f7c0535cc764f4dba36c2953e2e93691d6d58173c3b59e0c40732f13d" Feb 27 08:35:07 crc kubenswrapper[4612]: I0227 08:35:07.897195 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"65cbc37f7c0535cc764f4dba36c2953e2e93691d6d58173c3b59e0c40732f13d"} err="failed to get container status \"65cbc37f7c0535cc764f4dba36c2953e2e93691d6d58173c3b59e0c40732f13d\": rpc error: code = NotFound desc = could not find container \"65cbc37f7c0535cc764f4dba36c2953e2e93691d6d58173c3b59e0c40732f13d\": container with ID starting with 65cbc37f7c0535cc764f4dba36c2953e2e93691d6d58173c3b59e0c40732f13d not found: ID does not exist" Feb 27 08:35:08 crc kubenswrapper[4612]: I0227 08:35:08.863188 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="63a0ed0a-40a9-4250-b0f7-f23bd25675f8" path="/var/lib/kubelet/pods/63a0ed0a-40a9-4250-b0f7-f23bd25675f8/volumes" Feb 27 08:35:12 crc kubenswrapper[4612]: I0227 08:35:12.938068 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-gdv97"] Feb 27 08:35:12 crc kubenswrapper[4612]: E0227 08:35:12.939180 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63a0ed0a-40a9-4250-b0f7-f23bd25675f8" containerName="extract-utilities" Feb 27 08:35:12 crc kubenswrapper[4612]: I0227 08:35:12.939199 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="63a0ed0a-40a9-4250-b0f7-f23bd25675f8" containerName="extract-utilities" Feb 27 08:35:12 crc kubenswrapper[4612]: E0227 08:35:12.939225 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63a0ed0a-40a9-4250-b0f7-f23bd25675f8" containerName="registry-server" Feb 27 08:35:12 crc kubenswrapper[4612]: I0227 08:35:12.939233 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="63a0ed0a-40a9-4250-b0f7-f23bd25675f8" containerName="registry-server" Feb 27 08:35:12 crc kubenswrapper[4612]: E0227 08:35:12.939255 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63a0ed0a-40a9-4250-b0f7-f23bd25675f8" containerName="extract-content" Feb 27 08:35:12 crc kubenswrapper[4612]: I0227 08:35:12.939264 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="63a0ed0a-40a9-4250-b0f7-f23bd25675f8" containerName="extract-content" Feb 27 08:35:12 crc kubenswrapper[4612]: I0227 08:35:12.939483 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="63a0ed0a-40a9-4250-b0f7-f23bd25675f8" containerName="registry-server" Feb 27 08:35:12 crc kubenswrapper[4612]: I0227 08:35:12.941133 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-gdv97" Feb 27 08:35:12 crc kubenswrapper[4612]: I0227 08:35:12.964684 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-gdv97"] Feb 27 08:35:13 crc kubenswrapper[4612]: I0227 08:35:13.035432 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nmdm2\" (UniqueName: \"kubernetes.io/projected/65d07354-a95f-4a47-a4df-bd5cb66814b2-kube-api-access-nmdm2\") pod \"redhat-operators-gdv97\" (UID: \"65d07354-a95f-4a47-a4df-bd5cb66814b2\") " pod="openshift-marketplace/redhat-operators-gdv97" Feb 27 08:35:13 crc kubenswrapper[4612]: I0227 08:35:13.035565 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/65d07354-a95f-4a47-a4df-bd5cb66814b2-utilities\") pod \"redhat-operators-gdv97\" (UID: \"65d07354-a95f-4a47-a4df-bd5cb66814b2\") " pod="openshift-marketplace/redhat-operators-gdv97" Feb 27 08:35:13 crc kubenswrapper[4612]: I0227 08:35:13.035712 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/65d07354-a95f-4a47-a4df-bd5cb66814b2-catalog-content\") pod \"redhat-operators-gdv97\" (UID: \"65d07354-a95f-4a47-a4df-bd5cb66814b2\") " pod="openshift-marketplace/redhat-operators-gdv97" Feb 27 08:35:13 crc kubenswrapper[4612]: I0227 08:35:13.137133 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/65d07354-a95f-4a47-a4df-bd5cb66814b2-utilities\") pod \"redhat-operators-gdv97\" (UID: \"65d07354-a95f-4a47-a4df-bd5cb66814b2\") " pod="openshift-marketplace/redhat-operators-gdv97" Feb 27 08:35:13 crc kubenswrapper[4612]: I0227 08:35:13.137240 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/65d07354-a95f-4a47-a4df-bd5cb66814b2-catalog-content\") pod \"redhat-operators-gdv97\" (UID: \"65d07354-a95f-4a47-a4df-bd5cb66814b2\") " pod="openshift-marketplace/redhat-operators-gdv97" Feb 27 08:35:13 crc kubenswrapper[4612]: I0227 08:35:13.137294 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nmdm2\" (UniqueName: \"kubernetes.io/projected/65d07354-a95f-4a47-a4df-bd5cb66814b2-kube-api-access-nmdm2\") pod \"redhat-operators-gdv97\" (UID: \"65d07354-a95f-4a47-a4df-bd5cb66814b2\") " pod="openshift-marketplace/redhat-operators-gdv97" Feb 27 08:35:13 crc kubenswrapper[4612]: I0227 08:35:13.137777 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/65d07354-a95f-4a47-a4df-bd5cb66814b2-utilities\") pod \"redhat-operators-gdv97\" (UID: \"65d07354-a95f-4a47-a4df-bd5cb66814b2\") " pod="openshift-marketplace/redhat-operators-gdv97" Feb 27 08:35:13 crc kubenswrapper[4612]: I0227 08:35:13.138007 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/65d07354-a95f-4a47-a4df-bd5cb66814b2-catalog-content\") pod \"redhat-operators-gdv97\" (UID: \"65d07354-a95f-4a47-a4df-bd5cb66814b2\") " pod="openshift-marketplace/redhat-operators-gdv97" Feb 27 08:35:13 crc kubenswrapper[4612]: I0227 08:35:13.155238 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nmdm2\" (UniqueName: \"kubernetes.io/projected/65d07354-a95f-4a47-a4df-bd5cb66814b2-kube-api-access-nmdm2\") pod \"redhat-operators-gdv97\" (UID: \"65d07354-a95f-4a47-a4df-bd5cb66814b2\") " pod="openshift-marketplace/redhat-operators-gdv97" Feb 27 08:35:13 crc kubenswrapper[4612]: I0227 08:35:13.260791 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-gdv97" Feb 27 08:35:13 crc kubenswrapper[4612]: I0227 08:35:13.742960 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-gdv97"] Feb 27 08:35:13 crc kubenswrapper[4612]: I0227 08:35:13.855475 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gdv97" event={"ID":"65d07354-a95f-4a47-a4df-bd5cb66814b2","Type":"ContainerStarted","Data":"c62584d072e40452df7d5acb90275cb93b8a8946815c73c559b506a8008f2c81"} Feb 27 08:35:14 crc kubenswrapper[4612]: I0227 08:35:14.864807 4612 generic.go:334] "Generic (PLEG): container finished" podID="65d07354-a95f-4a47-a4df-bd5cb66814b2" containerID="f212e679fd155e68767180732b06409c017f70ee0527988dc40fcf1e1abf1c68" exitCode=0 Feb 27 08:35:14 crc kubenswrapper[4612]: I0227 08:35:14.864913 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gdv97" event={"ID":"65d07354-a95f-4a47-a4df-bd5cb66814b2","Type":"ContainerDied","Data":"f212e679fd155e68767180732b06409c017f70ee0527988dc40fcf1e1abf1c68"} Feb 27 08:35:14 crc kubenswrapper[4612]: I0227 08:35:14.867758 4612 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 27 08:35:15 crc kubenswrapper[4612]: I0227 08:35:15.874363 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gdv97" event={"ID":"65d07354-a95f-4a47-a4df-bd5cb66814b2","Type":"ContainerStarted","Data":"b9f5584d493c78e63fd469606cdd96b1b2490e974a491b170611420536233f78"} Feb 27 08:35:16 crc kubenswrapper[4612]: I0227 08:35:16.026907 4612 patch_prober.go:28] interesting pod/machine-config-daemon-924vb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 27 08:35:16 crc kubenswrapper[4612]: I0227 08:35:16.026970 4612 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 27 08:35:16 crc kubenswrapper[4612]: I0227 08:35:16.027020 4612 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-924vb" Feb 27 08:35:16 crc kubenswrapper[4612]: I0227 08:35:16.027807 4612 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"542804ac84a667bc92e9df5b1d69c8ad7e48579a796d2b39142a8890f0acd72f"} pod="openshift-machine-config-operator/machine-config-daemon-924vb" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 27 08:35:16 crc kubenswrapper[4612]: I0227 08:35:16.027884 4612 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" containerName="machine-config-daemon" containerID="cri-o://542804ac84a667bc92e9df5b1d69c8ad7e48579a796d2b39142a8890f0acd72f" gracePeriod=600 Feb 27 08:35:16 crc kubenswrapper[4612]: I0227 08:35:16.889079 4612 generic.go:334] "Generic (PLEG): container finished" podID="24599373-7adc-4a1b-8bb4-797bf726f43d" containerID="542804ac84a667bc92e9df5b1d69c8ad7e48579a796d2b39142a8890f0acd72f" exitCode=0 Feb 27 08:35:16 crc kubenswrapper[4612]: I0227 08:35:16.889156 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-924vb" event={"ID":"24599373-7adc-4a1b-8bb4-797bf726f43d","Type":"ContainerDied","Data":"542804ac84a667bc92e9df5b1d69c8ad7e48579a796d2b39142a8890f0acd72f"} Feb 27 08:35:16 crc kubenswrapper[4612]: I0227 08:35:16.889566 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-924vb" event={"ID":"24599373-7adc-4a1b-8bb4-797bf726f43d","Type":"ContainerStarted","Data":"384f63ec5e89debe3f33788a53a38ede9872a5e515bb3ecc12869cd886606dba"} Feb 27 08:35:16 crc kubenswrapper[4612]: I0227 08:35:16.889598 4612 scope.go:117] "RemoveContainer" containerID="76c2de0055c53fc536f48b64de557ea4f7b56a2de457f8f1b8708b99c7c612cb" Feb 27 08:35:22 crc kubenswrapper[4612]: I0227 08:35:22.951770 4612 generic.go:334] "Generic (PLEG): container finished" podID="65d07354-a95f-4a47-a4df-bd5cb66814b2" containerID="b9f5584d493c78e63fd469606cdd96b1b2490e974a491b170611420536233f78" exitCode=0 Feb 27 08:35:22 crc kubenswrapper[4612]: I0227 08:35:22.952293 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gdv97" event={"ID":"65d07354-a95f-4a47-a4df-bd5cb66814b2","Type":"ContainerDied","Data":"b9f5584d493c78e63fd469606cdd96b1b2490e974a491b170611420536233f78"} Feb 27 08:35:22 crc kubenswrapper[4612]: E0227 08:35:22.969959 4612 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod65d07354_a95f_4a47_a4df_bd5cb66814b2.slice/crio-b9f5584d493c78e63fd469606cdd96b1b2490e974a491b170611420536233f78.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod65d07354_a95f_4a47_a4df_bd5cb66814b2.slice/crio-conmon-b9f5584d493c78e63fd469606cdd96b1b2490e974a491b170611420536233f78.scope\": RecentStats: unable to find data in memory cache]" Feb 27 08:35:23 crc kubenswrapper[4612]: I0227 08:35:23.964663 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gdv97" event={"ID":"65d07354-a95f-4a47-a4df-bd5cb66814b2","Type":"ContainerStarted","Data":"ebfd7e4c78fc97741adbfdd7a222bfc0bf0b3f788355b29312d40b6b4e3fdb48"} Feb 27 08:35:23 crc kubenswrapper[4612]: I0227 08:35:23.986855 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-gdv97" podStartSLOduration=3.263532883 podStartE2EDuration="11.986838617s" podCreationTimestamp="2026-02-27 08:35:12 +0000 UTC" firstStartedPulling="2026-02-27 08:35:14.867314047 +0000 UTC m=+2772.721244045" lastFinishedPulling="2026-02-27 08:35:23.590619771 +0000 UTC m=+2781.444549779" observedRunningTime="2026-02-27 08:35:23.986349063 +0000 UTC m=+2781.840279081" watchObservedRunningTime="2026-02-27 08:35:23.986838617 +0000 UTC m=+2781.840768615" Feb 27 08:35:33 crc kubenswrapper[4612]: I0227 08:35:33.261572 4612 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-gdv97" Feb 27 08:35:33 crc kubenswrapper[4612]: I0227 08:35:33.262123 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-gdv97" Feb 27 08:35:34 crc kubenswrapper[4612]: I0227 08:35:34.325852 4612 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-gdv97" podUID="65d07354-a95f-4a47-a4df-bd5cb66814b2" containerName="registry-server" probeResult="failure" output=< Feb 27 08:35:34 crc kubenswrapper[4612]: timeout: failed to connect service ":50051" within 1s Feb 27 08:35:34 crc kubenswrapper[4612]: > Feb 27 08:35:44 crc kubenswrapper[4612]: I0227 08:35:44.302303 4612 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-gdv97" podUID="65d07354-a95f-4a47-a4df-bd5cb66814b2" containerName="registry-server" probeResult="failure" output=< Feb 27 08:35:44 crc kubenswrapper[4612]: timeout: failed to connect service ":50051" within 1s Feb 27 08:35:44 crc kubenswrapper[4612]: > Feb 27 08:35:54 crc kubenswrapper[4612]: I0227 08:35:54.310296 4612 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-gdv97" podUID="65d07354-a95f-4a47-a4df-bd5cb66814b2" containerName="registry-server" probeResult="failure" output=< Feb 27 08:35:54 crc kubenswrapper[4612]: timeout: failed to connect service ":50051" within 1s Feb 27 08:35:54 crc kubenswrapper[4612]: > Feb 27 08:36:00 crc kubenswrapper[4612]: I0227 08:36:00.143797 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29536356-txvz5"] Feb 27 08:36:00 crc kubenswrapper[4612]: I0227 08:36:00.146462 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536356-txvz5" Feb 27 08:36:00 crc kubenswrapper[4612]: I0227 08:36:00.158948 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536356-txvz5"] Feb 27 08:36:00 crc kubenswrapper[4612]: I0227 08:36:00.198379 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 27 08:36:00 crc kubenswrapper[4612]: I0227 08:36:00.199025 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-5zwfx" Feb 27 08:36:00 crc kubenswrapper[4612]: I0227 08:36:00.199188 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 27 08:36:00 crc kubenswrapper[4612]: I0227 08:36:00.345050 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ppn5q\" (UniqueName: \"kubernetes.io/projected/2b382b34-72d2-4144-a87c-50330a8396ad-kube-api-access-ppn5q\") pod \"auto-csr-approver-29536356-txvz5\" (UID: \"2b382b34-72d2-4144-a87c-50330a8396ad\") " pod="openshift-infra/auto-csr-approver-29536356-txvz5" Feb 27 08:36:00 crc kubenswrapper[4612]: I0227 08:36:00.446997 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ppn5q\" (UniqueName: \"kubernetes.io/projected/2b382b34-72d2-4144-a87c-50330a8396ad-kube-api-access-ppn5q\") pod \"auto-csr-approver-29536356-txvz5\" (UID: \"2b382b34-72d2-4144-a87c-50330a8396ad\") " pod="openshift-infra/auto-csr-approver-29536356-txvz5" Feb 27 08:36:00 crc kubenswrapper[4612]: I0227 08:36:00.467452 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ppn5q\" (UniqueName: \"kubernetes.io/projected/2b382b34-72d2-4144-a87c-50330a8396ad-kube-api-access-ppn5q\") pod \"auto-csr-approver-29536356-txvz5\" (UID: \"2b382b34-72d2-4144-a87c-50330a8396ad\") " pod="openshift-infra/auto-csr-approver-29536356-txvz5" Feb 27 08:36:00 crc kubenswrapper[4612]: I0227 08:36:00.512868 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536356-txvz5" Feb 27 08:36:00 crc kubenswrapper[4612]: I0227 08:36:00.963402 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536356-txvz5"] Feb 27 08:36:01 crc kubenswrapper[4612]: I0227 08:36:01.360125 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536356-txvz5" event={"ID":"2b382b34-72d2-4144-a87c-50330a8396ad","Type":"ContainerStarted","Data":"69ef493bb6cf04fcaece44620463a393e5d3f9bb1de883d41173142331848774"} Feb 27 08:36:04 crc kubenswrapper[4612]: I0227 08:36:04.315599 4612 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-gdv97" podUID="65d07354-a95f-4a47-a4df-bd5cb66814b2" containerName="registry-server" probeResult="failure" output=< Feb 27 08:36:04 crc kubenswrapper[4612]: timeout: failed to connect service ":50051" within 1s Feb 27 08:36:04 crc kubenswrapper[4612]: > Feb 27 08:36:05 crc kubenswrapper[4612]: I0227 08:36:05.395209 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536356-txvz5" event={"ID":"2b382b34-72d2-4144-a87c-50330a8396ad","Type":"ContainerStarted","Data":"524d1065917e1007784752569a9692b8810799b2065f018e1798ad7b08e5e650"} Feb 27 08:36:06 crc kubenswrapper[4612]: I0227 08:36:06.405405 4612 generic.go:334] "Generic (PLEG): container finished" podID="2b382b34-72d2-4144-a87c-50330a8396ad" containerID="524d1065917e1007784752569a9692b8810799b2065f018e1798ad7b08e5e650" exitCode=0 Feb 27 08:36:06 crc kubenswrapper[4612]: I0227 08:36:06.405457 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536356-txvz5" event={"ID":"2b382b34-72d2-4144-a87c-50330a8396ad","Type":"ContainerDied","Data":"524d1065917e1007784752569a9692b8810799b2065f018e1798ad7b08e5e650"} Feb 27 08:36:07 crc kubenswrapper[4612]: I0227 08:36:07.788294 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536356-txvz5" Feb 27 08:36:07 crc kubenswrapper[4612]: I0227 08:36:07.889816 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ppn5q\" (UniqueName: \"kubernetes.io/projected/2b382b34-72d2-4144-a87c-50330a8396ad-kube-api-access-ppn5q\") pod \"2b382b34-72d2-4144-a87c-50330a8396ad\" (UID: \"2b382b34-72d2-4144-a87c-50330a8396ad\") " Feb 27 08:36:07 crc kubenswrapper[4612]: I0227 08:36:07.894399 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2b382b34-72d2-4144-a87c-50330a8396ad-kube-api-access-ppn5q" (OuterVolumeSpecName: "kube-api-access-ppn5q") pod "2b382b34-72d2-4144-a87c-50330a8396ad" (UID: "2b382b34-72d2-4144-a87c-50330a8396ad"). InnerVolumeSpecName "kube-api-access-ppn5q". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 08:36:07 crc kubenswrapper[4612]: I0227 08:36:07.992366 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ppn5q\" (UniqueName: \"kubernetes.io/projected/2b382b34-72d2-4144-a87c-50330a8396ad-kube-api-access-ppn5q\") on node \"crc\" DevicePath \"\"" Feb 27 08:36:08 crc kubenswrapper[4612]: I0227 08:36:08.423237 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536356-txvz5" event={"ID":"2b382b34-72d2-4144-a87c-50330a8396ad","Type":"ContainerDied","Data":"69ef493bb6cf04fcaece44620463a393e5d3f9bb1de883d41173142331848774"} Feb 27 08:36:08 crc kubenswrapper[4612]: I0227 08:36:08.423284 4612 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="69ef493bb6cf04fcaece44620463a393e5d3f9bb1de883d41173142331848774" Feb 27 08:36:08 crc kubenswrapper[4612]: I0227 08:36:08.423331 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536356-txvz5" Feb 27 08:36:08 crc kubenswrapper[4612]: I0227 08:36:08.900228 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29536350-jwm78"] Feb 27 08:36:08 crc kubenswrapper[4612]: I0227 08:36:08.907082 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29536350-jwm78"] Feb 27 08:36:10 crc kubenswrapper[4612]: I0227 08:36:10.883619 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="98ea71d9-1b11-4cac-8281-c1220056a8ec" path="/var/lib/kubelet/pods/98ea71d9-1b11-4cac-8281-c1220056a8ec/volumes" Feb 27 08:36:14 crc kubenswrapper[4612]: I0227 08:36:14.315313 4612 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-gdv97" podUID="65d07354-a95f-4a47-a4df-bd5cb66814b2" containerName="registry-server" probeResult="failure" output=< Feb 27 08:36:14 crc kubenswrapper[4612]: timeout: failed to connect service ":50051" within 1s Feb 27 08:36:14 crc kubenswrapper[4612]: > Feb 27 08:36:18 crc kubenswrapper[4612]: I0227 08:36:18.203210 4612 scope.go:117] "RemoveContainer" containerID="8f940b943ef90992f45e40e478a0bc75fb1875a96059db043b0b71454915012f" Feb 27 08:36:24 crc kubenswrapper[4612]: I0227 08:36:24.327034 4612 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-gdv97" podUID="65d07354-a95f-4a47-a4df-bd5cb66814b2" containerName="registry-server" probeResult="failure" output=< Feb 27 08:36:24 crc kubenswrapper[4612]: timeout: failed to connect service ":50051" within 1s Feb 27 08:36:24 crc kubenswrapper[4612]: > Feb 27 08:36:33 crc kubenswrapper[4612]: I0227 08:36:33.318517 4612 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-gdv97" Feb 27 08:36:33 crc kubenswrapper[4612]: I0227 08:36:33.378839 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-gdv97" Feb 27 08:36:33 crc kubenswrapper[4612]: I0227 08:36:33.560427 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-gdv97"] Feb 27 08:36:33 crc kubenswrapper[4612]: I0227 08:36:33.644524 4612 generic.go:334] "Generic (PLEG): container finished" podID="f69e6377-1d38-4128-b6ad-af91cf8f7793" containerID="7c1ba70c93c91474ac80bc9be188fb1d4b6d5703023259f6d87838d2c096101c" exitCode=0 Feb 27 08:36:33 crc kubenswrapper[4612]: I0227 08:36:33.645055 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qthr6" event={"ID":"f69e6377-1d38-4128-b6ad-af91cf8f7793","Type":"ContainerDied","Data":"7c1ba70c93c91474ac80bc9be188fb1d4b6d5703023259f6d87838d2c096101c"} Feb 27 08:36:34 crc kubenswrapper[4612]: I0227 08:36:34.653633 4612 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-gdv97" podUID="65d07354-a95f-4a47-a4df-bd5cb66814b2" containerName="registry-server" containerID="cri-o://ebfd7e4c78fc97741adbfdd7a222bfc0bf0b3f788355b29312d40b6b4e3fdb48" gracePeriod=2 Feb 27 08:36:34 crc kubenswrapper[4612]: E0227 08:36:34.857448 4612 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod65d07354_a95f_4a47_a4df_bd5cb66814b2.slice/crio-ebfd7e4c78fc97741adbfdd7a222bfc0bf0b3f788355b29312d40b6b4e3fdb48.scope\": RecentStats: unable to find data in memory cache]" Feb 27 08:36:35 crc kubenswrapper[4612]: I0227 08:36:35.207555 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qthr6" Feb 27 08:36:35 crc kubenswrapper[4612]: I0227 08:36:35.219549 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-gdv97" Feb 27 08:36:35 crc kubenswrapper[4612]: I0227 08:36:35.341908 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f69e6377-1d38-4128-b6ad-af91cf8f7793-telemetry-combined-ca-bundle\") pod \"f69e6377-1d38-4128-b6ad-af91cf8f7793\" (UID: \"f69e6377-1d38-4128-b6ad-af91cf8f7793\") " Feb 27 08:36:35 crc kubenswrapper[4612]: I0227 08:36:35.341956 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f69e6377-1d38-4128-b6ad-af91cf8f7793-inventory\") pod \"f69e6377-1d38-4128-b6ad-af91cf8f7793\" (UID: \"f69e6377-1d38-4128-b6ad-af91cf8f7793\") " Feb 27 08:36:35 crc kubenswrapper[4612]: I0227 08:36:35.341988 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jl2bw\" (UniqueName: \"kubernetes.io/projected/f69e6377-1d38-4128-b6ad-af91cf8f7793-kube-api-access-jl2bw\") pod \"f69e6377-1d38-4128-b6ad-af91cf8f7793\" (UID: \"f69e6377-1d38-4128-b6ad-af91cf8f7793\") " Feb 27 08:36:35 crc kubenswrapper[4612]: I0227 08:36:35.342088 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nmdm2\" (UniqueName: \"kubernetes.io/projected/65d07354-a95f-4a47-a4df-bd5cb66814b2-kube-api-access-nmdm2\") pod \"65d07354-a95f-4a47-a4df-bd5cb66814b2\" (UID: \"65d07354-a95f-4a47-a4df-bd5cb66814b2\") " Feb 27 08:36:35 crc kubenswrapper[4612]: I0227 08:36:35.342122 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/f69e6377-1d38-4128-b6ad-af91cf8f7793-ceilometer-compute-config-data-1\") pod \"f69e6377-1d38-4128-b6ad-af91cf8f7793\" (UID: \"f69e6377-1d38-4128-b6ad-af91cf8f7793\") " Feb 27 08:36:35 crc kubenswrapper[4612]: I0227 08:36:35.342163 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/f69e6377-1d38-4128-b6ad-af91cf8f7793-ceilometer-compute-config-data-0\") pod \"f69e6377-1d38-4128-b6ad-af91cf8f7793\" (UID: \"f69e6377-1d38-4128-b6ad-af91cf8f7793\") " Feb 27 08:36:35 crc kubenswrapper[4612]: I0227 08:36:35.342268 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/65d07354-a95f-4a47-a4df-bd5cb66814b2-utilities\") pod \"65d07354-a95f-4a47-a4df-bd5cb66814b2\" (UID: \"65d07354-a95f-4a47-a4df-bd5cb66814b2\") " Feb 27 08:36:35 crc kubenswrapper[4612]: I0227 08:36:35.342301 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/f69e6377-1d38-4128-b6ad-af91cf8f7793-ssh-key-openstack-edpm-ipam\") pod \"f69e6377-1d38-4128-b6ad-af91cf8f7793\" (UID: \"f69e6377-1d38-4128-b6ad-af91cf8f7793\") " Feb 27 08:36:35 crc kubenswrapper[4612]: I0227 08:36:35.342360 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/f69e6377-1d38-4128-b6ad-af91cf8f7793-ceilometer-compute-config-data-2\") pod \"f69e6377-1d38-4128-b6ad-af91cf8f7793\" (UID: \"f69e6377-1d38-4128-b6ad-af91cf8f7793\") " Feb 27 08:36:35 crc kubenswrapper[4612]: I0227 08:36:35.342397 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/65d07354-a95f-4a47-a4df-bd5cb66814b2-catalog-content\") pod \"65d07354-a95f-4a47-a4df-bd5cb66814b2\" (UID: \"65d07354-a95f-4a47-a4df-bd5cb66814b2\") " Feb 27 08:36:35 crc kubenswrapper[4612]: I0227 08:36:35.344674 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/65d07354-a95f-4a47-a4df-bd5cb66814b2-utilities" (OuterVolumeSpecName: "utilities") pod "65d07354-a95f-4a47-a4df-bd5cb66814b2" (UID: "65d07354-a95f-4a47-a4df-bd5cb66814b2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 08:36:35 crc kubenswrapper[4612]: I0227 08:36:35.366326 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/65d07354-a95f-4a47-a4df-bd5cb66814b2-kube-api-access-nmdm2" (OuterVolumeSpecName: "kube-api-access-nmdm2") pod "65d07354-a95f-4a47-a4df-bd5cb66814b2" (UID: "65d07354-a95f-4a47-a4df-bd5cb66814b2"). InnerVolumeSpecName "kube-api-access-nmdm2". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 08:36:35 crc kubenswrapper[4612]: I0227 08:36:35.366412 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f69e6377-1d38-4128-b6ad-af91cf8f7793-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "f69e6377-1d38-4128-b6ad-af91cf8f7793" (UID: "f69e6377-1d38-4128-b6ad-af91cf8f7793"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:36:35 crc kubenswrapper[4612]: I0227 08:36:35.370666 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f69e6377-1d38-4128-b6ad-af91cf8f7793-ceilometer-compute-config-data-2" (OuterVolumeSpecName: "ceilometer-compute-config-data-2") pod "f69e6377-1d38-4128-b6ad-af91cf8f7793" (UID: "f69e6377-1d38-4128-b6ad-af91cf8f7793"). InnerVolumeSpecName "ceilometer-compute-config-data-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:36:35 crc kubenswrapper[4612]: I0227 08:36:35.375201 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f69e6377-1d38-4128-b6ad-af91cf8f7793-kube-api-access-jl2bw" (OuterVolumeSpecName: "kube-api-access-jl2bw") pod "f69e6377-1d38-4128-b6ad-af91cf8f7793" (UID: "f69e6377-1d38-4128-b6ad-af91cf8f7793"). InnerVolumeSpecName "kube-api-access-jl2bw". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 08:36:35 crc kubenswrapper[4612]: I0227 08:36:35.381869 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f69e6377-1d38-4128-b6ad-af91cf8f7793-ceilometer-compute-config-data-0" (OuterVolumeSpecName: "ceilometer-compute-config-data-0") pod "f69e6377-1d38-4128-b6ad-af91cf8f7793" (UID: "f69e6377-1d38-4128-b6ad-af91cf8f7793"). InnerVolumeSpecName "ceilometer-compute-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:36:35 crc kubenswrapper[4612]: I0227 08:36:35.384618 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f69e6377-1d38-4128-b6ad-af91cf8f7793-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "f69e6377-1d38-4128-b6ad-af91cf8f7793" (UID: "f69e6377-1d38-4128-b6ad-af91cf8f7793"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:36:35 crc kubenswrapper[4612]: I0227 08:36:35.414845 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f69e6377-1d38-4128-b6ad-af91cf8f7793-ceilometer-compute-config-data-1" (OuterVolumeSpecName: "ceilometer-compute-config-data-1") pod "f69e6377-1d38-4128-b6ad-af91cf8f7793" (UID: "f69e6377-1d38-4128-b6ad-af91cf8f7793"). InnerVolumeSpecName "ceilometer-compute-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:36:35 crc kubenswrapper[4612]: I0227 08:36:35.416554 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f69e6377-1d38-4128-b6ad-af91cf8f7793-inventory" (OuterVolumeSpecName: "inventory") pod "f69e6377-1d38-4128-b6ad-af91cf8f7793" (UID: "f69e6377-1d38-4128-b6ad-af91cf8f7793"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:36:35 crc kubenswrapper[4612]: I0227 08:36:35.444748 4612 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/f69e6377-1d38-4128-b6ad-af91cf8f7793-ceilometer-compute-config-data-2\") on node \"crc\" DevicePath \"\"" Feb 27 08:36:35 crc kubenswrapper[4612]: I0227 08:36:35.444784 4612 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f69e6377-1d38-4128-b6ad-af91cf8f7793-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 08:36:35 crc kubenswrapper[4612]: I0227 08:36:35.444800 4612 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f69e6377-1d38-4128-b6ad-af91cf8f7793-inventory\") on node \"crc\" DevicePath \"\"" Feb 27 08:36:35 crc kubenswrapper[4612]: I0227 08:36:35.444815 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jl2bw\" (UniqueName: \"kubernetes.io/projected/f69e6377-1d38-4128-b6ad-af91cf8f7793-kube-api-access-jl2bw\") on node \"crc\" DevicePath \"\"" Feb 27 08:36:35 crc kubenswrapper[4612]: I0227 08:36:35.444827 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nmdm2\" (UniqueName: \"kubernetes.io/projected/65d07354-a95f-4a47-a4df-bd5cb66814b2-kube-api-access-nmdm2\") on node \"crc\" DevicePath \"\"" Feb 27 08:36:35 crc kubenswrapper[4612]: I0227 08:36:35.444838 4612 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/f69e6377-1d38-4128-b6ad-af91cf8f7793-ceilometer-compute-config-data-1\") on node \"crc\" DevicePath \"\"" Feb 27 08:36:35 crc kubenswrapper[4612]: I0227 08:36:35.444849 4612 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/f69e6377-1d38-4128-b6ad-af91cf8f7793-ceilometer-compute-config-data-0\") on node \"crc\" DevicePath \"\"" Feb 27 08:36:35 crc kubenswrapper[4612]: I0227 08:36:35.444859 4612 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/65d07354-a95f-4a47-a4df-bd5cb66814b2-utilities\") on node \"crc\" DevicePath \"\"" Feb 27 08:36:35 crc kubenswrapper[4612]: I0227 08:36:35.444873 4612 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/f69e6377-1d38-4128-b6ad-af91cf8f7793-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Feb 27 08:36:35 crc kubenswrapper[4612]: I0227 08:36:35.464591 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/65d07354-a95f-4a47-a4df-bd5cb66814b2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "65d07354-a95f-4a47-a4df-bd5cb66814b2" (UID: "65d07354-a95f-4a47-a4df-bd5cb66814b2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 08:36:35 crc kubenswrapper[4612]: I0227 08:36:35.547364 4612 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/65d07354-a95f-4a47-a4df-bd5cb66814b2-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 27 08:36:35 crc kubenswrapper[4612]: I0227 08:36:35.663624 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qthr6" event={"ID":"f69e6377-1d38-4128-b6ad-af91cf8f7793","Type":"ContainerDied","Data":"1b5ff7e22385037c7f3214c5f5fa630513f8e13ca03388a1932e01441a1b3e9a"} Feb 27 08:36:35 crc kubenswrapper[4612]: I0227 08:36:35.663657 4612 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1b5ff7e22385037c7f3214c5f5fa630513f8e13ca03388a1932e01441a1b3e9a" Feb 27 08:36:35 crc kubenswrapper[4612]: I0227 08:36:35.663741 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qthr6" Feb 27 08:36:35 crc kubenswrapper[4612]: I0227 08:36:35.667365 4612 generic.go:334] "Generic (PLEG): container finished" podID="65d07354-a95f-4a47-a4df-bd5cb66814b2" containerID="ebfd7e4c78fc97741adbfdd7a222bfc0bf0b3f788355b29312d40b6b4e3fdb48" exitCode=0 Feb 27 08:36:35 crc kubenswrapper[4612]: I0227 08:36:35.667533 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gdv97" event={"ID":"65d07354-a95f-4a47-a4df-bd5cb66814b2","Type":"ContainerDied","Data":"ebfd7e4c78fc97741adbfdd7a222bfc0bf0b3f788355b29312d40b6b4e3fdb48"} Feb 27 08:36:35 crc kubenswrapper[4612]: I0227 08:36:35.667632 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gdv97" event={"ID":"65d07354-a95f-4a47-a4df-bd5cb66814b2","Type":"ContainerDied","Data":"c62584d072e40452df7d5acb90275cb93b8a8946815c73c559b506a8008f2c81"} Feb 27 08:36:35 crc kubenswrapper[4612]: I0227 08:36:35.667743 4612 scope.go:117] "RemoveContainer" containerID="ebfd7e4c78fc97741adbfdd7a222bfc0bf0b3f788355b29312d40b6b4e3fdb48" Feb 27 08:36:35 crc kubenswrapper[4612]: I0227 08:36:35.667549 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-gdv97" Feb 27 08:36:35 crc kubenswrapper[4612]: I0227 08:36:35.693839 4612 scope.go:117] "RemoveContainer" containerID="b9f5584d493c78e63fd469606cdd96b1b2490e974a491b170611420536233f78" Feb 27 08:36:35 crc kubenswrapper[4612]: I0227 08:36:35.728786 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-gdv97"] Feb 27 08:36:35 crc kubenswrapper[4612]: I0227 08:36:35.736740 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-gdv97"] Feb 27 08:36:35 crc kubenswrapper[4612]: I0227 08:36:35.740715 4612 scope.go:117] "RemoveContainer" containerID="f212e679fd155e68767180732b06409c017f70ee0527988dc40fcf1e1abf1c68" Feb 27 08:36:35 crc kubenswrapper[4612]: I0227 08:36:35.761298 4612 scope.go:117] "RemoveContainer" containerID="ebfd7e4c78fc97741adbfdd7a222bfc0bf0b3f788355b29312d40b6b4e3fdb48" Feb 27 08:36:35 crc kubenswrapper[4612]: E0227 08:36:35.761750 4612 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ebfd7e4c78fc97741adbfdd7a222bfc0bf0b3f788355b29312d40b6b4e3fdb48\": container with ID starting with ebfd7e4c78fc97741adbfdd7a222bfc0bf0b3f788355b29312d40b6b4e3fdb48 not found: ID does not exist" containerID="ebfd7e4c78fc97741adbfdd7a222bfc0bf0b3f788355b29312d40b6b4e3fdb48" Feb 27 08:36:35 crc kubenswrapper[4612]: I0227 08:36:35.761803 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ebfd7e4c78fc97741adbfdd7a222bfc0bf0b3f788355b29312d40b6b4e3fdb48"} err="failed to get container status \"ebfd7e4c78fc97741adbfdd7a222bfc0bf0b3f788355b29312d40b6b4e3fdb48\": rpc error: code = NotFound desc = could not find container \"ebfd7e4c78fc97741adbfdd7a222bfc0bf0b3f788355b29312d40b6b4e3fdb48\": container with ID starting with ebfd7e4c78fc97741adbfdd7a222bfc0bf0b3f788355b29312d40b6b4e3fdb48 not found: ID does not exist" Feb 27 08:36:35 crc kubenswrapper[4612]: I0227 08:36:35.761844 4612 scope.go:117] "RemoveContainer" containerID="b9f5584d493c78e63fd469606cdd96b1b2490e974a491b170611420536233f78" Feb 27 08:36:35 crc kubenswrapper[4612]: E0227 08:36:35.762226 4612 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b9f5584d493c78e63fd469606cdd96b1b2490e974a491b170611420536233f78\": container with ID starting with b9f5584d493c78e63fd469606cdd96b1b2490e974a491b170611420536233f78 not found: ID does not exist" containerID="b9f5584d493c78e63fd469606cdd96b1b2490e974a491b170611420536233f78" Feb 27 08:36:35 crc kubenswrapper[4612]: I0227 08:36:35.762257 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b9f5584d493c78e63fd469606cdd96b1b2490e974a491b170611420536233f78"} err="failed to get container status \"b9f5584d493c78e63fd469606cdd96b1b2490e974a491b170611420536233f78\": rpc error: code = NotFound desc = could not find container \"b9f5584d493c78e63fd469606cdd96b1b2490e974a491b170611420536233f78\": container with ID starting with b9f5584d493c78e63fd469606cdd96b1b2490e974a491b170611420536233f78 not found: ID does not exist" Feb 27 08:36:35 crc kubenswrapper[4612]: I0227 08:36:35.762279 4612 scope.go:117] "RemoveContainer" containerID="f212e679fd155e68767180732b06409c017f70ee0527988dc40fcf1e1abf1c68" Feb 27 08:36:35 crc kubenswrapper[4612]: E0227 08:36:35.762549 4612 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f212e679fd155e68767180732b06409c017f70ee0527988dc40fcf1e1abf1c68\": container with ID starting with f212e679fd155e68767180732b06409c017f70ee0527988dc40fcf1e1abf1c68 not found: ID does not exist" containerID="f212e679fd155e68767180732b06409c017f70ee0527988dc40fcf1e1abf1c68" Feb 27 08:36:35 crc kubenswrapper[4612]: I0227 08:36:35.762590 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f212e679fd155e68767180732b06409c017f70ee0527988dc40fcf1e1abf1c68"} err="failed to get container status \"f212e679fd155e68767180732b06409c017f70ee0527988dc40fcf1e1abf1c68\": rpc error: code = NotFound desc = could not find container \"f212e679fd155e68767180732b06409c017f70ee0527988dc40fcf1e1abf1c68\": container with ID starting with f212e679fd155e68767180732b06409c017f70ee0527988dc40fcf1e1abf1c68 not found: ID does not exist" Feb 27 08:36:36 crc kubenswrapper[4612]: I0227 08:36:36.864872 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="65d07354-a95f-4a47-a4df-bd5cb66814b2" path="/var/lib/kubelet/pods/65d07354-a95f-4a47-a4df-bd5cb66814b2/volumes" Feb 27 08:36:47 crc kubenswrapper[4612]: I0227 08:36:47.245918 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-bxw2f"] Feb 27 08:36:47 crc kubenswrapper[4612]: E0227 08:36:47.246825 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f69e6377-1d38-4128-b6ad-af91cf8f7793" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Feb 27 08:36:47 crc kubenswrapper[4612]: I0227 08:36:47.246841 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="f69e6377-1d38-4128-b6ad-af91cf8f7793" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Feb 27 08:36:47 crc kubenswrapper[4612]: E0227 08:36:47.246853 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="65d07354-a95f-4a47-a4df-bd5cb66814b2" containerName="registry-server" Feb 27 08:36:47 crc kubenswrapper[4612]: I0227 08:36:47.246859 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="65d07354-a95f-4a47-a4df-bd5cb66814b2" containerName="registry-server" Feb 27 08:36:47 crc kubenswrapper[4612]: E0227 08:36:47.246878 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="65d07354-a95f-4a47-a4df-bd5cb66814b2" containerName="extract-utilities" Feb 27 08:36:47 crc kubenswrapper[4612]: I0227 08:36:47.246885 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="65d07354-a95f-4a47-a4df-bd5cb66814b2" containerName="extract-utilities" Feb 27 08:36:47 crc kubenswrapper[4612]: E0227 08:36:47.246899 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="65d07354-a95f-4a47-a4df-bd5cb66814b2" containerName="extract-content" Feb 27 08:36:47 crc kubenswrapper[4612]: I0227 08:36:47.246905 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="65d07354-a95f-4a47-a4df-bd5cb66814b2" containerName="extract-content" Feb 27 08:36:47 crc kubenswrapper[4612]: E0227 08:36:47.246913 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2b382b34-72d2-4144-a87c-50330a8396ad" containerName="oc" Feb 27 08:36:47 crc kubenswrapper[4612]: I0227 08:36:47.246919 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="2b382b34-72d2-4144-a87c-50330a8396ad" containerName="oc" Feb 27 08:36:47 crc kubenswrapper[4612]: I0227 08:36:47.247102 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="2b382b34-72d2-4144-a87c-50330a8396ad" containerName="oc" Feb 27 08:36:47 crc kubenswrapper[4612]: I0227 08:36:47.247122 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="65d07354-a95f-4a47-a4df-bd5cb66814b2" containerName="registry-server" Feb 27 08:36:47 crc kubenswrapper[4612]: I0227 08:36:47.247133 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="f69e6377-1d38-4128-b6ad-af91cf8f7793" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Feb 27 08:36:47 crc kubenswrapper[4612]: I0227 08:36:47.248357 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bxw2f" Feb 27 08:36:47 crc kubenswrapper[4612]: I0227 08:36:47.270933 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-bxw2f"] Feb 27 08:36:47 crc kubenswrapper[4612]: I0227 08:36:47.399191 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nrpch\" (UniqueName: \"kubernetes.io/projected/81417ea9-86df-41fd-ae9c-d3ff112f0949-kube-api-access-nrpch\") pod \"certified-operators-bxw2f\" (UID: \"81417ea9-86df-41fd-ae9c-d3ff112f0949\") " pod="openshift-marketplace/certified-operators-bxw2f" Feb 27 08:36:47 crc kubenswrapper[4612]: I0227 08:36:47.399247 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/81417ea9-86df-41fd-ae9c-d3ff112f0949-catalog-content\") pod \"certified-operators-bxw2f\" (UID: \"81417ea9-86df-41fd-ae9c-d3ff112f0949\") " pod="openshift-marketplace/certified-operators-bxw2f" Feb 27 08:36:47 crc kubenswrapper[4612]: I0227 08:36:47.399484 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/81417ea9-86df-41fd-ae9c-d3ff112f0949-utilities\") pod \"certified-operators-bxw2f\" (UID: \"81417ea9-86df-41fd-ae9c-d3ff112f0949\") " pod="openshift-marketplace/certified-operators-bxw2f" Feb 27 08:36:47 crc kubenswrapper[4612]: I0227 08:36:47.501152 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nrpch\" (UniqueName: \"kubernetes.io/projected/81417ea9-86df-41fd-ae9c-d3ff112f0949-kube-api-access-nrpch\") pod \"certified-operators-bxw2f\" (UID: \"81417ea9-86df-41fd-ae9c-d3ff112f0949\") " pod="openshift-marketplace/certified-operators-bxw2f" Feb 27 08:36:47 crc kubenswrapper[4612]: I0227 08:36:47.501212 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/81417ea9-86df-41fd-ae9c-d3ff112f0949-catalog-content\") pod \"certified-operators-bxw2f\" (UID: \"81417ea9-86df-41fd-ae9c-d3ff112f0949\") " pod="openshift-marketplace/certified-operators-bxw2f" Feb 27 08:36:47 crc kubenswrapper[4612]: I0227 08:36:47.501281 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/81417ea9-86df-41fd-ae9c-d3ff112f0949-utilities\") pod \"certified-operators-bxw2f\" (UID: \"81417ea9-86df-41fd-ae9c-d3ff112f0949\") " pod="openshift-marketplace/certified-operators-bxw2f" Feb 27 08:36:47 crc kubenswrapper[4612]: I0227 08:36:47.501776 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/81417ea9-86df-41fd-ae9c-d3ff112f0949-utilities\") pod \"certified-operators-bxw2f\" (UID: \"81417ea9-86df-41fd-ae9c-d3ff112f0949\") " pod="openshift-marketplace/certified-operators-bxw2f" Feb 27 08:36:47 crc kubenswrapper[4612]: I0227 08:36:47.502302 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/81417ea9-86df-41fd-ae9c-d3ff112f0949-catalog-content\") pod \"certified-operators-bxw2f\" (UID: \"81417ea9-86df-41fd-ae9c-d3ff112f0949\") " pod="openshift-marketplace/certified-operators-bxw2f" Feb 27 08:36:47 crc kubenswrapper[4612]: I0227 08:36:47.533942 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nrpch\" (UniqueName: \"kubernetes.io/projected/81417ea9-86df-41fd-ae9c-d3ff112f0949-kube-api-access-nrpch\") pod \"certified-operators-bxw2f\" (UID: \"81417ea9-86df-41fd-ae9c-d3ff112f0949\") " pod="openshift-marketplace/certified-operators-bxw2f" Feb 27 08:36:47 crc kubenswrapper[4612]: I0227 08:36:47.568457 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bxw2f" Feb 27 08:36:48 crc kubenswrapper[4612]: I0227 08:36:48.126807 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-bxw2f"] Feb 27 08:36:48 crc kubenswrapper[4612]: I0227 08:36:48.815882 4612 generic.go:334] "Generic (PLEG): container finished" podID="81417ea9-86df-41fd-ae9c-d3ff112f0949" containerID="eb876bcea6868373bf5744f2d466b7b3ce745600c7608419df8ff8389a74ca6d" exitCode=0 Feb 27 08:36:48 crc kubenswrapper[4612]: I0227 08:36:48.816153 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bxw2f" event={"ID":"81417ea9-86df-41fd-ae9c-d3ff112f0949","Type":"ContainerDied","Data":"eb876bcea6868373bf5744f2d466b7b3ce745600c7608419df8ff8389a74ca6d"} Feb 27 08:36:48 crc kubenswrapper[4612]: I0227 08:36:48.816183 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bxw2f" event={"ID":"81417ea9-86df-41fd-ae9c-d3ff112f0949","Type":"ContainerStarted","Data":"b34ad5a281cb392dc96ef5c788bb7604ccb3e5fb1abf64a477e92ade76a338b6"} Feb 27 08:36:50 crc kubenswrapper[4612]: I0227 08:36:50.873192 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bxw2f" event={"ID":"81417ea9-86df-41fd-ae9c-d3ff112f0949","Type":"ContainerStarted","Data":"770bc9c904832f64ea872a3f8a56154198490a304e5a8cacdf940d5125ef3dee"} Feb 27 08:36:51 crc kubenswrapper[4612]: I0227 08:36:51.872908 4612 generic.go:334] "Generic (PLEG): container finished" podID="81417ea9-86df-41fd-ae9c-d3ff112f0949" containerID="770bc9c904832f64ea872a3f8a56154198490a304e5a8cacdf940d5125ef3dee" exitCode=0 Feb 27 08:36:51 crc kubenswrapper[4612]: I0227 08:36:51.873500 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bxw2f" event={"ID":"81417ea9-86df-41fd-ae9c-d3ff112f0949","Type":"ContainerDied","Data":"770bc9c904832f64ea872a3f8a56154198490a304e5a8cacdf940d5125ef3dee"} Feb 27 08:36:52 crc kubenswrapper[4612]: I0227 08:36:52.885202 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bxw2f" event={"ID":"81417ea9-86df-41fd-ae9c-d3ff112f0949","Type":"ContainerStarted","Data":"ad391a53c9ea3035b5961d9cb3ce6874154e1f94d88fbebff9a5dcf9c791df4c"} Feb 27 08:36:52 crc kubenswrapper[4612]: I0227 08:36:52.906462 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-bxw2f" podStartSLOduration=2.436529628 podStartE2EDuration="5.90643888s" podCreationTimestamp="2026-02-27 08:36:47 +0000 UTC" firstStartedPulling="2026-02-27 08:36:48.818024871 +0000 UTC m=+2866.671954869" lastFinishedPulling="2026-02-27 08:36:52.287934103 +0000 UTC m=+2870.141864121" observedRunningTime="2026-02-27 08:36:52.899825572 +0000 UTC m=+2870.753755570" watchObservedRunningTime="2026-02-27 08:36:52.90643888 +0000 UTC m=+2870.760368878" Feb 27 08:36:57 crc kubenswrapper[4612]: I0227 08:36:57.569854 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-bxw2f" Feb 27 08:36:57 crc kubenswrapper[4612]: I0227 08:36:57.570202 4612 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-bxw2f" Feb 27 08:36:57 crc kubenswrapper[4612]: I0227 08:36:57.643442 4612 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-bxw2f" Feb 27 08:36:57 crc kubenswrapper[4612]: I0227 08:36:57.970168 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-bxw2f" Feb 27 08:36:58 crc kubenswrapper[4612]: I0227 08:36:58.019219 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-bxw2f"] Feb 27 08:36:59 crc kubenswrapper[4612]: I0227 08:36:59.943047 4612 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-bxw2f" podUID="81417ea9-86df-41fd-ae9c-d3ff112f0949" containerName="registry-server" containerID="cri-o://ad391a53c9ea3035b5961d9cb3ce6874154e1f94d88fbebff9a5dcf9c791df4c" gracePeriod=2 Feb 27 08:37:00 crc kubenswrapper[4612]: I0227 08:37:00.409128 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bxw2f" Feb 27 08:37:00 crc kubenswrapper[4612]: I0227 08:37:00.571264 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/81417ea9-86df-41fd-ae9c-d3ff112f0949-utilities\") pod \"81417ea9-86df-41fd-ae9c-d3ff112f0949\" (UID: \"81417ea9-86df-41fd-ae9c-d3ff112f0949\") " Feb 27 08:37:00 crc kubenswrapper[4612]: I0227 08:37:00.571753 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nrpch\" (UniqueName: \"kubernetes.io/projected/81417ea9-86df-41fd-ae9c-d3ff112f0949-kube-api-access-nrpch\") pod \"81417ea9-86df-41fd-ae9c-d3ff112f0949\" (UID: \"81417ea9-86df-41fd-ae9c-d3ff112f0949\") " Feb 27 08:37:00 crc kubenswrapper[4612]: I0227 08:37:00.571960 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/81417ea9-86df-41fd-ae9c-d3ff112f0949-catalog-content\") pod \"81417ea9-86df-41fd-ae9c-d3ff112f0949\" (UID: \"81417ea9-86df-41fd-ae9c-d3ff112f0949\") " Feb 27 08:37:00 crc kubenswrapper[4612]: I0227 08:37:00.572080 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/81417ea9-86df-41fd-ae9c-d3ff112f0949-utilities" (OuterVolumeSpecName: "utilities") pod "81417ea9-86df-41fd-ae9c-d3ff112f0949" (UID: "81417ea9-86df-41fd-ae9c-d3ff112f0949"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 08:37:00 crc kubenswrapper[4612]: I0227 08:37:00.572623 4612 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/81417ea9-86df-41fd-ae9c-d3ff112f0949-utilities\") on node \"crc\" DevicePath \"\"" Feb 27 08:37:00 crc kubenswrapper[4612]: I0227 08:37:00.577459 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/81417ea9-86df-41fd-ae9c-d3ff112f0949-kube-api-access-nrpch" (OuterVolumeSpecName: "kube-api-access-nrpch") pod "81417ea9-86df-41fd-ae9c-d3ff112f0949" (UID: "81417ea9-86df-41fd-ae9c-d3ff112f0949"). InnerVolumeSpecName "kube-api-access-nrpch". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 08:37:00 crc kubenswrapper[4612]: I0227 08:37:00.634813 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/81417ea9-86df-41fd-ae9c-d3ff112f0949-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "81417ea9-86df-41fd-ae9c-d3ff112f0949" (UID: "81417ea9-86df-41fd-ae9c-d3ff112f0949"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 08:37:00 crc kubenswrapper[4612]: I0227 08:37:00.674642 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nrpch\" (UniqueName: \"kubernetes.io/projected/81417ea9-86df-41fd-ae9c-d3ff112f0949-kube-api-access-nrpch\") on node \"crc\" DevicePath \"\"" Feb 27 08:37:00 crc kubenswrapper[4612]: I0227 08:37:00.674680 4612 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/81417ea9-86df-41fd-ae9c-d3ff112f0949-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 27 08:37:00 crc kubenswrapper[4612]: I0227 08:37:00.954343 4612 generic.go:334] "Generic (PLEG): container finished" podID="81417ea9-86df-41fd-ae9c-d3ff112f0949" containerID="ad391a53c9ea3035b5961d9cb3ce6874154e1f94d88fbebff9a5dcf9c791df4c" exitCode=0 Feb 27 08:37:00 crc kubenswrapper[4612]: I0227 08:37:00.954636 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bxw2f" event={"ID":"81417ea9-86df-41fd-ae9c-d3ff112f0949","Type":"ContainerDied","Data":"ad391a53c9ea3035b5961d9cb3ce6874154e1f94d88fbebff9a5dcf9c791df4c"} Feb 27 08:37:00 crc kubenswrapper[4612]: I0227 08:37:00.954663 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bxw2f" event={"ID":"81417ea9-86df-41fd-ae9c-d3ff112f0949","Type":"ContainerDied","Data":"b34ad5a281cb392dc96ef5c788bb7604ccb3e5fb1abf64a477e92ade76a338b6"} Feb 27 08:37:00 crc kubenswrapper[4612]: I0227 08:37:00.954682 4612 scope.go:117] "RemoveContainer" containerID="ad391a53c9ea3035b5961d9cb3ce6874154e1f94d88fbebff9a5dcf9c791df4c" Feb 27 08:37:00 crc kubenswrapper[4612]: I0227 08:37:00.954874 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bxw2f" Feb 27 08:37:00 crc kubenswrapper[4612]: I0227 08:37:00.982114 4612 scope.go:117] "RemoveContainer" containerID="770bc9c904832f64ea872a3f8a56154198490a304e5a8cacdf940d5125ef3dee" Feb 27 08:37:00 crc kubenswrapper[4612]: I0227 08:37:00.989171 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-bxw2f"] Feb 27 08:37:00 crc kubenswrapper[4612]: I0227 08:37:00.999494 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-bxw2f"] Feb 27 08:37:01 crc kubenswrapper[4612]: I0227 08:37:01.018798 4612 scope.go:117] "RemoveContainer" containerID="eb876bcea6868373bf5744f2d466b7b3ce745600c7608419df8ff8389a74ca6d" Feb 27 08:37:01 crc kubenswrapper[4612]: I0227 08:37:01.057699 4612 scope.go:117] "RemoveContainer" containerID="ad391a53c9ea3035b5961d9cb3ce6874154e1f94d88fbebff9a5dcf9c791df4c" Feb 27 08:37:01 crc kubenswrapper[4612]: E0227 08:37:01.058824 4612 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ad391a53c9ea3035b5961d9cb3ce6874154e1f94d88fbebff9a5dcf9c791df4c\": container with ID starting with ad391a53c9ea3035b5961d9cb3ce6874154e1f94d88fbebff9a5dcf9c791df4c not found: ID does not exist" containerID="ad391a53c9ea3035b5961d9cb3ce6874154e1f94d88fbebff9a5dcf9c791df4c" Feb 27 08:37:01 crc kubenswrapper[4612]: I0227 08:37:01.058856 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ad391a53c9ea3035b5961d9cb3ce6874154e1f94d88fbebff9a5dcf9c791df4c"} err="failed to get container status \"ad391a53c9ea3035b5961d9cb3ce6874154e1f94d88fbebff9a5dcf9c791df4c\": rpc error: code = NotFound desc = could not find container \"ad391a53c9ea3035b5961d9cb3ce6874154e1f94d88fbebff9a5dcf9c791df4c\": container with ID starting with ad391a53c9ea3035b5961d9cb3ce6874154e1f94d88fbebff9a5dcf9c791df4c not found: ID does not exist" Feb 27 08:37:01 crc kubenswrapper[4612]: I0227 08:37:01.058876 4612 scope.go:117] "RemoveContainer" containerID="770bc9c904832f64ea872a3f8a56154198490a304e5a8cacdf940d5125ef3dee" Feb 27 08:37:01 crc kubenswrapper[4612]: E0227 08:37:01.059396 4612 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"770bc9c904832f64ea872a3f8a56154198490a304e5a8cacdf940d5125ef3dee\": container with ID starting with 770bc9c904832f64ea872a3f8a56154198490a304e5a8cacdf940d5125ef3dee not found: ID does not exist" containerID="770bc9c904832f64ea872a3f8a56154198490a304e5a8cacdf940d5125ef3dee" Feb 27 08:37:01 crc kubenswrapper[4612]: I0227 08:37:01.059424 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"770bc9c904832f64ea872a3f8a56154198490a304e5a8cacdf940d5125ef3dee"} err="failed to get container status \"770bc9c904832f64ea872a3f8a56154198490a304e5a8cacdf940d5125ef3dee\": rpc error: code = NotFound desc = could not find container \"770bc9c904832f64ea872a3f8a56154198490a304e5a8cacdf940d5125ef3dee\": container with ID starting with 770bc9c904832f64ea872a3f8a56154198490a304e5a8cacdf940d5125ef3dee not found: ID does not exist" Feb 27 08:37:01 crc kubenswrapper[4612]: I0227 08:37:01.059443 4612 scope.go:117] "RemoveContainer" containerID="eb876bcea6868373bf5744f2d466b7b3ce745600c7608419df8ff8389a74ca6d" Feb 27 08:37:01 crc kubenswrapper[4612]: E0227 08:37:01.059883 4612 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eb876bcea6868373bf5744f2d466b7b3ce745600c7608419df8ff8389a74ca6d\": container with ID starting with eb876bcea6868373bf5744f2d466b7b3ce745600c7608419df8ff8389a74ca6d not found: ID does not exist" containerID="eb876bcea6868373bf5744f2d466b7b3ce745600c7608419df8ff8389a74ca6d" Feb 27 08:37:01 crc kubenswrapper[4612]: I0227 08:37:01.059916 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eb876bcea6868373bf5744f2d466b7b3ce745600c7608419df8ff8389a74ca6d"} err="failed to get container status \"eb876bcea6868373bf5744f2d466b7b3ce745600c7608419df8ff8389a74ca6d\": rpc error: code = NotFound desc = could not find container \"eb876bcea6868373bf5744f2d466b7b3ce745600c7608419df8ff8389a74ca6d\": container with ID starting with eb876bcea6868373bf5744f2d466b7b3ce745600c7608419df8ff8389a74ca6d not found: ID does not exist" Feb 27 08:37:02 crc kubenswrapper[4612]: I0227 08:37:02.864615 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="81417ea9-86df-41fd-ae9c-d3ff112f0949" path="/var/lib/kubelet/pods/81417ea9-86df-41fd-ae9c-d3ff112f0949/volumes" Feb 27 08:37:16 crc kubenswrapper[4612]: I0227 08:37:16.027643 4612 patch_prober.go:28] interesting pod/machine-config-daemon-924vb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 27 08:37:16 crc kubenswrapper[4612]: I0227 08:37:16.029159 4612 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 27 08:37:35 crc kubenswrapper[4612]: I0227 08:37:35.664337 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tempest-tests-tempest"] Feb 27 08:37:35 crc kubenswrapper[4612]: E0227 08:37:35.665279 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="81417ea9-86df-41fd-ae9c-d3ff112f0949" containerName="extract-content" Feb 27 08:37:35 crc kubenswrapper[4612]: I0227 08:37:35.665297 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="81417ea9-86df-41fd-ae9c-d3ff112f0949" containerName="extract-content" Feb 27 08:37:35 crc kubenswrapper[4612]: E0227 08:37:35.665316 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="81417ea9-86df-41fd-ae9c-d3ff112f0949" containerName="extract-utilities" Feb 27 08:37:35 crc kubenswrapper[4612]: I0227 08:37:35.665324 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="81417ea9-86df-41fd-ae9c-d3ff112f0949" containerName="extract-utilities" Feb 27 08:37:35 crc kubenswrapper[4612]: E0227 08:37:35.665351 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="81417ea9-86df-41fd-ae9c-d3ff112f0949" containerName="registry-server" Feb 27 08:37:35 crc kubenswrapper[4612]: I0227 08:37:35.665358 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="81417ea9-86df-41fd-ae9c-d3ff112f0949" containerName="registry-server" Feb 27 08:37:35 crc kubenswrapper[4612]: I0227 08:37:35.665569 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="81417ea9-86df-41fd-ae9c-d3ff112f0949" containerName="registry-server" Feb 27 08:37:35 crc kubenswrapper[4612]: I0227 08:37:35.666274 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Feb 27 08:37:35 crc kubenswrapper[4612]: I0227 08:37:35.669464 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"test-operator-controller-priv-key" Feb 27 08:37:35 crc kubenswrapper[4612]: I0227 08:37:35.669646 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-custom-data-s0" Feb 27 08:37:35 crc kubenswrapper[4612]: I0227 08:37:35.671482 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-9rfwb" Feb 27 08:37:35 crc kubenswrapper[4612]: I0227 08:37:35.671774 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Feb 27 08:37:35 crc kubenswrapper[4612]: I0227 08:37:35.688401 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Feb 27 08:37:35 crc kubenswrapper[4612]: I0227 08:37:35.816005 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4700642e-da3a-4e53-84de-0bf4c1870ccd-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"4700642e-da3a-4e53-84de-0bf4c1870ccd\") " pod="openstack/tempest-tests-tempest" Feb 27 08:37:35 crc kubenswrapper[4612]: I0227 08:37:35.816076 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"tempest-tests-tempest\" (UID: \"4700642e-da3a-4e53-84de-0bf4c1870ccd\") " pod="openstack/tempest-tests-tempest" Feb 27 08:37:35 crc kubenswrapper[4612]: I0227 08:37:35.816377 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/4700642e-da3a-4e53-84de-0bf4c1870ccd-config-data\") pod \"tempest-tests-tempest\" (UID: \"4700642e-da3a-4e53-84de-0bf4c1870ccd\") " pod="openstack/tempest-tests-tempest" Feb 27 08:37:35 crc kubenswrapper[4612]: I0227 08:37:35.816507 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/4700642e-da3a-4e53-84de-0bf4c1870ccd-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"4700642e-da3a-4e53-84de-0bf4c1870ccd\") " pod="openstack/tempest-tests-tempest" Feb 27 08:37:35 crc kubenswrapper[4612]: I0227 08:37:35.816595 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/4700642e-da3a-4e53-84de-0bf4c1870ccd-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"4700642e-da3a-4e53-84de-0bf4c1870ccd\") " pod="openstack/tempest-tests-tempest" Feb 27 08:37:35 crc kubenswrapper[4612]: I0227 08:37:35.816629 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-69g4w\" (UniqueName: \"kubernetes.io/projected/4700642e-da3a-4e53-84de-0bf4c1870ccd-kube-api-access-69g4w\") pod \"tempest-tests-tempest\" (UID: \"4700642e-da3a-4e53-84de-0bf4c1870ccd\") " pod="openstack/tempest-tests-tempest" Feb 27 08:37:35 crc kubenswrapper[4612]: I0227 08:37:35.816648 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/4700642e-da3a-4e53-84de-0bf4c1870ccd-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"4700642e-da3a-4e53-84de-0bf4c1870ccd\") " pod="openstack/tempest-tests-tempest" Feb 27 08:37:35 crc kubenswrapper[4612]: I0227 08:37:35.816805 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/4700642e-da3a-4e53-84de-0bf4c1870ccd-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"4700642e-da3a-4e53-84de-0bf4c1870ccd\") " pod="openstack/tempest-tests-tempest" Feb 27 08:37:35 crc kubenswrapper[4612]: I0227 08:37:35.816841 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/4700642e-da3a-4e53-84de-0bf4c1870ccd-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"4700642e-da3a-4e53-84de-0bf4c1870ccd\") " pod="openstack/tempest-tests-tempest" Feb 27 08:37:35 crc kubenswrapper[4612]: I0227 08:37:35.918345 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/4700642e-da3a-4e53-84de-0bf4c1870ccd-config-data\") pod \"tempest-tests-tempest\" (UID: \"4700642e-da3a-4e53-84de-0bf4c1870ccd\") " pod="openstack/tempest-tests-tempest" Feb 27 08:37:35 crc kubenswrapper[4612]: I0227 08:37:35.918410 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/4700642e-da3a-4e53-84de-0bf4c1870ccd-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"4700642e-da3a-4e53-84de-0bf4c1870ccd\") " pod="openstack/tempest-tests-tempest" Feb 27 08:37:35 crc kubenswrapper[4612]: I0227 08:37:35.918440 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/4700642e-da3a-4e53-84de-0bf4c1870ccd-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"4700642e-da3a-4e53-84de-0bf4c1870ccd\") " pod="openstack/tempest-tests-tempest" Feb 27 08:37:35 crc kubenswrapper[4612]: I0227 08:37:35.918456 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-69g4w\" (UniqueName: \"kubernetes.io/projected/4700642e-da3a-4e53-84de-0bf4c1870ccd-kube-api-access-69g4w\") pod \"tempest-tests-tempest\" (UID: \"4700642e-da3a-4e53-84de-0bf4c1870ccd\") " pod="openstack/tempest-tests-tempest" Feb 27 08:37:35 crc kubenswrapper[4612]: I0227 08:37:35.918473 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/4700642e-da3a-4e53-84de-0bf4c1870ccd-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"4700642e-da3a-4e53-84de-0bf4c1870ccd\") " pod="openstack/tempest-tests-tempest" Feb 27 08:37:35 crc kubenswrapper[4612]: I0227 08:37:35.918524 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/4700642e-da3a-4e53-84de-0bf4c1870ccd-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"4700642e-da3a-4e53-84de-0bf4c1870ccd\") " pod="openstack/tempest-tests-tempest" Feb 27 08:37:35 crc kubenswrapper[4612]: I0227 08:37:35.918546 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/4700642e-da3a-4e53-84de-0bf4c1870ccd-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"4700642e-da3a-4e53-84de-0bf4c1870ccd\") " pod="openstack/tempest-tests-tempest" Feb 27 08:37:35 crc kubenswrapper[4612]: I0227 08:37:35.918574 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4700642e-da3a-4e53-84de-0bf4c1870ccd-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"4700642e-da3a-4e53-84de-0bf4c1870ccd\") " pod="openstack/tempest-tests-tempest" Feb 27 08:37:35 crc kubenswrapper[4612]: I0227 08:37:35.918599 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"tempest-tests-tempest\" (UID: \"4700642e-da3a-4e53-84de-0bf4c1870ccd\") " pod="openstack/tempest-tests-tempest" Feb 27 08:37:35 crc kubenswrapper[4612]: I0227 08:37:35.919013 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/4700642e-da3a-4e53-84de-0bf4c1870ccd-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"4700642e-da3a-4e53-84de-0bf4c1870ccd\") " pod="openstack/tempest-tests-tempest" Feb 27 08:37:35 crc kubenswrapper[4612]: I0227 08:37:35.919023 4612 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"tempest-tests-tempest\" (UID: \"4700642e-da3a-4e53-84de-0bf4c1870ccd\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/tempest-tests-tempest" Feb 27 08:37:35 crc kubenswrapper[4612]: I0227 08:37:35.919878 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/4700642e-da3a-4e53-84de-0bf4c1870ccd-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"4700642e-da3a-4e53-84de-0bf4c1870ccd\") " pod="openstack/tempest-tests-tempest" Feb 27 08:37:35 crc kubenswrapper[4612]: I0227 08:37:35.919917 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/4700642e-da3a-4e53-84de-0bf4c1870ccd-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"4700642e-da3a-4e53-84de-0bf4c1870ccd\") " pod="openstack/tempest-tests-tempest" Feb 27 08:37:35 crc kubenswrapper[4612]: I0227 08:37:35.920656 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/4700642e-da3a-4e53-84de-0bf4c1870ccd-config-data\") pod \"tempest-tests-tempest\" (UID: \"4700642e-da3a-4e53-84de-0bf4c1870ccd\") " pod="openstack/tempest-tests-tempest" Feb 27 08:37:35 crc kubenswrapper[4612]: I0227 08:37:35.925487 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4700642e-da3a-4e53-84de-0bf4c1870ccd-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"4700642e-da3a-4e53-84de-0bf4c1870ccd\") " pod="openstack/tempest-tests-tempest" Feb 27 08:37:35 crc kubenswrapper[4612]: I0227 08:37:35.926248 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/4700642e-da3a-4e53-84de-0bf4c1870ccd-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"4700642e-da3a-4e53-84de-0bf4c1870ccd\") " pod="openstack/tempest-tests-tempest" Feb 27 08:37:35 crc kubenswrapper[4612]: I0227 08:37:35.926528 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/4700642e-da3a-4e53-84de-0bf4c1870ccd-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"4700642e-da3a-4e53-84de-0bf4c1870ccd\") " pod="openstack/tempest-tests-tempest" Feb 27 08:37:35 crc kubenswrapper[4612]: I0227 08:37:35.942706 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-69g4w\" (UniqueName: \"kubernetes.io/projected/4700642e-da3a-4e53-84de-0bf4c1870ccd-kube-api-access-69g4w\") pod \"tempest-tests-tempest\" (UID: \"4700642e-da3a-4e53-84de-0bf4c1870ccd\") " pod="openstack/tempest-tests-tempest" Feb 27 08:37:35 crc kubenswrapper[4612]: I0227 08:37:35.944844 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"tempest-tests-tempest\" (UID: \"4700642e-da3a-4e53-84de-0bf4c1870ccd\") " pod="openstack/tempest-tests-tempest" Feb 27 08:37:36 crc kubenswrapper[4612]: I0227 08:37:36.002209 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Feb 27 08:37:36 crc kubenswrapper[4612]: I0227 08:37:36.494750 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Feb 27 08:37:37 crc kubenswrapper[4612]: I0227 08:37:37.299757 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"4700642e-da3a-4e53-84de-0bf4c1870ccd","Type":"ContainerStarted","Data":"cc7f3ff65e1f9decbe0960576c5071da6b09e9a686a9761a1b3bc857d80ff2f0"} Feb 27 08:37:46 crc kubenswrapper[4612]: I0227 08:37:46.027227 4612 patch_prober.go:28] interesting pod/machine-config-daemon-924vb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 27 08:37:46 crc kubenswrapper[4612]: I0227 08:37:46.027718 4612 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 27 08:38:00 crc kubenswrapper[4612]: I0227 08:38:00.154974 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29536358-mxzmn"] Feb 27 08:38:00 crc kubenswrapper[4612]: I0227 08:38:00.157282 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536358-mxzmn" Feb 27 08:38:00 crc kubenswrapper[4612]: I0227 08:38:00.159239 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 27 08:38:00 crc kubenswrapper[4612]: I0227 08:38:00.159426 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 27 08:38:00 crc kubenswrapper[4612]: I0227 08:38:00.160833 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-5zwfx" Feb 27 08:38:00 crc kubenswrapper[4612]: I0227 08:38:00.171332 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536358-mxzmn"] Feb 27 08:38:00 crc kubenswrapper[4612]: I0227 08:38:00.281044 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c8hf4\" (UniqueName: \"kubernetes.io/projected/1269c80c-6ede-4228-8d61-82efff8942e2-kube-api-access-c8hf4\") pod \"auto-csr-approver-29536358-mxzmn\" (UID: \"1269c80c-6ede-4228-8d61-82efff8942e2\") " pod="openshift-infra/auto-csr-approver-29536358-mxzmn" Feb 27 08:38:00 crc kubenswrapper[4612]: I0227 08:38:00.383058 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c8hf4\" (UniqueName: \"kubernetes.io/projected/1269c80c-6ede-4228-8d61-82efff8942e2-kube-api-access-c8hf4\") pod \"auto-csr-approver-29536358-mxzmn\" (UID: \"1269c80c-6ede-4228-8d61-82efff8942e2\") " pod="openshift-infra/auto-csr-approver-29536358-mxzmn" Feb 27 08:38:00 crc kubenswrapper[4612]: I0227 08:38:00.403180 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c8hf4\" (UniqueName: \"kubernetes.io/projected/1269c80c-6ede-4228-8d61-82efff8942e2-kube-api-access-c8hf4\") pod \"auto-csr-approver-29536358-mxzmn\" (UID: \"1269c80c-6ede-4228-8d61-82efff8942e2\") " pod="openshift-infra/auto-csr-approver-29536358-mxzmn" Feb 27 08:38:00 crc kubenswrapper[4612]: I0227 08:38:00.489615 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536358-mxzmn" Feb 27 08:38:16 crc kubenswrapper[4612]: I0227 08:38:16.027108 4612 patch_prober.go:28] interesting pod/machine-config-daemon-924vb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 27 08:38:16 crc kubenswrapper[4612]: I0227 08:38:16.027603 4612 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 27 08:38:16 crc kubenswrapper[4612]: I0227 08:38:16.027645 4612 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-924vb" Feb 27 08:38:16 crc kubenswrapper[4612]: I0227 08:38:16.028331 4612 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"384f63ec5e89debe3f33788a53a38ede9872a5e515bb3ecc12869cd886606dba"} pod="openshift-machine-config-operator/machine-config-daemon-924vb" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 27 08:38:16 crc kubenswrapper[4612]: I0227 08:38:16.028380 4612 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" containerName="machine-config-daemon" containerID="cri-o://384f63ec5e89debe3f33788a53a38ede9872a5e515bb3ecc12869cd886606dba" gracePeriod=600 Feb 27 08:38:16 crc kubenswrapper[4612]: I0227 08:38:16.720236 4612 generic.go:334] "Generic (PLEG): container finished" podID="24599373-7adc-4a1b-8bb4-797bf726f43d" containerID="384f63ec5e89debe3f33788a53a38ede9872a5e515bb3ecc12869cd886606dba" exitCode=0 Feb 27 08:38:16 crc kubenswrapper[4612]: I0227 08:38:16.720276 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-924vb" event={"ID":"24599373-7adc-4a1b-8bb4-797bf726f43d","Type":"ContainerDied","Data":"384f63ec5e89debe3f33788a53a38ede9872a5e515bb3ecc12869cd886606dba"} Feb 27 08:38:16 crc kubenswrapper[4612]: I0227 08:38:16.720342 4612 scope.go:117] "RemoveContainer" containerID="542804ac84a667bc92e9df5b1d69c8ad7e48579a796d2b39142a8890f0acd72f" Feb 27 08:38:19 crc kubenswrapper[4612]: E0227 08:38:19.104118 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-924vb_openshift-machine-config-operator(24599373-7adc-4a1b-8bb4-797bf726f43d)\"" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" Feb 27 08:38:19 crc kubenswrapper[4612]: E0227 08:38:19.179649 4612 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified" Feb 27 08:38:19 crc kubenswrapper[4612]: E0227 08:38:19.195297 4612 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:tempest-tests-tempest-tests-runner,Image:quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:false,MountPath:/etc/test_operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-workdir,ReadOnly:false,MountPath:/var/lib/tempest,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-temporary,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-logs,ReadOnly:false,MountPath:/var/lib/tempest/external_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/etc/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/var/lib/tempest/.config/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config-secret,ReadOnly:false,MountPath:/etc/openstack/secure.yaml,SubPath:secure.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ca-certs,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ssh-key,ReadOnly:false,MountPath:/var/lib/tempest/id_ecdsa,SubPath:ssh_key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-69g4w,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42480,RunAsNonRoot:*false,ReadOnlyRootFilesystem:*false,AllowPrivilegeEscalation:*true,RunAsGroup:*42480,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-custom-data-s0,},Optional:nil,},SecretRef:nil,},EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-env-vars-s0,},Optional:nil,},SecretRef:nil,},},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod tempest-tests-tempest_openstack(4700642e-da3a-4e53-84de-0bf4c1870ccd): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 27 08:38:19 crc kubenswrapper[4612]: E0227 08:38:19.197526 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/tempest-tests-tempest" podUID="4700642e-da3a-4e53-84de-0bf4c1870ccd" Feb 27 08:38:19 crc kubenswrapper[4612]: I0227 08:38:19.626671 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536358-mxzmn"] Feb 27 08:38:19 crc kubenswrapper[4612]: I0227 08:38:19.755341 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536358-mxzmn" event={"ID":"1269c80c-6ede-4228-8d61-82efff8942e2","Type":"ContainerStarted","Data":"35a1bfc885c1747587a13e5e4fd3ebfa0633686bfdb304fb5a6e222712ab49c8"} Feb 27 08:38:19 crc kubenswrapper[4612]: I0227 08:38:19.762510 4612 scope.go:117] "RemoveContainer" containerID="384f63ec5e89debe3f33788a53a38ede9872a5e515bb3ecc12869cd886606dba" Feb 27 08:38:19 crc kubenswrapper[4612]: E0227 08:38:19.763043 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-924vb_openshift-machine-config-operator(24599373-7adc-4a1b-8bb4-797bf726f43d)\"" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" Feb 27 08:38:19 crc kubenswrapper[4612]: E0227 08:38:19.768121 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified\\\"\"" pod="openstack/tempest-tests-tempest" podUID="4700642e-da3a-4e53-84de-0bf4c1870ccd" Feb 27 08:38:21 crc kubenswrapper[4612]: I0227 08:38:21.781790 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536358-mxzmn" event={"ID":"1269c80c-6ede-4228-8d61-82efff8942e2","Type":"ContainerStarted","Data":"9659d5d0ff3b5c78fcb3ea29d3fe15de918a7e2a964522b8abf3deb7f8cc283f"} Feb 27 08:38:21 crc kubenswrapper[4612]: I0227 08:38:21.797621 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29536358-mxzmn" podStartSLOduration=20.824363341 podStartE2EDuration="21.797603344s" podCreationTimestamp="2026-02-27 08:38:00 +0000 UTC" firstStartedPulling="2026-02-27 08:38:19.634663754 +0000 UTC m=+2957.488593752" lastFinishedPulling="2026-02-27 08:38:20.607903747 +0000 UTC m=+2958.461833755" observedRunningTime="2026-02-27 08:38:21.794710522 +0000 UTC m=+2959.648640510" watchObservedRunningTime="2026-02-27 08:38:21.797603344 +0000 UTC m=+2959.651533342" Feb 27 08:38:22 crc kubenswrapper[4612]: I0227 08:38:22.792444 4612 generic.go:334] "Generic (PLEG): container finished" podID="1269c80c-6ede-4228-8d61-82efff8942e2" containerID="9659d5d0ff3b5c78fcb3ea29d3fe15de918a7e2a964522b8abf3deb7f8cc283f" exitCode=0 Feb 27 08:38:22 crc kubenswrapper[4612]: I0227 08:38:22.792492 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536358-mxzmn" event={"ID":"1269c80c-6ede-4228-8d61-82efff8942e2","Type":"ContainerDied","Data":"9659d5d0ff3b5c78fcb3ea29d3fe15de918a7e2a964522b8abf3deb7f8cc283f"} Feb 27 08:38:24 crc kubenswrapper[4612]: I0227 08:38:24.226167 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536358-mxzmn" Feb 27 08:38:24 crc kubenswrapper[4612]: I0227 08:38:24.395162 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c8hf4\" (UniqueName: \"kubernetes.io/projected/1269c80c-6ede-4228-8d61-82efff8942e2-kube-api-access-c8hf4\") pod \"1269c80c-6ede-4228-8d61-82efff8942e2\" (UID: \"1269c80c-6ede-4228-8d61-82efff8942e2\") " Feb 27 08:38:24 crc kubenswrapper[4612]: I0227 08:38:24.400051 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1269c80c-6ede-4228-8d61-82efff8942e2-kube-api-access-c8hf4" (OuterVolumeSpecName: "kube-api-access-c8hf4") pod "1269c80c-6ede-4228-8d61-82efff8942e2" (UID: "1269c80c-6ede-4228-8d61-82efff8942e2"). InnerVolumeSpecName "kube-api-access-c8hf4". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 08:38:24 crc kubenswrapper[4612]: I0227 08:38:24.497233 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c8hf4\" (UniqueName: \"kubernetes.io/projected/1269c80c-6ede-4228-8d61-82efff8942e2-kube-api-access-c8hf4\") on node \"crc\" DevicePath \"\"" Feb 27 08:38:24 crc kubenswrapper[4612]: I0227 08:38:24.814600 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536358-mxzmn" event={"ID":"1269c80c-6ede-4228-8d61-82efff8942e2","Type":"ContainerDied","Data":"35a1bfc885c1747587a13e5e4fd3ebfa0633686bfdb304fb5a6e222712ab49c8"} Feb 27 08:38:24 crc kubenswrapper[4612]: I0227 08:38:24.814634 4612 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="35a1bfc885c1747587a13e5e4fd3ebfa0633686bfdb304fb5a6e222712ab49c8" Feb 27 08:38:24 crc kubenswrapper[4612]: I0227 08:38:24.814685 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536358-mxzmn" Feb 27 08:38:24 crc kubenswrapper[4612]: I0227 08:38:24.883290 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29536352-brq8t"] Feb 27 08:38:24 crc kubenswrapper[4612]: I0227 08:38:24.891219 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29536352-brq8t"] Feb 27 08:38:26 crc kubenswrapper[4612]: I0227 08:38:26.867272 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="318ef53b-57e4-4954-867e-d845fc77a484" path="/var/lib/kubelet/pods/318ef53b-57e4-4954-867e-d845fc77a484/volumes" Feb 27 08:38:33 crc kubenswrapper[4612]: I0227 08:38:33.332544 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Feb 27 08:38:34 crc kubenswrapper[4612]: I0227 08:38:34.853663 4612 scope.go:117] "RemoveContainer" containerID="384f63ec5e89debe3f33788a53a38ede9872a5e515bb3ecc12869cd886606dba" Feb 27 08:38:34 crc kubenswrapper[4612]: E0227 08:38:34.854454 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-924vb_openshift-machine-config-operator(24599373-7adc-4a1b-8bb4-797bf726f43d)\"" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" Feb 27 08:38:34 crc kubenswrapper[4612]: I0227 08:38:34.916423 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"4700642e-da3a-4e53-84de-0bf4c1870ccd","Type":"ContainerStarted","Data":"9420be57aa23b5a8049b44b757f2d7a73097b7dd7c3f1abfba3c6c1c0242cf7e"} Feb 27 08:38:34 crc kubenswrapper[4612]: I0227 08:38:34.943562 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tempest-tests-tempest" podStartSLOduration=4.119795498 podStartE2EDuration="1m0.943541423s" podCreationTimestamp="2026-02-27 08:37:34 +0000 UTC" firstStartedPulling="2026-02-27 08:37:36.505367133 +0000 UTC m=+2914.359297141" lastFinishedPulling="2026-02-27 08:38:33.329113068 +0000 UTC m=+2971.183043066" observedRunningTime="2026-02-27 08:38:34.93992555 +0000 UTC m=+2972.793855548" watchObservedRunningTime="2026-02-27 08:38:34.943541423 +0000 UTC m=+2972.797471421" Feb 27 08:38:48 crc kubenswrapper[4612]: I0227 08:38:48.853284 4612 scope.go:117] "RemoveContainer" containerID="384f63ec5e89debe3f33788a53a38ede9872a5e515bb3ecc12869cd886606dba" Feb 27 08:38:48 crc kubenswrapper[4612]: E0227 08:38:48.855382 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-924vb_openshift-machine-config-operator(24599373-7adc-4a1b-8bb4-797bf726f43d)\"" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" Feb 27 08:39:01 crc kubenswrapper[4612]: I0227 08:39:01.853966 4612 scope.go:117] "RemoveContainer" containerID="384f63ec5e89debe3f33788a53a38ede9872a5e515bb3ecc12869cd886606dba" Feb 27 08:39:01 crc kubenswrapper[4612]: E0227 08:39:01.854856 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-924vb_openshift-machine-config-operator(24599373-7adc-4a1b-8bb4-797bf726f43d)\"" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" Feb 27 08:39:14 crc kubenswrapper[4612]: I0227 08:39:14.852776 4612 scope.go:117] "RemoveContainer" containerID="384f63ec5e89debe3f33788a53a38ede9872a5e515bb3ecc12869cd886606dba" Feb 27 08:39:14 crc kubenswrapper[4612]: E0227 08:39:14.853564 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-924vb_openshift-machine-config-operator(24599373-7adc-4a1b-8bb4-797bf726f43d)\"" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" Feb 27 08:39:18 crc kubenswrapper[4612]: I0227 08:39:18.348579 4612 scope.go:117] "RemoveContainer" containerID="126c9df520238578841820ddbaf8bb9375205640276c7baf7814f85751306ecb" Feb 27 08:39:28 crc kubenswrapper[4612]: I0227 08:39:28.853459 4612 scope.go:117] "RemoveContainer" containerID="384f63ec5e89debe3f33788a53a38ede9872a5e515bb3ecc12869cd886606dba" Feb 27 08:39:28 crc kubenswrapper[4612]: E0227 08:39:28.854157 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-924vb_openshift-machine-config-operator(24599373-7adc-4a1b-8bb4-797bf726f43d)\"" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" Feb 27 08:39:39 crc kubenswrapper[4612]: I0227 08:39:39.853224 4612 scope.go:117] "RemoveContainer" containerID="384f63ec5e89debe3f33788a53a38ede9872a5e515bb3ecc12869cd886606dba" Feb 27 08:39:39 crc kubenswrapper[4612]: E0227 08:39:39.854242 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-924vb_openshift-machine-config-operator(24599373-7adc-4a1b-8bb4-797bf726f43d)\"" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" Feb 27 08:39:50 crc kubenswrapper[4612]: I0227 08:39:50.853977 4612 scope.go:117] "RemoveContainer" containerID="384f63ec5e89debe3f33788a53a38ede9872a5e515bb3ecc12869cd886606dba" Feb 27 08:39:50 crc kubenswrapper[4612]: E0227 08:39:50.854707 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-924vb_openshift-machine-config-operator(24599373-7adc-4a1b-8bb4-797bf726f43d)\"" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" Feb 27 08:40:00 crc kubenswrapper[4612]: I0227 08:40:00.146365 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29536360-pl294"] Feb 27 08:40:00 crc kubenswrapper[4612]: E0227 08:40:00.147264 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1269c80c-6ede-4228-8d61-82efff8942e2" containerName="oc" Feb 27 08:40:00 crc kubenswrapper[4612]: I0227 08:40:00.147282 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="1269c80c-6ede-4228-8d61-82efff8942e2" containerName="oc" Feb 27 08:40:00 crc kubenswrapper[4612]: I0227 08:40:00.147478 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="1269c80c-6ede-4228-8d61-82efff8942e2" containerName="oc" Feb 27 08:40:00 crc kubenswrapper[4612]: I0227 08:40:00.149685 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536360-pl294" Feb 27 08:40:00 crc kubenswrapper[4612]: I0227 08:40:00.152756 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 27 08:40:00 crc kubenswrapper[4612]: I0227 08:40:00.153057 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 27 08:40:00 crc kubenswrapper[4612]: I0227 08:40:00.159380 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-5zwfx" Feb 27 08:40:00 crc kubenswrapper[4612]: I0227 08:40:00.161278 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536360-pl294"] Feb 27 08:40:00 crc kubenswrapper[4612]: I0227 08:40:00.236067 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-glkrp\" (UniqueName: \"kubernetes.io/projected/faff5924-030b-4a8a-883e-830c195454ea-kube-api-access-glkrp\") pod \"auto-csr-approver-29536360-pl294\" (UID: \"faff5924-030b-4a8a-883e-830c195454ea\") " pod="openshift-infra/auto-csr-approver-29536360-pl294" Feb 27 08:40:00 crc kubenswrapper[4612]: I0227 08:40:00.338718 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-glkrp\" (UniqueName: \"kubernetes.io/projected/faff5924-030b-4a8a-883e-830c195454ea-kube-api-access-glkrp\") pod \"auto-csr-approver-29536360-pl294\" (UID: \"faff5924-030b-4a8a-883e-830c195454ea\") " pod="openshift-infra/auto-csr-approver-29536360-pl294" Feb 27 08:40:00 crc kubenswrapper[4612]: I0227 08:40:00.365549 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-glkrp\" (UniqueName: \"kubernetes.io/projected/faff5924-030b-4a8a-883e-830c195454ea-kube-api-access-glkrp\") pod \"auto-csr-approver-29536360-pl294\" (UID: \"faff5924-030b-4a8a-883e-830c195454ea\") " pod="openshift-infra/auto-csr-approver-29536360-pl294" Feb 27 08:40:00 crc kubenswrapper[4612]: I0227 08:40:00.471128 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536360-pl294" Feb 27 08:40:01 crc kubenswrapper[4612]: I0227 08:40:01.428352 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536360-pl294"] Feb 27 08:40:01 crc kubenswrapper[4612]: I0227 08:40:01.923739 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536360-pl294" event={"ID":"faff5924-030b-4a8a-883e-830c195454ea","Type":"ContainerStarted","Data":"8a51bf013f0d3c2627ccc8ebabf6843b8c509759f91eac7069eee1589d630254"} Feb 27 08:40:03 crc kubenswrapper[4612]: I0227 08:40:03.947055 4612 generic.go:334] "Generic (PLEG): container finished" podID="faff5924-030b-4a8a-883e-830c195454ea" containerID="76d3e2ca66bf6d6789f5256de1012d200acbda39a4c58d2c2515735b44cde6f6" exitCode=0 Feb 27 08:40:03 crc kubenswrapper[4612]: I0227 08:40:03.947122 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536360-pl294" event={"ID":"faff5924-030b-4a8a-883e-830c195454ea","Type":"ContainerDied","Data":"76d3e2ca66bf6d6789f5256de1012d200acbda39a4c58d2c2515735b44cde6f6"} Feb 27 08:40:05 crc kubenswrapper[4612]: I0227 08:40:05.402800 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536360-pl294" Feb 27 08:40:05 crc kubenswrapper[4612]: I0227 08:40:05.561193 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-glkrp\" (UniqueName: \"kubernetes.io/projected/faff5924-030b-4a8a-883e-830c195454ea-kube-api-access-glkrp\") pod \"faff5924-030b-4a8a-883e-830c195454ea\" (UID: \"faff5924-030b-4a8a-883e-830c195454ea\") " Feb 27 08:40:05 crc kubenswrapper[4612]: I0227 08:40:05.569397 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/faff5924-030b-4a8a-883e-830c195454ea-kube-api-access-glkrp" (OuterVolumeSpecName: "kube-api-access-glkrp") pod "faff5924-030b-4a8a-883e-830c195454ea" (UID: "faff5924-030b-4a8a-883e-830c195454ea"). InnerVolumeSpecName "kube-api-access-glkrp". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 08:40:05 crc kubenswrapper[4612]: I0227 08:40:05.663111 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-glkrp\" (UniqueName: \"kubernetes.io/projected/faff5924-030b-4a8a-883e-830c195454ea-kube-api-access-glkrp\") on node \"crc\" DevicePath \"\"" Feb 27 08:40:05 crc kubenswrapper[4612]: I0227 08:40:05.853677 4612 scope.go:117] "RemoveContainer" containerID="384f63ec5e89debe3f33788a53a38ede9872a5e515bb3ecc12869cd886606dba" Feb 27 08:40:05 crc kubenswrapper[4612]: E0227 08:40:05.854097 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-924vb_openshift-machine-config-operator(24599373-7adc-4a1b-8bb4-797bf726f43d)\"" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" Feb 27 08:40:05 crc kubenswrapper[4612]: I0227 08:40:05.964940 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536360-pl294" event={"ID":"faff5924-030b-4a8a-883e-830c195454ea","Type":"ContainerDied","Data":"8a51bf013f0d3c2627ccc8ebabf6843b8c509759f91eac7069eee1589d630254"} Feb 27 08:40:05 crc kubenswrapper[4612]: I0227 08:40:05.964975 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536360-pl294" Feb 27 08:40:05 crc kubenswrapper[4612]: I0227 08:40:05.964986 4612 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8a51bf013f0d3c2627ccc8ebabf6843b8c509759f91eac7069eee1589d630254" Feb 27 08:40:06 crc kubenswrapper[4612]: I0227 08:40:06.519956 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29536354-tb4sm"] Feb 27 08:40:06 crc kubenswrapper[4612]: I0227 08:40:06.528043 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29536354-tb4sm"] Feb 27 08:40:06 crc kubenswrapper[4612]: I0227 08:40:06.863629 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6973ea36-6dbd-4c24-bc4d-04b2180d9614" path="/var/lib/kubelet/pods/6973ea36-6dbd-4c24-bc4d-04b2180d9614/volumes" Feb 27 08:40:18 crc kubenswrapper[4612]: I0227 08:40:18.480595 4612 scope.go:117] "RemoveContainer" containerID="e537681481b958e073c9bb153dc93414b2a60df2afb30af474d132c00ab6819e" Feb 27 08:40:18 crc kubenswrapper[4612]: I0227 08:40:18.852971 4612 scope.go:117] "RemoveContainer" containerID="384f63ec5e89debe3f33788a53a38ede9872a5e515bb3ecc12869cd886606dba" Feb 27 08:40:18 crc kubenswrapper[4612]: E0227 08:40:18.853553 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-924vb_openshift-machine-config-operator(24599373-7adc-4a1b-8bb4-797bf726f43d)\"" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" Feb 27 08:40:30 crc kubenswrapper[4612]: I0227 08:40:30.862359 4612 scope.go:117] "RemoveContainer" containerID="384f63ec5e89debe3f33788a53a38ede9872a5e515bb3ecc12869cd886606dba" Feb 27 08:40:30 crc kubenswrapper[4612]: E0227 08:40:30.863729 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-924vb_openshift-machine-config-operator(24599373-7adc-4a1b-8bb4-797bf726f43d)\"" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" Feb 27 08:40:43 crc kubenswrapper[4612]: I0227 08:40:43.852745 4612 scope.go:117] "RemoveContainer" containerID="384f63ec5e89debe3f33788a53a38ede9872a5e515bb3ecc12869cd886606dba" Feb 27 08:40:43 crc kubenswrapper[4612]: E0227 08:40:43.853487 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-924vb_openshift-machine-config-operator(24599373-7adc-4a1b-8bb4-797bf726f43d)\"" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" Feb 27 08:40:55 crc kubenswrapper[4612]: I0227 08:40:55.853136 4612 scope.go:117] "RemoveContainer" containerID="384f63ec5e89debe3f33788a53a38ede9872a5e515bb3ecc12869cd886606dba" Feb 27 08:40:55 crc kubenswrapper[4612]: E0227 08:40:55.853922 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-924vb_openshift-machine-config-operator(24599373-7adc-4a1b-8bb4-797bf726f43d)\"" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" Feb 27 08:41:08 crc kubenswrapper[4612]: I0227 08:41:08.852994 4612 scope.go:117] "RemoveContainer" containerID="384f63ec5e89debe3f33788a53a38ede9872a5e515bb3ecc12869cd886606dba" Feb 27 08:41:08 crc kubenswrapper[4612]: E0227 08:41:08.854771 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-924vb_openshift-machine-config-operator(24599373-7adc-4a1b-8bb4-797bf726f43d)\"" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" Feb 27 08:41:19 crc kubenswrapper[4612]: I0227 08:41:19.855504 4612 scope.go:117] "RemoveContainer" containerID="384f63ec5e89debe3f33788a53a38ede9872a5e515bb3ecc12869cd886606dba" Feb 27 08:41:19 crc kubenswrapper[4612]: E0227 08:41:19.856293 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-924vb_openshift-machine-config-operator(24599373-7adc-4a1b-8bb4-797bf726f43d)\"" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" Feb 27 08:41:21 crc kubenswrapper[4612]: I0227 08:41:21.517466 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-68jxf"] Feb 27 08:41:21 crc kubenswrapper[4612]: E0227 08:41:21.518398 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="faff5924-030b-4a8a-883e-830c195454ea" containerName="oc" Feb 27 08:41:21 crc kubenswrapper[4612]: I0227 08:41:21.518416 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="faff5924-030b-4a8a-883e-830c195454ea" containerName="oc" Feb 27 08:41:21 crc kubenswrapper[4612]: I0227 08:41:21.518668 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="faff5924-030b-4a8a-883e-830c195454ea" containerName="oc" Feb 27 08:41:21 crc kubenswrapper[4612]: I0227 08:41:21.520347 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-68jxf" Feb 27 08:41:21 crc kubenswrapper[4612]: I0227 08:41:21.535082 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-68jxf"] Feb 27 08:41:21 crc kubenswrapper[4612]: I0227 08:41:21.629216 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2gb2s\" (UniqueName: \"kubernetes.io/projected/3f08af03-f7ff-42cf-ab46-f2a17fe2f6ed-kube-api-access-2gb2s\") pod \"community-operators-68jxf\" (UID: \"3f08af03-f7ff-42cf-ab46-f2a17fe2f6ed\") " pod="openshift-marketplace/community-operators-68jxf" Feb 27 08:41:21 crc kubenswrapper[4612]: I0227 08:41:21.629270 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3f08af03-f7ff-42cf-ab46-f2a17fe2f6ed-catalog-content\") pod \"community-operators-68jxf\" (UID: \"3f08af03-f7ff-42cf-ab46-f2a17fe2f6ed\") " pod="openshift-marketplace/community-operators-68jxf" Feb 27 08:41:21 crc kubenswrapper[4612]: I0227 08:41:21.629303 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3f08af03-f7ff-42cf-ab46-f2a17fe2f6ed-utilities\") pod \"community-operators-68jxf\" (UID: \"3f08af03-f7ff-42cf-ab46-f2a17fe2f6ed\") " pod="openshift-marketplace/community-operators-68jxf" Feb 27 08:41:21 crc kubenswrapper[4612]: I0227 08:41:21.730770 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2gb2s\" (UniqueName: \"kubernetes.io/projected/3f08af03-f7ff-42cf-ab46-f2a17fe2f6ed-kube-api-access-2gb2s\") pod \"community-operators-68jxf\" (UID: \"3f08af03-f7ff-42cf-ab46-f2a17fe2f6ed\") " pod="openshift-marketplace/community-operators-68jxf" Feb 27 08:41:21 crc kubenswrapper[4612]: I0227 08:41:21.730839 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3f08af03-f7ff-42cf-ab46-f2a17fe2f6ed-catalog-content\") pod \"community-operators-68jxf\" (UID: \"3f08af03-f7ff-42cf-ab46-f2a17fe2f6ed\") " pod="openshift-marketplace/community-operators-68jxf" Feb 27 08:41:21 crc kubenswrapper[4612]: I0227 08:41:21.730877 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3f08af03-f7ff-42cf-ab46-f2a17fe2f6ed-utilities\") pod \"community-operators-68jxf\" (UID: \"3f08af03-f7ff-42cf-ab46-f2a17fe2f6ed\") " pod="openshift-marketplace/community-operators-68jxf" Feb 27 08:41:21 crc kubenswrapper[4612]: I0227 08:41:21.731372 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3f08af03-f7ff-42cf-ab46-f2a17fe2f6ed-utilities\") pod \"community-operators-68jxf\" (UID: \"3f08af03-f7ff-42cf-ab46-f2a17fe2f6ed\") " pod="openshift-marketplace/community-operators-68jxf" Feb 27 08:41:21 crc kubenswrapper[4612]: I0227 08:41:21.731609 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3f08af03-f7ff-42cf-ab46-f2a17fe2f6ed-catalog-content\") pod \"community-operators-68jxf\" (UID: \"3f08af03-f7ff-42cf-ab46-f2a17fe2f6ed\") " pod="openshift-marketplace/community-operators-68jxf" Feb 27 08:41:21 crc kubenswrapper[4612]: I0227 08:41:21.762846 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2gb2s\" (UniqueName: \"kubernetes.io/projected/3f08af03-f7ff-42cf-ab46-f2a17fe2f6ed-kube-api-access-2gb2s\") pod \"community-operators-68jxf\" (UID: \"3f08af03-f7ff-42cf-ab46-f2a17fe2f6ed\") " pod="openshift-marketplace/community-operators-68jxf" Feb 27 08:41:21 crc kubenswrapper[4612]: I0227 08:41:21.869182 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-68jxf" Feb 27 08:41:22 crc kubenswrapper[4612]: I0227 08:41:22.571456 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-68jxf"] Feb 27 08:41:22 crc kubenswrapper[4612]: I0227 08:41:22.681209 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-68jxf" event={"ID":"3f08af03-f7ff-42cf-ab46-f2a17fe2f6ed","Type":"ContainerStarted","Data":"15780d47e135173bda675853e8ba9dd4a8d2688b7ae8a20d541fe4027baea9e5"} Feb 27 08:41:23 crc kubenswrapper[4612]: I0227 08:41:23.692939 4612 generic.go:334] "Generic (PLEG): container finished" podID="3f08af03-f7ff-42cf-ab46-f2a17fe2f6ed" containerID="450cc16e26f54b17c9f07beb8e8be1d4fd58de94c044deb43d7b417ac784abe2" exitCode=0 Feb 27 08:41:23 crc kubenswrapper[4612]: I0227 08:41:23.693026 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-68jxf" event={"ID":"3f08af03-f7ff-42cf-ab46-f2a17fe2f6ed","Type":"ContainerDied","Data":"450cc16e26f54b17c9f07beb8e8be1d4fd58de94c044deb43d7b417ac784abe2"} Feb 27 08:41:23 crc kubenswrapper[4612]: I0227 08:41:23.695973 4612 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 27 08:41:24 crc kubenswrapper[4612]: I0227 08:41:24.716391 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-68jxf" event={"ID":"3f08af03-f7ff-42cf-ab46-f2a17fe2f6ed","Type":"ContainerStarted","Data":"94132db57a57dbbc8c1e58515d4311e86e37cdd68908e9c9fbb5805e7a562e82"} Feb 27 08:41:27 crc kubenswrapper[4612]: I0227 08:41:27.744179 4612 generic.go:334] "Generic (PLEG): container finished" podID="3f08af03-f7ff-42cf-ab46-f2a17fe2f6ed" containerID="94132db57a57dbbc8c1e58515d4311e86e37cdd68908e9c9fbb5805e7a562e82" exitCode=0 Feb 27 08:41:27 crc kubenswrapper[4612]: I0227 08:41:27.744405 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-68jxf" event={"ID":"3f08af03-f7ff-42cf-ab46-f2a17fe2f6ed","Type":"ContainerDied","Data":"94132db57a57dbbc8c1e58515d4311e86e37cdd68908e9c9fbb5805e7a562e82"} Feb 27 08:41:28 crc kubenswrapper[4612]: I0227 08:41:28.755541 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-68jxf" event={"ID":"3f08af03-f7ff-42cf-ab46-f2a17fe2f6ed","Type":"ContainerStarted","Data":"927af5153a3f21986928cb4d47abfe0d003fe824c365b5642f7a8c394b92c83e"} Feb 27 08:41:28 crc kubenswrapper[4612]: I0227 08:41:28.779054 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-68jxf" podStartSLOduration=3.18623979 podStartE2EDuration="7.779036549s" podCreationTimestamp="2026-02-27 08:41:21 +0000 UTC" firstStartedPulling="2026-02-27 08:41:23.695484292 +0000 UTC m=+3141.549414290" lastFinishedPulling="2026-02-27 08:41:28.288281051 +0000 UTC m=+3146.142211049" observedRunningTime="2026-02-27 08:41:28.770588418 +0000 UTC m=+3146.624518416" watchObservedRunningTime="2026-02-27 08:41:28.779036549 +0000 UTC m=+3146.632966547" Feb 27 08:41:31 crc kubenswrapper[4612]: I0227 08:41:31.869669 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-68jxf" Feb 27 08:41:31 crc kubenswrapper[4612]: I0227 08:41:31.870131 4612 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-68jxf" Feb 27 08:41:32 crc kubenswrapper[4612]: I0227 08:41:32.861899 4612 scope.go:117] "RemoveContainer" containerID="384f63ec5e89debe3f33788a53a38ede9872a5e515bb3ecc12869cd886606dba" Feb 27 08:41:32 crc kubenswrapper[4612]: E0227 08:41:32.862515 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-924vb_openshift-machine-config-operator(24599373-7adc-4a1b-8bb4-797bf726f43d)\"" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" Feb 27 08:41:32 crc kubenswrapper[4612]: I0227 08:41:32.927655 4612 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-68jxf" podUID="3f08af03-f7ff-42cf-ab46-f2a17fe2f6ed" containerName="registry-server" probeResult="failure" output=< Feb 27 08:41:32 crc kubenswrapper[4612]: timeout: failed to connect service ":50051" within 1s Feb 27 08:41:32 crc kubenswrapper[4612]: > Feb 27 08:41:41 crc kubenswrapper[4612]: I0227 08:41:41.922678 4612 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-68jxf" Feb 27 08:41:41 crc kubenswrapper[4612]: I0227 08:41:41.967372 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-68jxf" Feb 27 08:41:42 crc kubenswrapper[4612]: I0227 08:41:42.170199 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-68jxf"] Feb 27 08:41:43 crc kubenswrapper[4612]: I0227 08:41:43.853734 4612 scope.go:117] "RemoveContainer" containerID="384f63ec5e89debe3f33788a53a38ede9872a5e515bb3ecc12869cd886606dba" Feb 27 08:41:43 crc kubenswrapper[4612]: E0227 08:41:43.854325 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-924vb_openshift-machine-config-operator(24599373-7adc-4a1b-8bb4-797bf726f43d)\"" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" Feb 27 08:41:43 crc kubenswrapper[4612]: I0227 08:41:43.885187 4612 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-68jxf" podUID="3f08af03-f7ff-42cf-ab46-f2a17fe2f6ed" containerName="registry-server" containerID="cri-o://927af5153a3f21986928cb4d47abfe0d003fe824c365b5642f7a8c394b92c83e" gracePeriod=2 Feb 27 08:41:44 crc kubenswrapper[4612]: I0227 08:41:44.624335 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-68jxf" Feb 27 08:41:44 crc kubenswrapper[4612]: I0227 08:41:44.727100 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3f08af03-f7ff-42cf-ab46-f2a17fe2f6ed-utilities\") pod \"3f08af03-f7ff-42cf-ab46-f2a17fe2f6ed\" (UID: \"3f08af03-f7ff-42cf-ab46-f2a17fe2f6ed\") " Feb 27 08:41:44 crc kubenswrapper[4612]: I0227 08:41:44.727446 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2gb2s\" (UniqueName: \"kubernetes.io/projected/3f08af03-f7ff-42cf-ab46-f2a17fe2f6ed-kube-api-access-2gb2s\") pod \"3f08af03-f7ff-42cf-ab46-f2a17fe2f6ed\" (UID: \"3f08af03-f7ff-42cf-ab46-f2a17fe2f6ed\") " Feb 27 08:41:44 crc kubenswrapper[4612]: I0227 08:41:44.727505 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3f08af03-f7ff-42cf-ab46-f2a17fe2f6ed-catalog-content\") pod \"3f08af03-f7ff-42cf-ab46-f2a17fe2f6ed\" (UID: \"3f08af03-f7ff-42cf-ab46-f2a17fe2f6ed\") " Feb 27 08:41:44 crc kubenswrapper[4612]: I0227 08:41:44.728280 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3f08af03-f7ff-42cf-ab46-f2a17fe2f6ed-utilities" (OuterVolumeSpecName: "utilities") pod "3f08af03-f7ff-42cf-ab46-f2a17fe2f6ed" (UID: "3f08af03-f7ff-42cf-ab46-f2a17fe2f6ed"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 08:41:44 crc kubenswrapper[4612]: I0227 08:41:44.768024 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3f08af03-f7ff-42cf-ab46-f2a17fe2f6ed-kube-api-access-2gb2s" (OuterVolumeSpecName: "kube-api-access-2gb2s") pod "3f08af03-f7ff-42cf-ab46-f2a17fe2f6ed" (UID: "3f08af03-f7ff-42cf-ab46-f2a17fe2f6ed"). InnerVolumeSpecName "kube-api-access-2gb2s". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 08:41:44 crc kubenswrapper[4612]: I0227 08:41:44.820414 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3f08af03-f7ff-42cf-ab46-f2a17fe2f6ed-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3f08af03-f7ff-42cf-ab46-f2a17fe2f6ed" (UID: "3f08af03-f7ff-42cf-ab46-f2a17fe2f6ed"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 08:41:44 crc kubenswrapper[4612]: I0227 08:41:44.829852 4612 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3f08af03-f7ff-42cf-ab46-f2a17fe2f6ed-utilities\") on node \"crc\" DevicePath \"\"" Feb 27 08:41:44 crc kubenswrapper[4612]: I0227 08:41:44.829894 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2gb2s\" (UniqueName: \"kubernetes.io/projected/3f08af03-f7ff-42cf-ab46-f2a17fe2f6ed-kube-api-access-2gb2s\") on node \"crc\" DevicePath \"\"" Feb 27 08:41:44 crc kubenswrapper[4612]: I0227 08:41:44.829910 4612 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3f08af03-f7ff-42cf-ab46-f2a17fe2f6ed-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 27 08:41:44 crc kubenswrapper[4612]: I0227 08:41:44.903454 4612 generic.go:334] "Generic (PLEG): container finished" podID="3f08af03-f7ff-42cf-ab46-f2a17fe2f6ed" containerID="927af5153a3f21986928cb4d47abfe0d003fe824c365b5642f7a8c394b92c83e" exitCode=0 Feb 27 08:41:44 crc kubenswrapper[4612]: I0227 08:41:44.903499 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-68jxf" event={"ID":"3f08af03-f7ff-42cf-ab46-f2a17fe2f6ed","Type":"ContainerDied","Data":"927af5153a3f21986928cb4d47abfe0d003fe824c365b5642f7a8c394b92c83e"} Feb 27 08:41:44 crc kubenswrapper[4612]: I0227 08:41:44.903525 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-68jxf" event={"ID":"3f08af03-f7ff-42cf-ab46-f2a17fe2f6ed","Type":"ContainerDied","Data":"15780d47e135173bda675853e8ba9dd4a8d2688b7ae8a20d541fe4027baea9e5"} Feb 27 08:41:44 crc kubenswrapper[4612]: I0227 08:41:44.903554 4612 scope.go:117] "RemoveContainer" containerID="927af5153a3f21986928cb4d47abfe0d003fe824c365b5642f7a8c394b92c83e" Feb 27 08:41:44 crc kubenswrapper[4612]: I0227 08:41:44.903930 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-68jxf" Feb 27 08:41:44 crc kubenswrapper[4612]: I0227 08:41:44.937144 4612 scope.go:117] "RemoveContainer" containerID="94132db57a57dbbc8c1e58515d4311e86e37cdd68908e9c9fbb5805e7a562e82" Feb 27 08:41:44 crc kubenswrapper[4612]: I0227 08:41:44.953943 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-68jxf"] Feb 27 08:41:44 crc kubenswrapper[4612]: I0227 08:41:44.963648 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-68jxf"] Feb 27 08:41:44 crc kubenswrapper[4612]: I0227 08:41:44.992435 4612 scope.go:117] "RemoveContainer" containerID="450cc16e26f54b17c9f07beb8e8be1d4fd58de94c044deb43d7b417ac784abe2" Feb 27 08:41:45 crc kubenswrapper[4612]: I0227 08:41:45.045618 4612 scope.go:117] "RemoveContainer" containerID="927af5153a3f21986928cb4d47abfe0d003fe824c365b5642f7a8c394b92c83e" Feb 27 08:41:45 crc kubenswrapper[4612]: E0227 08:41:45.047599 4612 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"927af5153a3f21986928cb4d47abfe0d003fe824c365b5642f7a8c394b92c83e\": container with ID starting with 927af5153a3f21986928cb4d47abfe0d003fe824c365b5642f7a8c394b92c83e not found: ID does not exist" containerID="927af5153a3f21986928cb4d47abfe0d003fe824c365b5642f7a8c394b92c83e" Feb 27 08:41:45 crc kubenswrapper[4612]: I0227 08:41:45.047648 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"927af5153a3f21986928cb4d47abfe0d003fe824c365b5642f7a8c394b92c83e"} err="failed to get container status \"927af5153a3f21986928cb4d47abfe0d003fe824c365b5642f7a8c394b92c83e\": rpc error: code = NotFound desc = could not find container \"927af5153a3f21986928cb4d47abfe0d003fe824c365b5642f7a8c394b92c83e\": container with ID starting with 927af5153a3f21986928cb4d47abfe0d003fe824c365b5642f7a8c394b92c83e not found: ID does not exist" Feb 27 08:41:45 crc kubenswrapper[4612]: I0227 08:41:45.047678 4612 scope.go:117] "RemoveContainer" containerID="94132db57a57dbbc8c1e58515d4311e86e37cdd68908e9c9fbb5805e7a562e82" Feb 27 08:41:45 crc kubenswrapper[4612]: E0227 08:41:45.048100 4612 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"94132db57a57dbbc8c1e58515d4311e86e37cdd68908e9c9fbb5805e7a562e82\": container with ID starting with 94132db57a57dbbc8c1e58515d4311e86e37cdd68908e9c9fbb5805e7a562e82 not found: ID does not exist" containerID="94132db57a57dbbc8c1e58515d4311e86e37cdd68908e9c9fbb5805e7a562e82" Feb 27 08:41:45 crc kubenswrapper[4612]: I0227 08:41:45.048124 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"94132db57a57dbbc8c1e58515d4311e86e37cdd68908e9c9fbb5805e7a562e82"} err="failed to get container status \"94132db57a57dbbc8c1e58515d4311e86e37cdd68908e9c9fbb5805e7a562e82\": rpc error: code = NotFound desc = could not find container \"94132db57a57dbbc8c1e58515d4311e86e37cdd68908e9c9fbb5805e7a562e82\": container with ID starting with 94132db57a57dbbc8c1e58515d4311e86e37cdd68908e9c9fbb5805e7a562e82 not found: ID does not exist" Feb 27 08:41:45 crc kubenswrapper[4612]: I0227 08:41:45.048142 4612 scope.go:117] "RemoveContainer" containerID="450cc16e26f54b17c9f07beb8e8be1d4fd58de94c044deb43d7b417ac784abe2" Feb 27 08:41:45 crc kubenswrapper[4612]: E0227 08:41:45.048366 4612 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"450cc16e26f54b17c9f07beb8e8be1d4fd58de94c044deb43d7b417ac784abe2\": container with ID starting with 450cc16e26f54b17c9f07beb8e8be1d4fd58de94c044deb43d7b417ac784abe2 not found: ID does not exist" containerID="450cc16e26f54b17c9f07beb8e8be1d4fd58de94c044deb43d7b417ac784abe2" Feb 27 08:41:45 crc kubenswrapper[4612]: I0227 08:41:45.048390 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"450cc16e26f54b17c9f07beb8e8be1d4fd58de94c044deb43d7b417ac784abe2"} err="failed to get container status \"450cc16e26f54b17c9f07beb8e8be1d4fd58de94c044deb43d7b417ac784abe2\": rpc error: code = NotFound desc = could not find container \"450cc16e26f54b17c9f07beb8e8be1d4fd58de94c044deb43d7b417ac784abe2\": container with ID starting with 450cc16e26f54b17c9f07beb8e8be1d4fd58de94c044deb43d7b417ac784abe2 not found: ID does not exist" Feb 27 08:41:46 crc kubenswrapper[4612]: I0227 08:41:46.864173 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3f08af03-f7ff-42cf-ab46-f2a17fe2f6ed" path="/var/lib/kubelet/pods/3f08af03-f7ff-42cf-ab46-f2a17fe2f6ed/volumes" Feb 27 08:41:58 crc kubenswrapper[4612]: I0227 08:41:58.857151 4612 scope.go:117] "RemoveContainer" containerID="384f63ec5e89debe3f33788a53a38ede9872a5e515bb3ecc12869cd886606dba" Feb 27 08:41:58 crc kubenswrapper[4612]: E0227 08:41:58.857918 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-924vb_openshift-machine-config-operator(24599373-7adc-4a1b-8bb4-797bf726f43d)\"" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" Feb 27 08:42:00 crc kubenswrapper[4612]: I0227 08:42:00.184377 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29536362-cphx2"] Feb 27 08:42:00 crc kubenswrapper[4612]: E0227 08:42:00.185076 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3f08af03-f7ff-42cf-ab46-f2a17fe2f6ed" containerName="extract-content" Feb 27 08:42:00 crc kubenswrapper[4612]: I0227 08:42:00.185090 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="3f08af03-f7ff-42cf-ab46-f2a17fe2f6ed" containerName="extract-content" Feb 27 08:42:00 crc kubenswrapper[4612]: E0227 08:42:00.185100 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3f08af03-f7ff-42cf-ab46-f2a17fe2f6ed" containerName="registry-server" Feb 27 08:42:00 crc kubenswrapper[4612]: I0227 08:42:00.185106 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="3f08af03-f7ff-42cf-ab46-f2a17fe2f6ed" containerName="registry-server" Feb 27 08:42:00 crc kubenswrapper[4612]: E0227 08:42:00.185124 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3f08af03-f7ff-42cf-ab46-f2a17fe2f6ed" containerName="extract-utilities" Feb 27 08:42:00 crc kubenswrapper[4612]: I0227 08:42:00.185130 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="3f08af03-f7ff-42cf-ab46-f2a17fe2f6ed" containerName="extract-utilities" Feb 27 08:42:00 crc kubenswrapper[4612]: I0227 08:42:00.185297 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="3f08af03-f7ff-42cf-ab46-f2a17fe2f6ed" containerName="registry-server" Feb 27 08:42:00 crc kubenswrapper[4612]: I0227 08:42:00.185954 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536362-cphx2" Feb 27 08:42:00 crc kubenswrapper[4612]: I0227 08:42:00.188583 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-5zwfx" Feb 27 08:42:00 crc kubenswrapper[4612]: I0227 08:42:00.188664 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 27 08:42:00 crc kubenswrapper[4612]: I0227 08:42:00.189170 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 27 08:42:00 crc kubenswrapper[4612]: I0227 08:42:00.226186 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536362-cphx2"] Feb 27 08:42:00 crc kubenswrapper[4612]: I0227 08:42:00.239589 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z6xmc\" (UniqueName: \"kubernetes.io/projected/7f89c45d-7de2-4b7b-93da-ea210d605384-kube-api-access-z6xmc\") pod \"auto-csr-approver-29536362-cphx2\" (UID: \"7f89c45d-7de2-4b7b-93da-ea210d605384\") " pod="openshift-infra/auto-csr-approver-29536362-cphx2" Feb 27 08:42:00 crc kubenswrapper[4612]: I0227 08:42:00.341348 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z6xmc\" (UniqueName: \"kubernetes.io/projected/7f89c45d-7de2-4b7b-93da-ea210d605384-kube-api-access-z6xmc\") pod \"auto-csr-approver-29536362-cphx2\" (UID: \"7f89c45d-7de2-4b7b-93da-ea210d605384\") " pod="openshift-infra/auto-csr-approver-29536362-cphx2" Feb 27 08:42:00 crc kubenswrapper[4612]: I0227 08:42:00.362363 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z6xmc\" (UniqueName: \"kubernetes.io/projected/7f89c45d-7de2-4b7b-93da-ea210d605384-kube-api-access-z6xmc\") pod \"auto-csr-approver-29536362-cphx2\" (UID: \"7f89c45d-7de2-4b7b-93da-ea210d605384\") " pod="openshift-infra/auto-csr-approver-29536362-cphx2" Feb 27 08:42:00 crc kubenswrapper[4612]: I0227 08:42:00.507213 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536362-cphx2" Feb 27 08:42:01 crc kubenswrapper[4612]: I0227 08:42:01.077426 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536362-cphx2"] Feb 27 08:42:02 crc kubenswrapper[4612]: I0227 08:42:02.041965 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536362-cphx2" event={"ID":"7f89c45d-7de2-4b7b-93da-ea210d605384","Type":"ContainerStarted","Data":"2fb76c33373bbb3102b45317e0b1b3f91eead5cb240207413beb750af757585a"} Feb 27 08:42:03 crc kubenswrapper[4612]: I0227 08:42:03.052621 4612 generic.go:334] "Generic (PLEG): container finished" podID="7f89c45d-7de2-4b7b-93da-ea210d605384" containerID="16c57272f23a9361528d5d4d7e2d957dcc474a740cf8ca2f0cb4ae3a6abb7126" exitCode=0 Feb 27 08:42:03 crc kubenswrapper[4612]: I0227 08:42:03.052946 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536362-cphx2" event={"ID":"7f89c45d-7de2-4b7b-93da-ea210d605384","Type":"ContainerDied","Data":"16c57272f23a9361528d5d4d7e2d957dcc474a740cf8ca2f0cb4ae3a6abb7126"} Feb 27 08:42:04 crc kubenswrapper[4612]: I0227 08:42:04.641406 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536362-cphx2" Feb 27 08:42:04 crc kubenswrapper[4612]: I0227 08:42:04.766266 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z6xmc\" (UniqueName: \"kubernetes.io/projected/7f89c45d-7de2-4b7b-93da-ea210d605384-kube-api-access-z6xmc\") pod \"7f89c45d-7de2-4b7b-93da-ea210d605384\" (UID: \"7f89c45d-7de2-4b7b-93da-ea210d605384\") " Feb 27 08:42:04 crc kubenswrapper[4612]: I0227 08:42:04.775828 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7f89c45d-7de2-4b7b-93da-ea210d605384-kube-api-access-z6xmc" (OuterVolumeSpecName: "kube-api-access-z6xmc") pod "7f89c45d-7de2-4b7b-93da-ea210d605384" (UID: "7f89c45d-7de2-4b7b-93da-ea210d605384"). InnerVolumeSpecName "kube-api-access-z6xmc". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 08:42:04 crc kubenswrapper[4612]: I0227 08:42:04.868469 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z6xmc\" (UniqueName: \"kubernetes.io/projected/7f89c45d-7de2-4b7b-93da-ea210d605384-kube-api-access-z6xmc\") on node \"crc\" DevicePath \"\"" Feb 27 08:42:05 crc kubenswrapper[4612]: I0227 08:42:05.069649 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536362-cphx2" event={"ID":"7f89c45d-7de2-4b7b-93da-ea210d605384","Type":"ContainerDied","Data":"2fb76c33373bbb3102b45317e0b1b3f91eead5cb240207413beb750af757585a"} Feb 27 08:42:05 crc kubenswrapper[4612]: I0227 08:42:05.069971 4612 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2fb76c33373bbb3102b45317e0b1b3f91eead5cb240207413beb750af757585a" Feb 27 08:42:05 crc kubenswrapper[4612]: I0227 08:42:05.069755 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536362-cphx2" Feb 27 08:42:05 crc kubenswrapper[4612]: I0227 08:42:05.714418 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29536356-txvz5"] Feb 27 08:42:05 crc kubenswrapper[4612]: I0227 08:42:05.725687 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29536356-txvz5"] Feb 27 08:42:06 crc kubenswrapper[4612]: I0227 08:42:06.863032 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2b382b34-72d2-4144-a87c-50330a8396ad" path="/var/lib/kubelet/pods/2b382b34-72d2-4144-a87c-50330a8396ad/volumes" Feb 27 08:42:13 crc kubenswrapper[4612]: I0227 08:42:13.853648 4612 scope.go:117] "RemoveContainer" containerID="384f63ec5e89debe3f33788a53a38ede9872a5e515bb3ecc12869cd886606dba" Feb 27 08:42:13 crc kubenswrapper[4612]: E0227 08:42:13.854396 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-924vb_openshift-machine-config-operator(24599373-7adc-4a1b-8bb4-797bf726f43d)\"" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" Feb 27 08:42:18 crc kubenswrapper[4612]: I0227 08:42:18.595445 4612 scope.go:117] "RemoveContainer" containerID="524d1065917e1007784752569a9692b8810799b2065f018e1798ad7b08e5e650" Feb 27 08:42:27 crc kubenswrapper[4612]: I0227 08:42:27.853215 4612 scope.go:117] "RemoveContainer" containerID="384f63ec5e89debe3f33788a53a38ede9872a5e515bb3ecc12869cd886606dba" Feb 27 08:42:27 crc kubenswrapper[4612]: E0227 08:42:27.853965 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-924vb_openshift-machine-config-operator(24599373-7adc-4a1b-8bb4-797bf726f43d)\"" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" Feb 27 08:42:42 crc kubenswrapper[4612]: I0227 08:42:42.859825 4612 scope.go:117] "RemoveContainer" containerID="384f63ec5e89debe3f33788a53a38ede9872a5e515bb3ecc12869cd886606dba" Feb 27 08:42:42 crc kubenswrapper[4612]: E0227 08:42:42.860607 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-924vb_openshift-machine-config-operator(24599373-7adc-4a1b-8bb4-797bf726f43d)\"" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" Feb 27 08:42:55 crc kubenswrapper[4612]: I0227 08:42:55.852917 4612 scope.go:117] "RemoveContainer" containerID="384f63ec5e89debe3f33788a53a38ede9872a5e515bb3ecc12869cd886606dba" Feb 27 08:42:55 crc kubenswrapper[4612]: E0227 08:42:55.853627 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-924vb_openshift-machine-config-operator(24599373-7adc-4a1b-8bb4-797bf726f43d)\"" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" Feb 27 08:43:08 crc kubenswrapper[4612]: I0227 08:43:08.854397 4612 scope.go:117] "RemoveContainer" containerID="384f63ec5e89debe3f33788a53a38ede9872a5e515bb3ecc12869cd886606dba" Feb 27 08:43:08 crc kubenswrapper[4612]: E0227 08:43:08.855208 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-924vb_openshift-machine-config-operator(24599373-7adc-4a1b-8bb4-797bf726f43d)\"" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" Feb 27 08:43:22 crc kubenswrapper[4612]: I0227 08:43:22.860602 4612 scope.go:117] "RemoveContainer" containerID="384f63ec5e89debe3f33788a53a38ede9872a5e515bb3ecc12869cd886606dba" Feb 27 08:43:23 crc kubenswrapper[4612]: I0227 08:43:23.757337 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-924vb" event={"ID":"24599373-7adc-4a1b-8bb4-797bf726f43d","Type":"ContainerStarted","Data":"bf86b408e02a5e06a79d4d8bbbe514621c6f07302864c707bf02845f4bf9beb7"} Feb 27 08:44:00 crc kubenswrapper[4612]: I0227 08:44:00.149372 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29536364-4ztkb"] Feb 27 08:44:00 crc kubenswrapper[4612]: E0227 08:44:00.150278 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7f89c45d-7de2-4b7b-93da-ea210d605384" containerName="oc" Feb 27 08:44:00 crc kubenswrapper[4612]: I0227 08:44:00.150297 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="7f89c45d-7de2-4b7b-93da-ea210d605384" containerName="oc" Feb 27 08:44:00 crc kubenswrapper[4612]: I0227 08:44:00.150535 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="7f89c45d-7de2-4b7b-93da-ea210d605384" containerName="oc" Feb 27 08:44:00 crc kubenswrapper[4612]: I0227 08:44:00.151241 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536364-4ztkb" Feb 27 08:44:00 crc kubenswrapper[4612]: I0227 08:44:00.153818 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 27 08:44:00 crc kubenswrapper[4612]: I0227 08:44:00.154073 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-5zwfx" Feb 27 08:44:00 crc kubenswrapper[4612]: I0227 08:44:00.154098 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 27 08:44:00 crc kubenswrapper[4612]: I0227 08:44:00.164004 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536364-4ztkb"] Feb 27 08:44:00 crc kubenswrapper[4612]: I0227 08:44:00.321908 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cnkql\" (UniqueName: \"kubernetes.io/projected/14ea3683-b0e3-4aaf-bc0f-e6e3e15bbe41-kube-api-access-cnkql\") pod \"auto-csr-approver-29536364-4ztkb\" (UID: \"14ea3683-b0e3-4aaf-bc0f-e6e3e15bbe41\") " pod="openshift-infra/auto-csr-approver-29536364-4ztkb" Feb 27 08:44:00 crc kubenswrapper[4612]: I0227 08:44:00.423811 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cnkql\" (UniqueName: \"kubernetes.io/projected/14ea3683-b0e3-4aaf-bc0f-e6e3e15bbe41-kube-api-access-cnkql\") pod \"auto-csr-approver-29536364-4ztkb\" (UID: \"14ea3683-b0e3-4aaf-bc0f-e6e3e15bbe41\") " pod="openshift-infra/auto-csr-approver-29536364-4ztkb" Feb 27 08:44:00 crc kubenswrapper[4612]: I0227 08:44:00.451398 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cnkql\" (UniqueName: \"kubernetes.io/projected/14ea3683-b0e3-4aaf-bc0f-e6e3e15bbe41-kube-api-access-cnkql\") pod \"auto-csr-approver-29536364-4ztkb\" (UID: \"14ea3683-b0e3-4aaf-bc0f-e6e3e15bbe41\") " pod="openshift-infra/auto-csr-approver-29536364-4ztkb" Feb 27 08:44:00 crc kubenswrapper[4612]: I0227 08:44:00.474658 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536364-4ztkb" Feb 27 08:44:00 crc kubenswrapper[4612]: I0227 08:44:00.954212 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536364-4ztkb"] Feb 27 08:44:01 crc kubenswrapper[4612]: I0227 08:44:01.062508 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536364-4ztkb" event={"ID":"14ea3683-b0e3-4aaf-bc0f-e6e3e15bbe41","Type":"ContainerStarted","Data":"563e4f22496776ff5048832e1d1def4e889b036d5665d27e3a31695ae9e7835e"} Feb 27 08:44:03 crc kubenswrapper[4612]: I0227 08:44:03.081447 4612 generic.go:334] "Generic (PLEG): container finished" podID="14ea3683-b0e3-4aaf-bc0f-e6e3e15bbe41" containerID="c410fb9e5df505d0bd2d98d4d9cf5ec4aa03d9c3b71166796b145c010b5a121c" exitCode=0 Feb 27 08:44:03 crc kubenswrapper[4612]: I0227 08:44:03.081921 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536364-4ztkb" event={"ID":"14ea3683-b0e3-4aaf-bc0f-e6e3e15bbe41","Type":"ContainerDied","Data":"c410fb9e5df505d0bd2d98d4d9cf5ec4aa03d9c3b71166796b145c010b5a121c"} Feb 27 08:44:04 crc kubenswrapper[4612]: I0227 08:44:04.715637 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536364-4ztkb" Feb 27 08:44:04 crc kubenswrapper[4612]: I0227 08:44:04.834513 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cnkql\" (UniqueName: \"kubernetes.io/projected/14ea3683-b0e3-4aaf-bc0f-e6e3e15bbe41-kube-api-access-cnkql\") pod \"14ea3683-b0e3-4aaf-bc0f-e6e3e15bbe41\" (UID: \"14ea3683-b0e3-4aaf-bc0f-e6e3e15bbe41\") " Feb 27 08:44:04 crc kubenswrapper[4612]: I0227 08:44:04.850133 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/14ea3683-b0e3-4aaf-bc0f-e6e3e15bbe41-kube-api-access-cnkql" (OuterVolumeSpecName: "kube-api-access-cnkql") pod "14ea3683-b0e3-4aaf-bc0f-e6e3e15bbe41" (UID: "14ea3683-b0e3-4aaf-bc0f-e6e3e15bbe41"). InnerVolumeSpecName "kube-api-access-cnkql". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 08:44:04 crc kubenswrapper[4612]: I0227 08:44:04.937474 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cnkql\" (UniqueName: \"kubernetes.io/projected/14ea3683-b0e3-4aaf-bc0f-e6e3e15bbe41-kube-api-access-cnkql\") on node \"crc\" DevicePath \"\"" Feb 27 08:44:05 crc kubenswrapper[4612]: I0227 08:44:05.105618 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536364-4ztkb" event={"ID":"14ea3683-b0e3-4aaf-bc0f-e6e3e15bbe41","Type":"ContainerDied","Data":"563e4f22496776ff5048832e1d1def4e889b036d5665d27e3a31695ae9e7835e"} Feb 27 08:44:05 crc kubenswrapper[4612]: I0227 08:44:05.105657 4612 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="563e4f22496776ff5048832e1d1def4e889b036d5665d27e3a31695ae9e7835e" Feb 27 08:44:05 crc kubenswrapper[4612]: I0227 08:44:05.106121 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536364-4ztkb" Feb 27 08:44:05 crc kubenswrapper[4612]: I0227 08:44:05.803120 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29536358-mxzmn"] Feb 27 08:44:05 crc kubenswrapper[4612]: I0227 08:44:05.812059 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29536358-mxzmn"] Feb 27 08:44:06 crc kubenswrapper[4612]: I0227 08:44:06.863283 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1269c80c-6ede-4228-8d61-82efff8942e2" path="/var/lib/kubelet/pods/1269c80c-6ede-4228-8d61-82efff8942e2/volumes" Feb 27 08:45:00 crc kubenswrapper[4612]: I0227 08:45:00.155229 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29536365-xxx6n"] Feb 27 08:45:00 crc kubenswrapper[4612]: E0227 08:45:00.156245 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="14ea3683-b0e3-4aaf-bc0f-e6e3e15bbe41" containerName="oc" Feb 27 08:45:00 crc kubenswrapper[4612]: I0227 08:45:00.156268 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="14ea3683-b0e3-4aaf-bc0f-e6e3e15bbe41" containerName="oc" Feb 27 08:45:00 crc kubenswrapper[4612]: I0227 08:45:00.156502 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="14ea3683-b0e3-4aaf-bc0f-e6e3e15bbe41" containerName="oc" Feb 27 08:45:00 crc kubenswrapper[4612]: I0227 08:45:00.157534 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29536365-xxx6n" Feb 27 08:45:00 crc kubenswrapper[4612]: I0227 08:45:00.160649 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Feb 27 08:45:00 crc kubenswrapper[4612]: I0227 08:45:00.160996 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Feb 27 08:45:00 crc kubenswrapper[4612]: I0227 08:45:00.168387 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29536365-xxx6n"] Feb 27 08:45:00 crc kubenswrapper[4612]: I0227 08:45:00.297591 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/de48600c-9cf3-4679-98b5-382cd3424e9b-config-volume\") pod \"collect-profiles-29536365-xxx6n\" (UID: \"de48600c-9cf3-4679-98b5-382cd3424e9b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29536365-xxx6n" Feb 27 08:45:00 crc kubenswrapper[4612]: I0227 08:45:00.297778 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/de48600c-9cf3-4679-98b5-382cd3424e9b-secret-volume\") pod \"collect-profiles-29536365-xxx6n\" (UID: \"de48600c-9cf3-4679-98b5-382cd3424e9b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29536365-xxx6n" Feb 27 08:45:00 crc kubenswrapper[4612]: I0227 08:45:00.297813 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7gv6j\" (UniqueName: \"kubernetes.io/projected/de48600c-9cf3-4679-98b5-382cd3424e9b-kube-api-access-7gv6j\") pod \"collect-profiles-29536365-xxx6n\" (UID: \"de48600c-9cf3-4679-98b5-382cd3424e9b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29536365-xxx6n" Feb 27 08:45:00 crc kubenswrapper[4612]: I0227 08:45:00.399443 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/de48600c-9cf3-4679-98b5-382cd3424e9b-config-volume\") pod \"collect-profiles-29536365-xxx6n\" (UID: \"de48600c-9cf3-4679-98b5-382cd3424e9b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29536365-xxx6n" Feb 27 08:45:00 crc kubenswrapper[4612]: I0227 08:45:00.399569 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/de48600c-9cf3-4679-98b5-382cd3424e9b-secret-volume\") pod \"collect-profiles-29536365-xxx6n\" (UID: \"de48600c-9cf3-4679-98b5-382cd3424e9b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29536365-xxx6n" Feb 27 08:45:00 crc kubenswrapper[4612]: I0227 08:45:00.399591 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7gv6j\" (UniqueName: \"kubernetes.io/projected/de48600c-9cf3-4679-98b5-382cd3424e9b-kube-api-access-7gv6j\") pod \"collect-profiles-29536365-xxx6n\" (UID: \"de48600c-9cf3-4679-98b5-382cd3424e9b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29536365-xxx6n" Feb 27 08:45:00 crc kubenswrapper[4612]: I0227 08:45:00.400439 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/de48600c-9cf3-4679-98b5-382cd3424e9b-config-volume\") pod \"collect-profiles-29536365-xxx6n\" (UID: \"de48600c-9cf3-4679-98b5-382cd3424e9b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29536365-xxx6n" Feb 27 08:45:00 crc kubenswrapper[4612]: I0227 08:45:00.415685 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/de48600c-9cf3-4679-98b5-382cd3424e9b-secret-volume\") pod \"collect-profiles-29536365-xxx6n\" (UID: \"de48600c-9cf3-4679-98b5-382cd3424e9b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29536365-xxx6n" Feb 27 08:45:00 crc kubenswrapper[4612]: I0227 08:45:00.421393 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7gv6j\" (UniqueName: \"kubernetes.io/projected/de48600c-9cf3-4679-98b5-382cd3424e9b-kube-api-access-7gv6j\") pod \"collect-profiles-29536365-xxx6n\" (UID: \"de48600c-9cf3-4679-98b5-382cd3424e9b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29536365-xxx6n" Feb 27 08:45:00 crc kubenswrapper[4612]: I0227 08:45:00.490660 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29536365-xxx6n" Feb 27 08:45:00 crc kubenswrapper[4612]: I0227 08:45:00.961798 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29536365-xxx6n"] Feb 27 08:45:00 crc kubenswrapper[4612]: W0227 08:45:00.974631 4612 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podde48600c_9cf3_4679_98b5_382cd3424e9b.slice/crio-c2e30240b07cc20c831350911c0ebf626b58256f50e4598d5991630a69c282c3 WatchSource:0}: Error finding container c2e30240b07cc20c831350911c0ebf626b58256f50e4598d5991630a69c282c3: Status 404 returned error can't find the container with id c2e30240b07cc20c831350911c0ebf626b58256f50e4598d5991630a69c282c3 Feb 27 08:45:01 crc kubenswrapper[4612]: I0227 08:45:01.637005 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29536365-xxx6n" event={"ID":"de48600c-9cf3-4679-98b5-382cd3424e9b","Type":"ContainerStarted","Data":"63a2780dc594ba6b0f0815b897080a4e593c1e94784218bf90c47710d8d83cdd"} Feb 27 08:45:01 crc kubenswrapper[4612]: I0227 08:45:01.637355 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29536365-xxx6n" event={"ID":"de48600c-9cf3-4679-98b5-382cd3424e9b","Type":"ContainerStarted","Data":"c2e30240b07cc20c831350911c0ebf626b58256f50e4598d5991630a69c282c3"} Feb 27 08:45:01 crc kubenswrapper[4612]: I0227 08:45:01.656994 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29536365-xxx6n" podStartSLOduration=1.6569770510000001 podStartE2EDuration="1.656977051s" podCreationTimestamp="2026-02-27 08:45:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 08:45:01.649542349 +0000 UTC m=+3359.503472347" watchObservedRunningTime="2026-02-27 08:45:01.656977051 +0000 UTC m=+3359.510907049" Feb 27 08:45:02 crc kubenswrapper[4612]: I0227 08:45:02.652141 4612 generic.go:334] "Generic (PLEG): container finished" podID="de48600c-9cf3-4679-98b5-382cd3424e9b" containerID="63a2780dc594ba6b0f0815b897080a4e593c1e94784218bf90c47710d8d83cdd" exitCode=0 Feb 27 08:45:02 crc kubenswrapper[4612]: I0227 08:45:02.652473 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29536365-xxx6n" event={"ID":"de48600c-9cf3-4679-98b5-382cd3424e9b","Type":"ContainerDied","Data":"63a2780dc594ba6b0f0815b897080a4e593c1e94784218bf90c47710d8d83cdd"} Feb 27 08:45:04 crc kubenswrapper[4612]: I0227 08:45:04.266096 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29536365-xxx6n" Feb 27 08:45:04 crc kubenswrapper[4612]: I0227 08:45:04.374789 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/de48600c-9cf3-4679-98b5-382cd3424e9b-secret-volume\") pod \"de48600c-9cf3-4679-98b5-382cd3424e9b\" (UID: \"de48600c-9cf3-4679-98b5-382cd3424e9b\") " Feb 27 08:45:04 crc kubenswrapper[4612]: I0227 08:45:04.374972 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7gv6j\" (UniqueName: \"kubernetes.io/projected/de48600c-9cf3-4679-98b5-382cd3424e9b-kube-api-access-7gv6j\") pod \"de48600c-9cf3-4679-98b5-382cd3424e9b\" (UID: \"de48600c-9cf3-4679-98b5-382cd3424e9b\") " Feb 27 08:45:04 crc kubenswrapper[4612]: I0227 08:45:04.375134 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/de48600c-9cf3-4679-98b5-382cd3424e9b-config-volume\") pod \"de48600c-9cf3-4679-98b5-382cd3424e9b\" (UID: \"de48600c-9cf3-4679-98b5-382cd3424e9b\") " Feb 27 08:45:04 crc kubenswrapper[4612]: I0227 08:45:04.376314 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/de48600c-9cf3-4679-98b5-382cd3424e9b-config-volume" (OuterVolumeSpecName: "config-volume") pod "de48600c-9cf3-4679-98b5-382cd3424e9b" (UID: "de48600c-9cf3-4679-98b5-382cd3424e9b"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 08:45:04 crc kubenswrapper[4612]: I0227 08:45:04.382935 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/de48600c-9cf3-4679-98b5-382cd3424e9b-kube-api-access-7gv6j" (OuterVolumeSpecName: "kube-api-access-7gv6j") pod "de48600c-9cf3-4679-98b5-382cd3424e9b" (UID: "de48600c-9cf3-4679-98b5-382cd3424e9b"). InnerVolumeSpecName "kube-api-access-7gv6j". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 08:45:04 crc kubenswrapper[4612]: I0227 08:45:04.384963 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/de48600c-9cf3-4679-98b5-382cd3424e9b-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "de48600c-9cf3-4679-98b5-382cd3424e9b" (UID: "de48600c-9cf3-4679-98b5-382cd3424e9b"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:45:04 crc kubenswrapper[4612]: I0227 08:45:04.477363 4612 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/de48600c-9cf3-4679-98b5-382cd3424e9b-secret-volume\") on node \"crc\" DevicePath \"\"" Feb 27 08:45:04 crc kubenswrapper[4612]: I0227 08:45:04.477401 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7gv6j\" (UniqueName: \"kubernetes.io/projected/de48600c-9cf3-4679-98b5-382cd3424e9b-kube-api-access-7gv6j\") on node \"crc\" DevicePath \"\"" Feb 27 08:45:04 crc kubenswrapper[4612]: I0227 08:45:04.477414 4612 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/de48600c-9cf3-4679-98b5-382cd3424e9b-config-volume\") on node \"crc\" DevicePath \"\"" Feb 27 08:45:04 crc kubenswrapper[4612]: I0227 08:45:04.671790 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29536365-xxx6n" event={"ID":"de48600c-9cf3-4679-98b5-382cd3424e9b","Type":"ContainerDied","Data":"c2e30240b07cc20c831350911c0ebf626b58256f50e4598d5991630a69c282c3"} Feb 27 08:45:04 crc kubenswrapper[4612]: I0227 08:45:04.672086 4612 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c2e30240b07cc20c831350911c0ebf626b58256f50e4598d5991630a69c282c3" Feb 27 08:45:04 crc kubenswrapper[4612]: I0227 08:45:04.671856 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29536365-xxx6n" Feb 27 08:45:04 crc kubenswrapper[4612]: I0227 08:45:04.746287 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29536320-2skmc"] Feb 27 08:45:04 crc kubenswrapper[4612]: I0227 08:45:04.756771 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29536320-2skmc"] Feb 27 08:45:04 crc kubenswrapper[4612]: I0227 08:45:04.868918 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="52c3c89b-5721-4bc5-b1af-c6354236896f" path="/var/lib/kubelet/pods/52c3c89b-5721-4bc5-b1af-c6354236896f/volumes" Feb 27 08:45:18 crc kubenswrapper[4612]: I0227 08:45:18.774126 4612 scope.go:117] "RemoveContainer" containerID="9659d5d0ff3b5c78fcb3ea29d3fe15de918a7e2a964522b8abf3deb7f8cc283f" Feb 27 08:45:18 crc kubenswrapper[4612]: I0227 08:45:18.837533 4612 scope.go:117] "RemoveContainer" containerID="873b6e2e81a9933f9218fb5e670f8181c4c24427e897138f061f66159a964999" Feb 27 08:45:46 crc kubenswrapper[4612]: I0227 08:45:46.026830 4612 patch_prober.go:28] interesting pod/machine-config-daemon-924vb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 27 08:45:46 crc kubenswrapper[4612]: I0227 08:45:46.027463 4612 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 27 08:45:54 crc kubenswrapper[4612]: I0227 08:45:54.133453 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-fqbjs"] Feb 27 08:45:54 crc kubenswrapper[4612]: E0227 08:45:54.134289 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de48600c-9cf3-4679-98b5-382cd3424e9b" containerName="collect-profiles" Feb 27 08:45:54 crc kubenswrapper[4612]: I0227 08:45:54.134300 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="de48600c-9cf3-4679-98b5-382cd3424e9b" containerName="collect-profiles" Feb 27 08:45:54 crc kubenswrapper[4612]: I0227 08:45:54.134527 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="de48600c-9cf3-4679-98b5-382cd3424e9b" containerName="collect-profiles" Feb 27 08:45:54 crc kubenswrapper[4612]: I0227 08:45:54.135774 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fqbjs" Feb 27 08:45:54 crc kubenswrapper[4612]: I0227 08:45:54.147973 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-fqbjs"] Feb 27 08:45:54 crc kubenswrapper[4612]: I0227 08:45:54.282655 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7a6fed09-4aab-4d9b-9511-39b6bba3156d-utilities\") pod \"redhat-marketplace-fqbjs\" (UID: \"7a6fed09-4aab-4d9b-9511-39b6bba3156d\") " pod="openshift-marketplace/redhat-marketplace-fqbjs" Feb 27 08:45:54 crc kubenswrapper[4612]: I0227 08:45:54.282740 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-24stq\" (UniqueName: \"kubernetes.io/projected/7a6fed09-4aab-4d9b-9511-39b6bba3156d-kube-api-access-24stq\") pod \"redhat-marketplace-fqbjs\" (UID: \"7a6fed09-4aab-4d9b-9511-39b6bba3156d\") " pod="openshift-marketplace/redhat-marketplace-fqbjs" Feb 27 08:45:54 crc kubenswrapper[4612]: I0227 08:45:54.282878 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7a6fed09-4aab-4d9b-9511-39b6bba3156d-catalog-content\") pod \"redhat-marketplace-fqbjs\" (UID: \"7a6fed09-4aab-4d9b-9511-39b6bba3156d\") " pod="openshift-marketplace/redhat-marketplace-fqbjs" Feb 27 08:45:54 crc kubenswrapper[4612]: I0227 08:45:54.384420 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7a6fed09-4aab-4d9b-9511-39b6bba3156d-utilities\") pod \"redhat-marketplace-fqbjs\" (UID: \"7a6fed09-4aab-4d9b-9511-39b6bba3156d\") " pod="openshift-marketplace/redhat-marketplace-fqbjs" Feb 27 08:45:54 crc kubenswrapper[4612]: I0227 08:45:54.384485 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-24stq\" (UniqueName: \"kubernetes.io/projected/7a6fed09-4aab-4d9b-9511-39b6bba3156d-kube-api-access-24stq\") pod \"redhat-marketplace-fqbjs\" (UID: \"7a6fed09-4aab-4d9b-9511-39b6bba3156d\") " pod="openshift-marketplace/redhat-marketplace-fqbjs" Feb 27 08:45:54 crc kubenswrapper[4612]: I0227 08:45:54.384565 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7a6fed09-4aab-4d9b-9511-39b6bba3156d-catalog-content\") pod \"redhat-marketplace-fqbjs\" (UID: \"7a6fed09-4aab-4d9b-9511-39b6bba3156d\") " pod="openshift-marketplace/redhat-marketplace-fqbjs" Feb 27 08:45:54 crc kubenswrapper[4612]: I0227 08:45:54.384971 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7a6fed09-4aab-4d9b-9511-39b6bba3156d-catalog-content\") pod \"redhat-marketplace-fqbjs\" (UID: \"7a6fed09-4aab-4d9b-9511-39b6bba3156d\") " pod="openshift-marketplace/redhat-marketplace-fqbjs" Feb 27 08:45:54 crc kubenswrapper[4612]: I0227 08:45:54.385176 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7a6fed09-4aab-4d9b-9511-39b6bba3156d-utilities\") pod \"redhat-marketplace-fqbjs\" (UID: \"7a6fed09-4aab-4d9b-9511-39b6bba3156d\") " pod="openshift-marketplace/redhat-marketplace-fqbjs" Feb 27 08:45:54 crc kubenswrapper[4612]: I0227 08:45:54.415802 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-24stq\" (UniqueName: \"kubernetes.io/projected/7a6fed09-4aab-4d9b-9511-39b6bba3156d-kube-api-access-24stq\") pod \"redhat-marketplace-fqbjs\" (UID: \"7a6fed09-4aab-4d9b-9511-39b6bba3156d\") " pod="openshift-marketplace/redhat-marketplace-fqbjs" Feb 27 08:45:54 crc kubenswrapper[4612]: I0227 08:45:54.471790 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fqbjs" Feb 27 08:45:55 crc kubenswrapper[4612]: I0227 08:45:55.000190 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-fqbjs"] Feb 27 08:45:55 crc kubenswrapper[4612]: W0227 08:45:55.014968 4612 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7a6fed09_4aab_4d9b_9511_39b6bba3156d.slice/crio-e1c79e4989bccc95dd2369ab734bc89e2ed90d94656821f83a95ebf106b6625b WatchSource:0}: Error finding container e1c79e4989bccc95dd2369ab734bc89e2ed90d94656821f83a95ebf106b6625b: Status 404 returned error can't find the container with id e1c79e4989bccc95dd2369ab734bc89e2ed90d94656821f83a95ebf106b6625b Feb 27 08:45:55 crc kubenswrapper[4612]: I0227 08:45:55.138049 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fqbjs" event={"ID":"7a6fed09-4aab-4d9b-9511-39b6bba3156d","Type":"ContainerStarted","Data":"e1c79e4989bccc95dd2369ab734bc89e2ed90d94656821f83a95ebf106b6625b"} Feb 27 08:45:56 crc kubenswrapper[4612]: I0227 08:45:56.153131 4612 generic.go:334] "Generic (PLEG): container finished" podID="7a6fed09-4aab-4d9b-9511-39b6bba3156d" containerID="822960bc6eb57b3c5c3d056e1838973861d41f80e7feb6df9f44bcbc8ab56487" exitCode=0 Feb 27 08:45:56 crc kubenswrapper[4612]: I0227 08:45:56.153213 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fqbjs" event={"ID":"7a6fed09-4aab-4d9b-9511-39b6bba3156d","Type":"ContainerDied","Data":"822960bc6eb57b3c5c3d056e1838973861d41f80e7feb6df9f44bcbc8ab56487"} Feb 27 08:45:58 crc kubenswrapper[4612]: I0227 08:45:58.247575 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fqbjs" event={"ID":"7a6fed09-4aab-4d9b-9511-39b6bba3156d","Type":"ContainerStarted","Data":"719f2b64be6ebb9c949e2b1777092b83fdcfcea65bbde14aaea3cbc401bd8110"} Feb 27 08:46:00 crc kubenswrapper[4612]: I0227 08:46:00.145198 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29536366-xfkw7"] Feb 27 08:46:00 crc kubenswrapper[4612]: I0227 08:46:00.146554 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536366-xfkw7" Feb 27 08:46:00 crc kubenswrapper[4612]: I0227 08:46:00.152959 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 27 08:46:00 crc kubenswrapper[4612]: I0227 08:46:00.153813 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 27 08:46:00 crc kubenswrapper[4612]: I0227 08:46:00.155253 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-5zwfx" Feb 27 08:46:00 crc kubenswrapper[4612]: I0227 08:46:00.158031 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536366-xfkw7"] Feb 27 08:46:00 crc kubenswrapper[4612]: I0227 08:46:00.270707 4612 generic.go:334] "Generic (PLEG): container finished" podID="7a6fed09-4aab-4d9b-9511-39b6bba3156d" containerID="719f2b64be6ebb9c949e2b1777092b83fdcfcea65bbde14aaea3cbc401bd8110" exitCode=0 Feb 27 08:46:00 crc kubenswrapper[4612]: I0227 08:46:00.270790 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fqbjs" event={"ID":"7a6fed09-4aab-4d9b-9511-39b6bba3156d","Type":"ContainerDied","Data":"719f2b64be6ebb9c949e2b1777092b83fdcfcea65bbde14aaea3cbc401bd8110"} Feb 27 08:46:00 crc kubenswrapper[4612]: I0227 08:46:00.280773 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v85dz\" (UniqueName: \"kubernetes.io/projected/aa5037cd-a5d3-4edf-81e6-32d67a9d08b7-kube-api-access-v85dz\") pod \"auto-csr-approver-29536366-xfkw7\" (UID: \"aa5037cd-a5d3-4edf-81e6-32d67a9d08b7\") " pod="openshift-infra/auto-csr-approver-29536366-xfkw7" Feb 27 08:46:00 crc kubenswrapper[4612]: I0227 08:46:00.382941 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v85dz\" (UniqueName: \"kubernetes.io/projected/aa5037cd-a5d3-4edf-81e6-32d67a9d08b7-kube-api-access-v85dz\") pod \"auto-csr-approver-29536366-xfkw7\" (UID: \"aa5037cd-a5d3-4edf-81e6-32d67a9d08b7\") " pod="openshift-infra/auto-csr-approver-29536366-xfkw7" Feb 27 08:46:00 crc kubenswrapper[4612]: I0227 08:46:00.404338 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v85dz\" (UniqueName: \"kubernetes.io/projected/aa5037cd-a5d3-4edf-81e6-32d67a9d08b7-kube-api-access-v85dz\") pod \"auto-csr-approver-29536366-xfkw7\" (UID: \"aa5037cd-a5d3-4edf-81e6-32d67a9d08b7\") " pod="openshift-infra/auto-csr-approver-29536366-xfkw7" Feb 27 08:46:00 crc kubenswrapper[4612]: I0227 08:46:00.485479 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536366-xfkw7" Feb 27 08:46:01 crc kubenswrapper[4612]: I0227 08:46:01.626516 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536366-xfkw7"] Feb 27 08:46:02 crc kubenswrapper[4612]: I0227 08:46:02.294176 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536366-xfkw7" event={"ID":"aa5037cd-a5d3-4edf-81e6-32d67a9d08b7","Type":"ContainerStarted","Data":"83f3c7eee1f980e2252a33828e173175ebab5c95c146975fbe94550f6a54952f"} Feb 27 08:46:04 crc kubenswrapper[4612]: I0227 08:46:04.324005 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fqbjs" event={"ID":"7a6fed09-4aab-4d9b-9511-39b6bba3156d","Type":"ContainerStarted","Data":"37cc3113b75498a323590ab8bed8d89898f1189d8e4d12ed0d2eb38acf06b239"} Feb 27 08:46:04 crc kubenswrapper[4612]: I0227 08:46:04.348900 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-fqbjs" podStartSLOduration=3.110486189 podStartE2EDuration="10.34887848s" podCreationTimestamp="2026-02-27 08:45:54 +0000 UTC" firstStartedPulling="2026-02-27 08:45:56.155274334 +0000 UTC m=+3414.009204332" lastFinishedPulling="2026-02-27 08:46:03.393666625 +0000 UTC m=+3421.247596623" observedRunningTime="2026-02-27 08:46:04.341921341 +0000 UTC m=+3422.195851349" watchObservedRunningTime="2026-02-27 08:46:04.34887848 +0000 UTC m=+3422.202808478" Feb 27 08:46:04 crc kubenswrapper[4612]: I0227 08:46:04.473010 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-fqbjs" Feb 27 08:46:04 crc kubenswrapper[4612]: I0227 08:46:04.473215 4612 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-fqbjs" Feb 27 08:46:05 crc kubenswrapper[4612]: I0227 08:46:05.341169 4612 generic.go:334] "Generic (PLEG): container finished" podID="aa5037cd-a5d3-4edf-81e6-32d67a9d08b7" containerID="11b472f6096690caa8865dd995bcfff584e2fc3df506306aaa71ec91d363c02b" exitCode=0 Feb 27 08:46:05 crc kubenswrapper[4612]: I0227 08:46:05.341264 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536366-xfkw7" event={"ID":"aa5037cd-a5d3-4edf-81e6-32d67a9d08b7","Type":"ContainerDied","Data":"11b472f6096690caa8865dd995bcfff584e2fc3df506306aaa71ec91d363c02b"} Feb 27 08:46:05 crc kubenswrapper[4612]: I0227 08:46:05.621268 4612 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-marketplace-fqbjs" podUID="7a6fed09-4aab-4d9b-9511-39b6bba3156d" containerName="registry-server" probeResult="failure" output=< Feb 27 08:46:05 crc kubenswrapper[4612]: timeout: failed to connect service ":50051" within 1s Feb 27 08:46:05 crc kubenswrapper[4612]: > Feb 27 08:46:06 crc kubenswrapper[4612]: I0227 08:46:06.885172 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536366-xfkw7" Feb 27 08:46:07 crc kubenswrapper[4612]: I0227 08:46:07.015833 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v85dz\" (UniqueName: \"kubernetes.io/projected/aa5037cd-a5d3-4edf-81e6-32d67a9d08b7-kube-api-access-v85dz\") pod \"aa5037cd-a5d3-4edf-81e6-32d67a9d08b7\" (UID: \"aa5037cd-a5d3-4edf-81e6-32d67a9d08b7\") " Feb 27 08:46:07 crc kubenswrapper[4612]: I0227 08:46:07.032938 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aa5037cd-a5d3-4edf-81e6-32d67a9d08b7-kube-api-access-v85dz" (OuterVolumeSpecName: "kube-api-access-v85dz") pod "aa5037cd-a5d3-4edf-81e6-32d67a9d08b7" (UID: "aa5037cd-a5d3-4edf-81e6-32d67a9d08b7"). InnerVolumeSpecName "kube-api-access-v85dz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 08:46:07 crc kubenswrapper[4612]: I0227 08:46:07.118295 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v85dz\" (UniqueName: \"kubernetes.io/projected/aa5037cd-a5d3-4edf-81e6-32d67a9d08b7-kube-api-access-v85dz\") on node \"crc\" DevicePath \"\"" Feb 27 08:46:07 crc kubenswrapper[4612]: I0227 08:46:07.360430 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536366-xfkw7" event={"ID":"aa5037cd-a5d3-4edf-81e6-32d67a9d08b7","Type":"ContainerDied","Data":"83f3c7eee1f980e2252a33828e173175ebab5c95c146975fbe94550f6a54952f"} Feb 27 08:46:07 crc kubenswrapper[4612]: I0227 08:46:07.360811 4612 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="83f3c7eee1f980e2252a33828e173175ebab5c95c146975fbe94550f6a54952f" Feb 27 08:46:07 crc kubenswrapper[4612]: I0227 08:46:07.360612 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536366-xfkw7" Feb 27 08:46:07 crc kubenswrapper[4612]: I0227 08:46:07.984083 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29536360-pl294"] Feb 27 08:46:07 crc kubenswrapper[4612]: I0227 08:46:07.995016 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29536360-pl294"] Feb 27 08:46:08 crc kubenswrapper[4612]: I0227 08:46:08.865654 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="faff5924-030b-4a8a-883e-830c195454ea" path="/var/lib/kubelet/pods/faff5924-030b-4a8a-883e-830c195454ea/volumes" Feb 27 08:46:15 crc kubenswrapper[4612]: I0227 08:46:15.512332 4612 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-marketplace-fqbjs" podUID="7a6fed09-4aab-4d9b-9511-39b6bba3156d" containerName="registry-server" probeResult="failure" output=< Feb 27 08:46:15 crc kubenswrapper[4612]: timeout: failed to connect service ":50051" within 1s Feb 27 08:46:15 crc kubenswrapper[4612]: > Feb 27 08:46:16 crc kubenswrapper[4612]: I0227 08:46:16.026944 4612 patch_prober.go:28] interesting pod/machine-config-daemon-924vb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 27 08:46:16 crc kubenswrapper[4612]: I0227 08:46:16.026997 4612 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 27 08:46:18 crc kubenswrapper[4612]: I0227 08:46:18.949418 4612 scope.go:117] "RemoveContainer" containerID="76d3e2ca66bf6d6789f5256de1012d200acbda39a4c58d2c2515735b44cde6f6" Feb 27 08:46:24 crc kubenswrapper[4612]: I0227 08:46:24.525216 4612 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-fqbjs" Feb 27 08:46:24 crc kubenswrapper[4612]: I0227 08:46:24.578400 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-fqbjs" Feb 27 08:46:25 crc kubenswrapper[4612]: I0227 08:46:25.338208 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-fqbjs"] Feb 27 08:46:26 crc kubenswrapper[4612]: I0227 08:46:26.510499 4612 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-fqbjs" podUID="7a6fed09-4aab-4d9b-9511-39b6bba3156d" containerName="registry-server" containerID="cri-o://37cc3113b75498a323590ab8bed8d89898f1189d8e4d12ed0d2eb38acf06b239" gracePeriod=2 Feb 27 08:46:27 crc kubenswrapper[4612]: I0227 08:46:27.177366 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fqbjs" Feb 27 08:46:27 crc kubenswrapper[4612]: I0227 08:46:27.318121 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7a6fed09-4aab-4d9b-9511-39b6bba3156d-catalog-content\") pod \"7a6fed09-4aab-4d9b-9511-39b6bba3156d\" (UID: \"7a6fed09-4aab-4d9b-9511-39b6bba3156d\") " Feb 27 08:46:27 crc kubenswrapper[4612]: I0227 08:46:27.318279 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-24stq\" (UniqueName: \"kubernetes.io/projected/7a6fed09-4aab-4d9b-9511-39b6bba3156d-kube-api-access-24stq\") pod \"7a6fed09-4aab-4d9b-9511-39b6bba3156d\" (UID: \"7a6fed09-4aab-4d9b-9511-39b6bba3156d\") " Feb 27 08:46:27 crc kubenswrapper[4612]: I0227 08:46:27.318307 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7a6fed09-4aab-4d9b-9511-39b6bba3156d-utilities\") pod \"7a6fed09-4aab-4d9b-9511-39b6bba3156d\" (UID: \"7a6fed09-4aab-4d9b-9511-39b6bba3156d\") " Feb 27 08:46:27 crc kubenswrapper[4612]: I0227 08:46:27.319335 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7a6fed09-4aab-4d9b-9511-39b6bba3156d-utilities" (OuterVolumeSpecName: "utilities") pod "7a6fed09-4aab-4d9b-9511-39b6bba3156d" (UID: "7a6fed09-4aab-4d9b-9511-39b6bba3156d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 08:46:27 crc kubenswrapper[4612]: I0227 08:46:27.325149 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7a6fed09-4aab-4d9b-9511-39b6bba3156d-kube-api-access-24stq" (OuterVolumeSpecName: "kube-api-access-24stq") pod "7a6fed09-4aab-4d9b-9511-39b6bba3156d" (UID: "7a6fed09-4aab-4d9b-9511-39b6bba3156d"). InnerVolumeSpecName "kube-api-access-24stq". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 08:46:27 crc kubenswrapper[4612]: I0227 08:46:27.354380 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7a6fed09-4aab-4d9b-9511-39b6bba3156d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7a6fed09-4aab-4d9b-9511-39b6bba3156d" (UID: "7a6fed09-4aab-4d9b-9511-39b6bba3156d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 08:46:27 crc kubenswrapper[4612]: I0227 08:46:27.420521 4612 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7a6fed09-4aab-4d9b-9511-39b6bba3156d-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 27 08:46:27 crc kubenswrapper[4612]: I0227 08:46:27.420561 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-24stq\" (UniqueName: \"kubernetes.io/projected/7a6fed09-4aab-4d9b-9511-39b6bba3156d-kube-api-access-24stq\") on node \"crc\" DevicePath \"\"" Feb 27 08:46:27 crc kubenswrapper[4612]: I0227 08:46:27.420573 4612 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7a6fed09-4aab-4d9b-9511-39b6bba3156d-utilities\") on node \"crc\" DevicePath \"\"" Feb 27 08:46:27 crc kubenswrapper[4612]: I0227 08:46:27.521462 4612 generic.go:334] "Generic (PLEG): container finished" podID="7a6fed09-4aab-4d9b-9511-39b6bba3156d" containerID="37cc3113b75498a323590ab8bed8d89898f1189d8e4d12ed0d2eb38acf06b239" exitCode=0 Feb 27 08:46:27 crc kubenswrapper[4612]: I0227 08:46:27.521515 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fqbjs" event={"ID":"7a6fed09-4aab-4d9b-9511-39b6bba3156d","Type":"ContainerDied","Data":"37cc3113b75498a323590ab8bed8d89898f1189d8e4d12ed0d2eb38acf06b239"} Feb 27 08:46:27 crc kubenswrapper[4612]: I0227 08:46:27.521531 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fqbjs" Feb 27 08:46:27 crc kubenswrapper[4612]: I0227 08:46:27.521549 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fqbjs" event={"ID":"7a6fed09-4aab-4d9b-9511-39b6bba3156d","Type":"ContainerDied","Data":"e1c79e4989bccc95dd2369ab734bc89e2ed90d94656821f83a95ebf106b6625b"} Feb 27 08:46:27 crc kubenswrapper[4612]: I0227 08:46:27.521569 4612 scope.go:117] "RemoveContainer" containerID="37cc3113b75498a323590ab8bed8d89898f1189d8e4d12ed0d2eb38acf06b239" Feb 27 08:46:27 crc kubenswrapper[4612]: I0227 08:46:27.551399 4612 scope.go:117] "RemoveContainer" containerID="719f2b64be6ebb9c949e2b1777092b83fdcfcea65bbde14aaea3cbc401bd8110" Feb 27 08:46:27 crc kubenswrapper[4612]: I0227 08:46:27.572098 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-fqbjs"] Feb 27 08:46:27 crc kubenswrapper[4612]: I0227 08:46:27.572152 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-fqbjs"] Feb 27 08:46:27 crc kubenswrapper[4612]: I0227 08:46:27.597850 4612 scope.go:117] "RemoveContainer" containerID="822960bc6eb57b3c5c3d056e1838973861d41f80e7feb6df9f44bcbc8ab56487" Feb 27 08:46:27 crc kubenswrapper[4612]: I0227 08:46:27.637956 4612 scope.go:117] "RemoveContainer" containerID="37cc3113b75498a323590ab8bed8d89898f1189d8e4d12ed0d2eb38acf06b239" Feb 27 08:46:27 crc kubenswrapper[4612]: E0227 08:46:27.639078 4612 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"37cc3113b75498a323590ab8bed8d89898f1189d8e4d12ed0d2eb38acf06b239\": container with ID starting with 37cc3113b75498a323590ab8bed8d89898f1189d8e4d12ed0d2eb38acf06b239 not found: ID does not exist" containerID="37cc3113b75498a323590ab8bed8d89898f1189d8e4d12ed0d2eb38acf06b239" Feb 27 08:46:27 crc kubenswrapper[4612]: I0227 08:46:27.639116 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"37cc3113b75498a323590ab8bed8d89898f1189d8e4d12ed0d2eb38acf06b239"} err="failed to get container status \"37cc3113b75498a323590ab8bed8d89898f1189d8e4d12ed0d2eb38acf06b239\": rpc error: code = NotFound desc = could not find container \"37cc3113b75498a323590ab8bed8d89898f1189d8e4d12ed0d2eb38acf06b239\": container with ID starting with 37cc3113b75498a323590ab8bed8d89898f1189d8e4d12ed0d2eb38acf06b239 not found: ID does not exist" Feb 27 08:46:27 crc kubenswrapper[4612]: I0227 08:46:27.639144 4612 scope.go:117] "RemoveContainer" containerID="719f2b64be6ebb9c949e2b1777092b83fdcfcea65bbde14aaea3cbc401bd8110" Feb 27 08:46:27 crc kubenswrapper[4612]: E0227 08:46:27.639546 4612 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"719f2b64be6ebb9c949e2b1777092b83fdcfcea65bbde14aaea3cbc401bd8110\": container with ID starting with 719f2b64be6ebb9c949e2b1777092b83fdcfcea65bbde14aaea3cbc401bd8110 not found: ID does not exist" containerID="719f2b64be6ebb9c949e2b1777092b83fdcfcea65bbde14aaea3cbc401bd8110" Feb 27 08:46:27 crc kubenswrapper[4612]: I0227 08:46:27.639587 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"719f2b64be6ebb9c949e2b1777092b83fdcfcea65bbde14aaea3cbc401bd8110"} err="failed to get container status \"719f2b64be6ebb9c949e2b1777092b83fdcfcea65bbde14aaea3cbc401bd8110\": rpc error: code = NotFound desc = could not find container \"719f2b64be6ebb9c949e2b1777092b83fdcfcea65bbde14aaea3cbc401bd8110\": container with ID starting with 719f2b64be6ebb9c949e2b1777092b83fdcfcea65bbde14aaea3cbc401bd8110 not found: ID does not exist" Feb 27 08:46:27 crc kubenswrapper[4612]: I0227 08:46:27.639611 4612 scope.go:117] "RemoveContainer" containerID="822960bc6eb57b3c5c3d056e1838973861d41f80e7feb6df9f44bcbc8ab56487" Feb 27 08:46:27 crc kubenswrapper[4612]: E0227 08:46:27.639973 4612 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"822960bc6eb57b3c5c3d056e1838973861d41f80e7feb6df9f44bcbc8ab56487\": container with ID starting with 822960bc6eb57b3c5c3d056e1838973861d41f80e7feb6df9f44bcbc8ab56487 not found: ID does not exist" containerID="822960bc6eb57b3c5c3d056e1838973861d41f80e7feb6df9f44bcbc8ab56487" Feb 27 08:46:27 crc kubenswrapper[4612]: I0227 08:46:27.640017 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"822960bc6eb57b3c5c3d056e1838973861d41f80e7feb6df9f44bcbc8ab56487"} err="failed to get container status \"822960bc6eb57b3c5c3d056e1838973861d41f80e7feb6df9f44bcbc8ab56487\": rpc error: code = NotFound desc = could not find container \"822960bc6eb57b3c5c3d056e1838973861d41f80e7feb6df9f44bcbc8ab56487\": container with ID starting with 822960bc6eb57b3c5c3d056e1838973861d41f80e7feb6df9f44bcbc8ab56487 not found: ID does not exist" Feb 27 08:46:28 crc kubenswrapper[4612]: I0227 08:46:28.865020 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7a6fed09-4aab-4d9b-9511-39b6bba3156d" path="/var/lib/kubelet/pods/7a6fed09-4aab-4d9b-9511-39b6bba3156d/volumes" Feb 27 08:46:46 crc kubenswrapper[4612]: I0227 08:46:46.026949 4612 patch_prober.go:28] interesting pod/machine-config-daemon-924vb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 27 08:46:46 crc kubenswrapper[4612]: I0227 08:46:46.028570 4612 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 27 08:46:46 crc kubenswrapper[4612]: I0227 08:46:46.028750 4612 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-924vb" Feb 27 08:46:46 crc kubenswrapper[4612]: I0227 08:46:46.029668 4612 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"bf86b408e02a5e06a79d4d8bbbe514621c6f07302864c707bf02845f4bf9beb7"} pod="openshift-machine-config-operator/machine-config-daemon-924vb" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 27 08:46:46 crc kubenswrapper[4612]: I0227 08:46:46.029883 4612 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" containerName="machine-config-daemon" containerID="cri-o://bf86b408e02a5e06a79d4d8bbbe514621c6f07302864c707bf02845f4bf9beb7" gracePeriod=600 Feb 27 08:46:46 crc kubenswrapper[4612]: I0227 08:46:46.744193 4612 generic.go:334] "Generic (PLEG): container finished" podID="24599373-7adc-4a1b-8bb4-797bf726f43d" containerID="bf86b408e02a5e06a79d4d8bbbe514621c6f07302864c707bf02845f4bf9beb7" exitCode=0 Feb 27 08:46:46 crc kubenswrapper[4612]: I0227 08:46:46.744503 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-924vb" event={"ID":"24599373-7adc-4a1b-8bb4-797bf726f43d","Type":"ContainerDied","Data":"bf86b408e02a5e06a79d4d8bbbe514621c6f07302864c707bf02845f4bf9beb7"} Feb 27 08:46:46 crc kubenswrapper[4612]: I0227 08:46:46.744577 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-924vb" event={"ID":"24599373-7adc-4a1b-8bb4-797bf726f43d","Type":"ContainerStarted","Data":"2ed59616da422731358981fd9cdfe7932267d24e60d616b16a3481f816f614da"} Feb 27 08:46:46 crc kubenswrapper[4612]: I0227 08:46:46.744607 4612 scope.go:117] "RemoveContainer" containerID="384f63ec5e89debe3f33788a53a38ede9872a5e515bb3ecc12869cd886606dba" Feb 27 08:47:03 crc kubenswrapper[4612]: I0227 08:47:03.042978 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-rgd9s"] Feb 27 08:47:03 crc kubenswrapper[4612]: E0227 08:47:03.044029 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa5037cd-a5d3-4edf-81e6-32d67a9d08b7" containerName="oc" Feb 27 08:47:03 crc kubenswrapper[4612]: I0227 08:47:03.044045 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa5037cd-a5d3-4edf-81e6-32d67a9d08b7" containerName="oc" Feb 27 08:47:03 crc kubenswrapper[4612]: E0227 08:47:03.044073 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a6fed09-4aab-4d9b-9511-39b6bba3156d" containerName="extract-content" Feb 27 08:47:03 crc kubenswrapper[4612]: I0227 08:47:03.044082 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a6fed09-4aab-4d9b-9511-39b6bba3156d" containerName="extract-content" Feb 27 08:47:03 crc kubenswrapper[4612]: E0227 08:47:03.044103 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a6fed09-4aab-4d9b-9511-39b6bba3156d" containerName="extract-utilities" Feb 27 08:47:03 crc kubenswrapper[4612]: I0227 08:47:03.044111 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a6fed09-4aab-4d9b-9511-39b6bba3156d" containerName="extract-utilities" Feb 27 08:47:03 crc kubenswrapper[4612]: E0227 08:47:03.044126 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a6fed09-4aab-4d9b-9511-39b6bba3156d" containerName="registry-server" Feb 27 08:47:03 crc kubenswrapper[4612]: I0227 08:47:03.044133 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a6fed09-4aab-4d9b-9511-39b6bba3156d" containerName="registry-server" Feb 27 08:47:03 crc kubenswrapper[4612]: I0227 08:47:03.044355 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="aa5037cd-a5d3-4edf-81e6-32d67a9d08b7" containerName="oc" Feb 27 08:47:03 crc kubenswrapper[4612]: I0227 08:47:03.044388 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="7a6fed09-4aab-4d9b-9511-39b6bba3156d" containerName="registry-server" Feb 27 08:47:03 crc kubenswrapper[4612]: I0227 08:47:03.046194 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rgd9s" Feb 27 08:47:03 crc kubenswrapper[4612]: I0227 08:47:03.067661 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-rgd9s"] Feb 27 08:47:03 crc kubenswrapper[4612]: I0227 08:47:03.205456 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nf69v\" (UniqueName: \"kubernetes.io/projected/6b1d40f1-b500-47bf-adb1-a08a745fecbf-kube-api-access-nf69v\") pod \"certified-operators-rgd9s\" (UID: \"6b1d40f1-b500-47bf-adb1-a08a745fecbf\") " pod="openshift-marketplace/certified-operators-rgd9s" Feb 27 08:47:03 crc kubenswrapper[4612]: I0227 08:47:03.205493 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6b1d40f1-b500-47bf-adb1-a08a745fecbf-catalog-content\") pod \"certified-operators-rgd9s\" (UID: \"6b1d40f1-b500-47bf-adb1-a08a745fecbf\") " pod="openshift-marketplace/certified-operators-rgd9s" Feb 27 08:47:03 crc kubenswrapper[4612]: I0227 08:47:03.205992 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6b1d40f1-b500-47bf-adb1-a08a745fecbf-utilities\") pod \"certified-operators-rgd9s\" (UID: \"6b1d40f1-b500-47bf-adb1-a08a745fecbf\") " pod="openshift-marketplace/certified-operators-rgd9s" Feb 27 08:47:03 crc kubenswrapper[4612]: I0227 08:47:03.307447 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6b1d40f1-b500-47bf-adb1-a08a745fecbf-utilities\") pod \"certified-operators-rgd9s\" (UID: \"6b1d40f1-b500-47bf-adb1-a08a745fecbf\") " pod="openshift-marketplace/certified-operators-rgd9s" Feb 27 08:47:03 crc kubenswrapper[4612]: I0227 08:47:03.307542 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nf69v\" (UniqueName: \"kubernetes.io/projected/6b1d40f1-b500-47bf-adb1-a08a745fecbf-kube-api-access-nf69v\") pod \"certified-operators-rgd9s\" (UID: \"6b1d40f1-b500-47bf-adb1-a08a745fecbf\") " pod="openshift-marketplace/certified-operators-rgd9s" Feb 27 08:47:03 crc kubenswrapper[4612]: I0227 08:47:03.307560 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6b1d40f1-b500-47bf-adb1-a08a745fecbf-catalog-content\") pod \"certified-operators-rgd9s\" (UID: \"6b1d40f1-b500-47bf-adb1-a08a745fecbf\") " pod="openshift-marketplace/certified-operators-rgd9s" Feb 27 08:47:03 crc kubenswrapper[4612]: I0227 08:47:03.308107 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6b1d40f1-b500-47bf-adb1-a08a745fecbf-catalog-content\") pod \"certified-operators-rgd9s\" (UID: \"6b1d40f1-b500-47bf-adb1-a08a745fecbf\") " pod="openshift-marketplace/certified-operators-rgd9s" Feb 27 08:47:03 crc kubenswrapper[4612]: I0227 08:47:03.308317 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6b1d40f1-b500-47bf-adb1-a08a745fecbf-utilities\") pod \"certified-operators-rgd9s\" (UID: \"6b1d40f1-b500-47bf-adb1-a08a745fecbf\") " pod="openshift-marketplace/certified-operators-rgd9s" Feb 27 08:47:03 crc kubenswrapper[4612]: I0227 08:47:03.336367 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nf69v\" (UniqueName: \"kubernetes.io/projected/6b1d40f1-b500-47bf-adb1-a08a745fecbf-kube-api-access-nf69v\") pod \"certified-operators-rgd9s\" (UID: \"6b1d40f1-b500-47bf-adb1-a08a745fecbf\") " pod="openshift-marketplace/certified-operators-rgd9s" Feb 27 08:47:03 crc kubenswrapper[4612]: I0227 08:47:03.397370 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rgd9s" Feb 27 08:47:03 crc kubenswrapper[4612]: I0227 08:47:03.943117 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-rgd9s"] Feb 27 08:47:04 crc kubenswrapper[4612]: I0227 08:47:04.921757 4612 generic.go:334] "Generic (PLEG): container finished" podID="6b1d40f1-b500-47bf-adb1-a08a745fecbf" containerID="310e7295d58b41be99c181a426495c7c2a6300a1579f5515488372527b3fbcb5" exitCode=0 Feb 27 08:47:04 crc kubenswrapper[4612]: I0227 08:47:04.922308 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rgd9s" event={"ID":"6b1d40f1-b500-47bf-adb1-a08a745fecbf","Type":"ContainerDied","Data":"310e7295d58b41be99c181a426495c7c2a6300a1579f5515488372527b3fbcb5"} Feb 27 08:47:04 crc kubenswrapper[4612]: I0227 08:47:04.922337 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rgd9s" event={"ID":"6b1d40f1-b500-47bf-adb1-a08a745fecbf","Type":"ContainerStarted","Data":"428db9f46180c0d7e4aa0cc5fb3ac8e0441760d1ac7bff33b3a2696719940a35"} Feb 27 08:47:04 crc kubenswrapper[4612]: I0227 08:47:04.925500 4612 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 27 08:47:05 crc kubenswrapper[4612]: I0227 08:47:05.932050 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rgd9s" event={"ID":"6b1d40f1-b500-47bf-adb1-a08a745fecbf","Type":"ContainerStarted","Data":"847d05d2ee0442c70bcc4fcabc5bad2b8fbfd1ba82376f871f2cb306bb5b160d"} Feb 27 08:47:09 crc kubenswrapper[4612]: I0227 08:47:09.968619 4612 generic.go:334] "Generic (PLEG): container finished" podID="6b1d40f1-b500-47bf-adb1-a08a745fecbf" containerID="847d05d2ee0442c70bcc4fcabc5bad2b8fbfd1ba82376f871f2cb306bb5b160d" exitCode=0 Feb 27 08:47:09 crc kubenswrapper[4612]: I0227 08:47:09.969191 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rgd9s" event={"ID":"6b1d40f1-b500-47bf-adb1-a08a745fecbf","Type":"ContainerDied","Data":"847d05d2ee0442c70bcc4fcabc5bad2b8fbfd1ba82376f871f2cb306bb5b160d"} Feb 27 08:47:10 crc kubenswrapper[4612]: I0227 08:47:10.980422 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rgd9s" event={"ID":"6b1d40f1-b500-47bf-adb1-a08a745fecbf","Type":"ContainerStarted","Data":"a01a7ccddbe32e35750e7477f070bb18aa77b4b9bbfbf51763f651ab18431b4b"} Feb 27 08:47:11 crc kubenswrapper[4612]: I0227 08:47:11.005093 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-rgd9s" podStartSLOduration=2.572846621 podStartE2EDuration="8.005070375s" podCreationTimestamp="2026-02-27 08:47:03 +0000 UTC" firstStartedPulling="2026-02-27 08:47:04.925115591 +0000 UTC m=+3482.779045589" lastFinishedPulling="2026-02-27 08:47:10.357339335 +0000 UTC m=+3488.211269343" observedRunningTime="2026-02-27 08:47:10.996494401 +0000 UTC m=+3488.850424399" watchObservedRunningTime="2026-02-27 08:47:11.005070375 +0000 UTC m=+3488.859000373" Feb 27 08:47:13 crc kubenswrapper[4612]: I0227 08:47:13.398886 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-rgd9s" Feb 27 08:47:13 crc kubenswrapper[4612]: I0227 08:47:13.399306 4612 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-rgd9s" Feb 27 08:47:13 crc kubenswrapper[4612]: I0227 08:47:13.450076 4612 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-rgd9s" Feb 27 08:47:23 crc kubenswrapper[4612]: I0227 08:47:23.466915 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-rgd9s" Feb 27 08:47:23 crc kubenswrapper[4612]: I0227 08:47:23.529648 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-rgd9s"] Feb 27 08:47:24 crc kubenswrapper[4612]: I0227 08:47:24.103070 4612 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-rgd9s" podUID="6b1d40f1-b500-47bf-adb1-a08a745fecbf" containerName="registry-server" containerID="cri-o://a01a7ccddbe32e35750e7477f070bb18aa77b4b9bbfbf51763f651ab18431b4b" gracePeriod=2 Feb 27 08:47:24 crc kubenswrapper[4612]: I0227 08:47:24.674476 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rgd9s" Feb 27 08:47:24 crc kubenswrapper[4612]: I0227 08:47:24.819920 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6b1d40f1-b500-47bf-adb1-a08a745fecbf-catalog-content\") pod \"6b1d40f1-b500-47bf-adb1-a08a745fecbf\" (UID: \"6b1d40f1-b500-47bf-adb1-a08a745fecbf\") " Feb 27 08:47:24 crc kubenswrapper[4612]: I0227 08:47:24.820019 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6b1d40f1-b500-47bf-adb1-a08a745fecbf-utilities\") pod \"6b1d40f1-b500-47bf-adb1-a08a745fecbf\" (UID: \"6b1d40f1-b500-47bf-adb1-a08a745fecbf\") " Feb 27 08:47:24 crc kubenswrapper[4612]: I0227 08:47:24.820060 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nf69v\" (UniqueName: \"kubernetes.io/projected/6b1d40f1-b500-47bf-adb1-a08a745fecbf-kube-api-access-nf69v\") pod \"6b1d40f1-b500-47bf-adb1-a08a745fecbf\" (UID: \"6b1d40f1-b500-47bf-adb1-a08a745fecbf\") " Feb 27 08:47:24 crc kubenswrapper[4612]: I0227 08:47:24.822959 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6b1d40f1-b500-47bf-adb1-a08a745fecbf-utilities" (OuterVolumeSpecName: "utilities") pod "6b1d40f1-b500-47bf-adb1-a08a745fecbf" (UID: "6b1d40f1-b500-47bf-adb1-a08a745fecbf"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 08:47:24 crc kubenswrapper[4612]: I0227 08:47:24.840882 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6b1d40f1-b500-47bf-adb1-a08a745fecbf-kube-api-access-nf69v" (OuterVolumeSpecName: "kube-api-access-nf69v") pod "6b1d40f1-b500-47bf-adb1-a08a745fecbf" (UID: "6b1d40f1-b500-47bf-adb1-a08a745fecbf"). InnerVolumeSpecName "kube-api-access-nf69v". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 08:47:24 crc kubenswrapper[4612]: I0227 08:47:24.879866 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6b1d40f1-b500-47bf-adb1-a08a745fecbf-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6b1d40f1-b500-47bf-adb1-a08a745fecbf" (UID: "6b1d40f1-b500-47bf-adb1-a08a745fecbf"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 08:47:24 crc kubenswrapper[4612]: I0227 08:47:24.921972 4612 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6b1d40f1-b500-47bf-adb1-a08a745fecbf-utilities\") on node \"crc\" DevicePath \"\"" Feb 27 08:47:24 crc kubenswrapper[4612]: I0227 08:47:24.921997 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nf69v\" (UniqueName: \"kubernetes.io/projected/6b1d40f1-b500-47bf-adb1-a08a745fecbf-kube-api-access-nf69v\") on node \"crc\" DevicePath \"\"" Feb 27 08:47:24 crc kubenswrapper[4612]: I0227 08:47:24.922008 4612 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6b1d40f1-b500-47bf-adb1-a08a745fecbf-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 27 08:47:25 crc kubenswrapper[4612]: I0227 08:47:25.112183 4612 generic.go:334] "Generic (PLEG): container finished" podID="6b1d40f1-b500-47bf-adb1-a08a745fecbf" containerID="a01a7ccddbe32e35750e7477f070bb18aa77b4b9bbfbf51763f651ab18431b4b" exitCode=0 Feb 27 08:47:25 crc kubenswrapper[4612]: I0227 08:47:25.112237 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rgd9s" event={"ID":"6b1d40f1-b500-47bf-adb1-a08a745fecbf","Type":"ContainerDied","Data":"a01a7ccddbe32e35750e7477f070bb18aa77b4b9bbfbf51763f651ab18431b4b"} Feb 27 08:47:25 crc kubenswrapper[4612]: I0227 08:47:25.112288 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rgd9s" event={"ID":"6b1d40f1-b500-47bf-adb1-a08a745fecbf","Type":"ContainerDied","Data":"428db9f46180c0d7e4aa0cc5fb3ac8e0441760d1ac7bff33b3a2696719940a35"} Feb 27 08:47:25 crc kubenswrapper[4612]: I0227 08:47:25.112308 4612 scope.go:117] "RemoveContainer" containerID="a01a7ccddbe32e35750e7477f070bb18aa77b4b9bbfbf51763f651ab18431b4b" Feb 27 08:47:25 crc kubenswrapper[4612]: I0227 08:47:25.112452 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rgd9s" Feb 27 08:47:25 crc kubenswrapper[4612]: I0227 08:47:25.134931 4612 scope.go:117] "RemoveContainer" containerID="847d05d2ee0442c70bcc4fcabc5bad2b8fbfd1ba82376f871f2cb306bb5b160d" Feb 27 08:47:25 crc kubenswrapper[4612]: I0227 08:47:25.152654 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-rgd9s"] Feb 27 08:47:25 crc kubenswrapper[4612]: I0227 08:47:25.164360 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-rgd9s"] Feb 27 08:47:25 crc kubenswrapper[4612]: I0227 08:47:25.169053 4612 scope.go:117] "RemoveContainer" containerID="310e7295d58b41be99c181a426495c7c2a6300a1579f5515488372527b3fbcb5" Feb 27 08:47:25 crc kubenswrapper[4612]: I0227 08:47:25.219293 4612 scope.go:117] "RemoveContainer" containerID="a01a7ccddbe32e35750e7477f070bb18aa77b4b9bbfbf51763f651ab18431b4b" Feb 27 08:47:25 crc kubenswrapper[4612]: E0227 08:47:25.220204 4612 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a01a7ccddbe32e35750e7477f070bb18aa77b4b9bbfbf51763f651ab18431b4b\": container with ID starting with a01a7ccddbe32e35750e7477f070bb18aa77b4b9bbfbf51763f651ab18431b4b not found: ID does not exist" containerID="a01a7ccddbe32e35750e7477f070bb18aa77b4b9bbfbf51763f651ab18431b4b" Feb 27 08:47:25 crc kubenswrapper[4612]: I0227 08:47:25.220237 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a01a7ccddbe32e35750e7477f070bb18aa77b4b9bbfbf51763f651ab18431b4b"} err="failed to get container status \"a01a7ccddbe32e35750e7477f070bb18aa77b4b9bbfbf51763f651ab18431b4b\": rpc error: code = NotFound desc = could not find container \"a01a7ccddbe32e35750e7477f070bb18aa77b4b9bbfbf51763f651ab18431b4b\": container with ID starting with a01a7ccddbe32e35750e7477f070bb18aa77b4b9bbfbf51763f651ab18431b4b not found: ID does not exist" Feb 27 08:47:25 crc kubenswrapper[4612]: I0227 08:47:25.220257 4612 scope.go:117] "RemoveContainer" containerID="847d05d2ee0442c70bcc4fcabc5bad2b8fbfd1ba82376f871f2cb306bb5b160d" Feb 27 08:47:25 crc kubenswrapper[4612]: E0227 08:47:25.220553 4612 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"847d05d2ee0442c70bcc4fcabc5bad2b8fbfd1ba82376f871f2cb306bb5b160d\": container with ID starting with 847d05d2ee0442c70bcc4fcabc5bad2b8fbfd1ba82376f871f2cb306bb5b160d not found: ID does not exist" containerID="847d05d2ee0442c70bcc4fcabc5bad2b8fbfd1ba82376f871f2cb306bb5b160d" Feb 27 08:47:25 crc kubenswrapper[4612]: I0227 08:47:25.220582 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"847d05d2ee0442c70bcc4fcabc5bad2b8fbfd1ba82376f871f2cb306bb5b160d"} err="failed to get container status \"847d05d2ee0442c70bcc4fcabc5bad2b8fbfd1ba82376f871f2cb306bb5b160d\": rpc error: code = NotFound desc = could not find container \"847d05d2ee0442c70bcc4fcabc5bad2b8fbfd1ba82376f871f2cb306bb5b160d\": container with ID starting with 847d05d2ee0442c70bcc4fcabc5bad2b8fbfd1ba82376f871f2cb306bb5b160d not found: ID does not exist" Feb 27 08:47:25 crc kubenswrapper[4612]: I0227 08:47:25.220602 4612 scope.go:117] "RemoveContainer" containerID="310e7295d58b41be99c181a426495c7c2a6300a1579f5515488372527b3fbcb5" Feb 27 08:47:25 crc kubenswrapper[4612]: E0227 08:47:25.220809 4612 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"310e7295d58b41be99c181a426495c7c2a6300a1579f5515488372527b3fbcb5\": container with ID starting with 310e7295d58b41be99c181a426495c7c2a6300a1579f5515488372527b3fbcb5 not found: ID does not exist" containerID="310e7295d58b41be99c181a426495c7c2a6300a1579f5515488372527b3fbcb5" Feb 27 08:47:25 crc kubenswrapper[4612]: I0227 08:47:25.220830 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"310e7295d58b41be99c181a426495c7c2a6300a1579f5515488372527b3fbcb5"} err="failed to get container status \"310e7295d58b41be99c181a426495c7c2a6300a1579f5515488372527b3fbcb5\": rpc error: code = NotFound desc = could not find container \"310e7295d58b41be99c181a426495c7c2a6300a1579f5515488372527b3fbcb5\": container with ID starting with 310e7295d58b41be99c181a426495c7c2a6300a1579f5515488372527b3fbcb5 not found: ID does not exist" Feb 27 08:47:26 crc kubenswrapper[4612]: I0227 08:47:26.866745 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6b1d40f1-b500-47bf-adb1-a08a745fecbf" path="/var/lib/kubelet/pods/6b1d40f1-b500-47bf-adb1-a08a745fecbf/volumes" Feb 27 08:47:33 crc kubenswrapper[4612]: I0227 08:47:33.831931 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-k67wt"] Feb 27 08:47:33 crc kubenswrapper[4612]: E0227 08:47:33.833012 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b1d40f1-b500-47bf-adb1-a08a745fecbf" containerName="extract-content" Feb 27 08:47:33 crc kubenswrapper[4612]: I0227 08:47:33.833027 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b1d40f1-b500-47bf-adb1-a08a745fecbf" containerName="extract-content" Feb 27 08:47:33 crc kubenswrapper[4612]: E0227 08:47:33.833070 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b1d40f1-b500-47bf-adb1-a08a745fecbf" containerName="extract-utilities" Feb 27 08:47:33 crc kubenswrapper[4612]: I0227 08:47:33.833078 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b1d40f1-b500-47bf-adb1-a08a745fecbf" containerName="extract-utilities" Feb 27 08:47:33 crc kubenswrapper[4612]: E0227 08:47:33.833091 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b1d40f1-b500-47bf-adb1-a08a745fecbf" containerName="registry-server" Feb 27 08:47:33 crc kubenswrapper[4612]: I0227 08:47:33.833099 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b1d40f1-b500-47bf-adb1-a08a745fecbf" containerName="registry-server" Feb 27 08:47:33 crc kubenswrapper[4612]: I0227 08:47:33.833322 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="6b1d40f1-b500-47bf-adb1-a08a745fecbf" containerName="registry-server" Feb 27 08:47:33 crc kubenswrapper[4612]: I0227 08:47:33.834992 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-k67wt" Feb 27 08:47:33 crc kubenswrapper[4612]: I0227 08:47:33.843604 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-k67wt"] Feb 27 08:47:33 crc kubenswrapper[4612]: I0227 08:47:33.845024 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7jsdw\" (UniqueName: \"kubernetes.io/projected/9ebc429e-a12b-4547-9d98-d2459194642a-kube-api-access-7jsdw\") pod \"redhat-operators-k67wt\" (UID: \"9ebc429e-a12b-4547-9d98-d2459194642a\") " pod="openshift-marketplace/redhat-operators-k67wt" Feb 27 08:47:33 crc kubenswrapper[4612]: I0227 08:47:33.845097 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9ebc429e-a12b-4547-9d98-d2459194642a-utilities\") pod \"redhat-operators-k67wt\" (UID: \"9ebc429e-a12b-4547-9d98-d2459194642a\") " pod="openshift-marketplace/redhat-operators-k67wt" Feb 27 08:47:33 crc kubenswrapper[4612]: I0227 08:47:33.845357 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9ebc429e-a12b-4547-9d98-d2459194642a-catalog-content\") pod \"redhat-operators-k67wt\" (UID: \"9ebc429e-a12b-4547-9d98-d2459194642a\") " pod="openshift-marketplace/redhat-operators-k67wt" Feb 27 08:47:33 crc kubenswrapper[4612]: I0227 08:47:33.947313 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7jsdw\" (UniqueName: \"kubernetes.io/projected/9ebc429e-a12b-4547-9d98-d2459194642a-kube-api-access-7jsdw\") pod \"redhat-operators-k67wt\" (UID: \"9ebc429e-a12b-4547-9d98-d2459194642a\") " pod="openshift-marketplace/redhat-operators-k67wt" Feb 27 08:47:33 crc kubenswrapper[4612]: I0227 08:47:33.947365 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9ebc429e-a12b-4547-9d98-d2459194642a-utilities\") pod \"redhat-operators-k67wt\" (UID: \"9ebc429e-a12b-4547-9d98-d2459194642a\") " pod="openshift-marketplace/redhat-operators-k67wt" Feb 27 08:47:33 crc kubenswrapper[4612]: I0227 08:47:33.947454 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9ebc429e-a12b-4547-9d98-d2459194642a-catalog-content\") pod \"redhat-operators-k67wt\" (UID: \"9ebc429e-a12b-4547-9d98-d2459194642a\") " pod="openshift-marketplace/redhat-operators-k67wt" Feb 27 08:47:33 crc kubenswrapper[4612]: I0227 08:47:33.948119 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9ebc429e-a12b-4547-9d98-d2459194642a-catalog-content\") pod \"redhat-operators-k67wt\" (UID: \"9ebc429e-a12b-4547-9d98-d2459194642a\") " pod="openshift-marketplace/redhat-operators-k67wt" Feb 27 08:47:33 crc kubenswrapper[4612]: I0227 08:47:33.948195 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9ebc429e-a12b-4547-9d98-d2459194642a-utilities\") pod \"redhat-operators-k67wt\" (UID: \"9ebc429e-a12b-4547-9d98-d2459194642a\") " pod="openshift-marketplace/redhat-operators-k67wt" Feb 27 08:47:33 crc kubenswrapper[4612]: I0227 08:47:33.967923 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7jsdw\" (UniqueName: \"kubernetes.io/projected/9ebc429e-a12b-4547-9d98-d2459194642a-kube-api-access-7jsdw\") pod \"redhat-operators-k67wt\" (UID: \"9ebc429e-a12b-4547-9d98-d2459194642a\") " pod="openshift-marketplace/redhat-operators-k67wt" Feb 27 08:47:34 crc kubenswrapper[4612]: I0227 08:47:34.157935 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-k67wt" Feb 27 08:47:34 crc kubenswrapper[4612]: I0227 08:47:34.675699 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-k67wt"] Feb 27 08:47:35 crc kubenswrapper[4612]: I0227 08:47:35.208295 4612 generic.go:334] "Generic (PLEG): container finished" podID="9ebc429e-a12b-4547-9d98-d2459194642a" containerID="ba4c7ee554e95f9f26ca5d030ff49d6cb147c68c8d9aac0553941f5b8c9b24ec" exitCode=0 Feb 27 08:47:35 crc kubenswrapper[4612]: I0227 08:47:35.209145 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-k67wt" event={"ID":"9ebc429e-a12b-4547-9d98-d2459194642a","Type":"ContainerDied","Data":"ba4c7ee554e95f9f26ca5d030ff49d6cb147c68c8d9aac0553941f5b8c9b24ec"} Feb 27 08:47:35 crc kubenswrapper[4612]: I0227 08:47:35.209190 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-k67wt" event={"ID":"9ebc429e-a12b-4547-9d98-d2459194642a","Type":"ContainerStarted","Data":"ae3690e3e858479fc21c2eefba5279f7a92f3d70de58ed547cd15c8ee3bdaef2"} Feb 27 08:47:37 crc kubenswrapper[4612]: I0227 08:47:37.226736 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-k67wt" event={"ID":"9ebc429e-a12b-4547-9d98-d2459194642a","Type":"ContainerStarted","Data":"9c616c2fd68f812eea535758f2514701d29f2f3daf88442d154c1d1f86f9d405"} Feb 27 08:47:43 crc kubenswrapper[4612]: I0227 08:47:43.286457 4612 generic.go:334] "Generic (PLEG): container finished" podID="9ebc429e-a12b-4547-9d98-d2459194642a" containerID="9c616c2fd68f812eea535758f2514701d29f2f3daf88442d154c1d1f86f9d405" exitCode=0 Feb 27 08:47:43 crc kubenswrapper[4612]: I0227 08:47:43.286576 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-k67wt" event={"ID":"9ebc429e-a12b-4547-9d98-d2459194642a","Type":"ContainerDied","Data":"9c616c2fd68f812eea535758f2514701d29f2f3daf88442d154c1d1f86f9d405"} Feb 27 08:47:44 crc kubenswrapper[4612]: I0227 08:47:44.297661 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-k67wt" event={"ID":"9ebc429e-a12b-4547-9d98-d2459194642a","Type":"ContainerStarted","Data":"67b9f2e4d7da062a5ac3239214252d579a1efc8485ab68dd028a36af63bf578e"} Feb 27 08:47:44 crc kubenswrapper[4612]: I0227 08:47:44.325628 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-k67wt" podStartSLOduration=2.576574067 podStartE2EDuration="11.325605292s" podCreationTimestamp="2026-02-27 08:47:33 +0000 UTC" firstStartedPulling="2026-02-27 08:47:35.210211475 +0000 UTC m=+3513.064141473" lastFinishedPulling="2026-02-27 08:47:43.95924269 +0000 UTC m=+3521.813172698" observedRunningTime="2026-02-27 08:47:44.316222495 +0000 UTC m=+3522.170152493" watchObservedRunningTime="2026-02-27 08:47:44.325605292 +0000 UTC m=+3522.179535290" Feb 27 08:47:54 crc kubenswrapper[4612]: I0227 08:47:54.158451 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-k67wt" Feb 27 08:47:54 crc kubenswrapper[4612]: I0227 08:47:54.159094 4612 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-k67wt" Feb 27 08:47:55 crc kubenswrapper[4612]: I0227 08:47:55.203846 4612 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-k67wt" podUID="9ebc429e-a12b-4547-9d98-d2459194642a" containerName="registry-server" probeResult="failure" output=< Feb 27 08:47:55 crc kubenswrapper[4612]: timeout: failed to connect service ":50051" within 1s Feb 27 08:47:55 crc kubenswrapper[4612]: > Feb 27 08:48:00 crc kubenswrapper[4612]: I0227 08:48:00.144441 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29536368-ftbbd"] Feb 27 08:48:00 crc kubenswrapper[4612]: I0227 08:48:00.145978 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536368-ftbbd" Feb 27 08:48:00 crc kubenswrapper[4612]: I0227 08:48:00.148576 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 27 08:48:00 crc kubenswrapper[4612]: I0227 08:48:00.148581 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-5zwfx" Feb 27 08:48:00 crc kubenswrapper[4612]: I0227 08:48:00.149192 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 27 08:48:00 crc kubenswrapper[4612]: I0227 08:48:00.155748 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536368-ftbbd"] Feb 27 08:48:00 crc kubenswrapper[4612]: I0227 08:48:00.236904 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lp7rg\" (UniqueName: \"kubernetes.io/projected/dfdd99d3-8279-46f6-a853-418a80e00f19-kube-api-access-lp7rg\") pod \"auto-csr-approver-29536368-ftbbd\" (UID: \"dfdd99d3-8279-46f6-a853-418a80e00f19\") " pod="openshift-infra/auto-csr-approver-29536368-ftbbd" Feb 27 08:48:00 crc kubenswrapper[4612]: I0227 08:48:00.338376 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lp7rg\" (UniqueName: \"kubernetes.io/projected/dfdd99d3-8279-46f6-a853-418a80e00f19-kube-api-access-lp7rg\") pod \"auto-csr-approver-29536368-ftbbd\" (UID: \"dfdd99d3-8279-46f6-a853-418a80e00f19\") " pod="openshift-infra/auto-csr-approver-29536368-ftbbd" Feb 27 08:48:00 crc kubenswrapper[4612]: I0227 08:48:00.359852 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lp7rg\" (UniqueName: \"kubernetes.io/projected/dfdd99d3-8279-46f6-a853-418a80e00f19-kube-api-access-lp7rg\") pod \"auto-csr-approver-29536368-ftbbd\" (UID: \"dfdd99d3-8279-46f6-a853-418a80e00f19\") " pod="openshift-infra/auto-csr-approver-29536368-ftbbd" Feb 27 08:48:00 crc kubenswrapper[4612]: I0227 08:48:00.469582 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536368-ftbbd" Feb 27 08:48:00 crc kubenswrapper[4612]: I0227 08:48:00.972394 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536368-ftbbd"] Feb 27 08:48:01 crc kubenswrapper[4612]: I0227 08:48:01.445635 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536368-ftbbd" event={"ID":"dfdd99d3-8279-46f6-a853-418a80e00f19","Type":"ContainerStarted","Data":"e306cb88b0558d400649c08688caa99a07180e6400831512d1fd6649b5d917d2"} Feb 27 08:48:02 crc kubenswrapper[4612]: I0227 08:48:02.454251 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536368-ftbbd" event={"ID":"dfdd99d3-8279-46f6-a853-418a80e00f19","Type":"ContainerStarted","Data":"851b5ed6baadc16ffbd744acc297674fdde486cc5393a8074492364f31670d24"} Feb 27 08:48:02 crc kubenswrapper[4612]: I0227 08:48:02.476877 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29536368-ftbbd" podStartSLOduration=1.571617979 podStartE2EDuration="2.47685967s" podCreationTimestamp="2026-02-27 08:48:00 +0000 UTC" firstStartedPulling="2026-02-27 08:48:00.985789193 +0000 UTC m=+3538.839719191" lastFinishedPulling="2026-02-27 08:48:01.891030884 +0000 UTC m=+3539.744960882" observedRunningTime="2026-02-27 08:48:02.467568716 +0000 UTC m=+3540.321498714" watchObservedRunningTime="2026-02-27 08:48:02.47685967 +0000 UTC m=+3540.330789668" Feb 27 08:48:03 crc kubenswrapper[4612]: I0227 08:48:03.463429 4612 generic.go:334] "Generic (PLEG): container finished" podID="dfdd99d3-8279-46f6-a853-418a80e00f19" containerID="851b5ed6baadc16ffbd744acc297674fdde486cc5393a8074492364f31670d24" exitCode=0 Feb 27 08:48:03 crc kubenswrapper[4612]: I0227 08:48:03.463793 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536368-ftbbd" event={"ID":"dfdd99d3-8279-46f6-a853-418a80e00f19","Type":"ContainerDied","Data":"851b5ed6baadc16ffbd744acc297674fdde486cc5393a8074492364f31670d24"} Feb 27 08:48:04 crc kubenswrapper[4612]: I0227 08:48:04.957554 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536368-ftbbd" Feb 27 08:48:05 crc kubenswrapper[4612]: I0227 08:48:05.046046 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lp7rg\" (UniqueName: \"kubernetes.io/projected/dfdd99d3-8279-46f6-a853-418a80e00f19-kube-api-access-lp7rg\") pod \"dfdd99d3-8279-46f6-a853-418a80e00f19\" (UID: \"dfdd99d3-8279-46f6-a853-418a80e00f19\") " Feb 27 08:48:05 crc kubenswrapper[4612]: I0227 08:48:05.051621 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dfdd99d3-8279-46f6-a853-418a80e00f19-kube-api-access-lp7rg" (OuterVolumeSpecName: "kube-api-access-lp7rg") pod "dfdd99d3-8279-46f6-a853-418a80e00f19" (UID: "dfdd99d3-8279-46f6-a853-418a80e00f19"). InnerVolumeSpecName "kube-api-access-lp7rg". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 08:48:05 crc kubenswrapper[4612]: I0227 08:48:05.148814 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lp7rg\" (UniqueName: \"kubernetes.io/projected/dfdd99d3-8279-46f6-a853-418a80e00f19-kube-api-access-lp7rg\") on node \"crc\" DevicePath \"\"" Feb 27 08:48:05 crc kubenswrapper[4612]: I0227 08:48:05.205217 4612 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-k67wt" podUID="9ebc429e-a12b-4547-9d98-d2459194642a" containerName="registry-server" probeResult="failure" output=< Feb 27 08:48:05 crc kubenswrapper[4612]: timeout: failed to connect service ":50051" within 1s Feb 27 08:48:05 crc kubenswrapper[4612]: > Feb 27 08:48:05 crc kubenswrapper[4612]: I0227 08:48:05.478938 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536368-ftbbd" event={"ID":"dfdd99d3-8279-46f6-a853-418a80e00f19","Type":"ContainerDied","Data":"e306cb88b0558d400649c08688caa99a07180e6400831512d1fd6649b5d917d2"} Feb 27 08:48:05 crc kubenswrapper[4612]: I0227 08:48:05.478976 4612 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e306cb88b0558d400649c08688caa99a07180e6400831512d1fd6649b5d917d2" Feb 27 08:48:05 crc kubenswrapper[4612]: I0227 08:48:05.479032 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536368-ftbbd" Feb 27 08:48:05 crc kubenswrapper[4612]: I0227 08:48:05.535602 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29536362-cphx2"] Feb 27 08:48:05 crc kubenswrapper[4612]: I0227 08:48:05.544401 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29536362-cphx2"] Feb 27 08:48:06 crc kubenswrapper[4612]: I0227 08:48:06.864850 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7f89c45d-7de2-4b7b-93da-ea210d605384" path="/var/lib/kubelet/pods/7f89c45d-7de2-4b7b-93da-ea210d605384/volumes" Feb 27 08:48:15 crc kubenswrapper[4612]: I0227 08:48:15.228933 4612 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-k67wt" podUID="9ebc429e-a12b-4547-9d98-d2459194642a" containerName="registry-server" probeResult="failure" output=< Feb 27 08:48:15 crc kubenswrapper[4612]: timeout: failed to connect service ":50051" within 1s Feb 27 08:48:15 crc kubenswrapper[4612]: > Feb 27 08:48:19 crc kubenswrapper[4612]: I0227 08:48:19.065646 4612 scope.go:117] "RemoveContainer" containerID="16c57272f23a9361528d5d4d7e2d957dcc474a740cf8ca2f0cb4ae3a6abb7126" Feb 27 08:48:24 crc kubenswrapper[4612]: I0227 08:48:24.228913 4612 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-k67wt" Feb 27 08:48:24 crc kubenswrapper[4612]: I0227 08:48:24.290240 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-k67wt" Feb 27 08:48:24 crc kubenswrapper[4612]: I0227 08:48:24.482997 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-k67wt"] Feb 27 08:48:25 crc kubenswrapper[4612]: I0227 08:48:25.670635 4612 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-k67wt" podUID="9ebc429e-a12b-4547-9d98-d2459194642a" containerName="registry-server" containerID="cri-o://67b9f2e4d7da062a5ac3239214252d579a1efc8485ab68dd028a36af63bf578e" gracePeriod=2 Feb 27 08:48:26 crc kubenswrapper[4612]: I0227 08:48:26.282182 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-k67wt" Feb 27 08:48:26 crc kubenswrapper[4612]: I0227 08:48:26.443221 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9ebc429e-a12b-4547-9d98-d2459194642a-utilities\") pod \"9ebc429e-a12b-4547-9d98-d2459194642a\" (UID: \"9ebc429e-a12b-4547-9d98-d2459194642a\") " Feb 27 08:48:26 crc kubenswrapper[4612]: I0227 08:48:26.443845 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7jsdw\" (UniqueName: \"kubernetes.io/projected/9ebc429e-a12b-4547-9d98-d2459194642a-kube-api-access-7jsdw\") pod \"9ebc429e-a12b-4547-9d98-d2459194642a\" (UID: \"9ebc429e-a12b-4547-9d98-d2459194642a\") " Feb 27 08:48:26 crc kubenswrapper[4612]: I0227 08:48:26.444122 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9ebc429e-a12b-4547-9d98-d2459194642a-utilities" (OuterVolumeSpecName: "utilities") pod "9ebc429e-a12b-4547-9d98-d2459194642a" (UID: "9ebc429e-a12b-4547-9d98-d2459194642a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 08:48:26 crc kubenswrapper[4612]: I0227 08:48:26.444138 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9ebc429e-a12b-4547-9d98-d2459194642a-catalog-content\") pod \"9ebc429e-a12b-4547-9d98-d2459194642a\" (UID: \"9ebc429e-a12b-4547-9d98-d2459194642a\") " Feb 27 08:48:26 crc kubenswrapper[4612]: I0227 08:48:26.445883 4612 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9ebc429e-a12b-4547-9d98-d2459194642a-utilities\") on node \"crc\" DevicePath \"\"" Feb 27 08:48:26 crc kubenswrapper[4612]: I0227 08:48:26.452587 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9ebc429e-a12b-4547-9d98-d2459194642a-kube-api-access-7jsdw" (OuterVolumeSpecName: "kube-api-access-7jsdw") pod "9ebc429e-a12b-4547-9d98-d2459194642a" (UID: "9ebc429e-a12b-4547-9d98-d2459194642a"). InnerVolumeSpecName "kube-api-access-7jsdw". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 08:48:26 crc kubenswrapper[4612]: I0227 08:48:26.547336 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7jsdw\" (UniqueName: \"kubernetes.io/projected/9ebc429e-a12b-4547-9d98-d2459194642a-kube-api-access-7jsdw\") on node \"crc\" DevicePath \"\"" Feb 27 08:48:26 crc kubenswrapper[4612]: I0227 08:48:26.567818 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9ebc429e-a12b-4547-9d98-d2459194642a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9ebc429e-a12b-4547-9d98-d2459194642a" (UID: "9ebc429e-a12b-4547-9d98-d2459194642a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 08:48:26 crc kubenswrapper[4612]: I0227 08:48:26.649164 4612 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9ebc429e-a12b-4547-9d98-d2459194642a-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 27 08:48:26 crc kubenswrapper[4612]: I0227 08:48:26.679195 4612 generic.go:334] "Generic (PLEG): container finished" podID="9ebc429e-a12b-4547-9d98-d2459194642a" containerID="67b9f2e4d7da062a5ac3239214252d579a1efc8485ab68dd028a36af63bf578e" exitCode=0 Feb 27 08:48:26 crc kubenswrapper[4612]: I0227 08:48:26.679256 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-k67wt" event={"ID":"9ebc429e-a12b-4547-9d98-d2459194642a","Type":"ContainerDied","Data":"67b9f2e4d7da062a5ac3239214252d579a1efc8485ab68dd028a36af63bf578e"} Feb 27 08:48:26 crc kubenswrapper[4612]: I0227 08:48:26.679305 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-k67wt" event={"ID":"9ebc429e-a12b-4547-9d98-d2459194642a","Type":"ContainerDied","Data":"ae3690e3e858479fc21c2eefba5279f7a92f3d70de58ed547cd15c8ee3bdaef2"} Feb 27 08:48:26 crc kubenswrapper[4612]: I0227 08:48:26.679348 4612 scope.go:117] "RemoveContainer" containerID="67b9f2e4d7da062a5ac3239214252d579a1efc8485ab68dd028a36af63bf578e" Feb 27 08:48:26 crc kubenswrapper[4612]: I0227 08:48:26.679558 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-k67wt" Feb 27 08:48:26 crc kubenswrapper[4612]: I0227 08:48:26.734957 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-k67wt"] Feb 27 08:48:26 crc kubenswrapper[4612]: I0227 08:48:26.738650 4612 scope.go:117] "RemoveContainer" containerID="9c616c2fd68f812eea535758f2514701d29f2f3daf88442d154c1d1f86f9d405" Feb 27 08:48:26 crc kubenswrapper[4612]: I0227 08:48:26.750878 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-k67wt"] Feb 27 08:48:26 crc kubenswrapper[4612]: I0227 08:48:26.766597 4612 scope.go:117] "RemoveContainer" containerID="ba4c7ee554e95f9f26ca5d030ff49d6cb147c68c8d9aac0553941f5b8c9b24ec" Feb 27 08:48:26 crc kubenswrapper[4612]: I0227 08:48:26.851776 4612 scope.go:117] "RemoveContainer" containerID="67b9f2e4d7da062a5ac3239214252d579a1efc8485ab68dd028a36af63bf578e" Feb 27 08:48:26 crc kubenswrapper[4612]: E0227 08:48:26.852222 4612 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"67b9f2e4d7da062a5ac3239214252d579a1efc8485ab68dd028a36af63bf578e\": container with ID starting with 67b9f2e4d7da062a5ac3239214252d579a1efc8485ab68dd028a36af63bf578e not found: ID does not exist" containerID="67b9f2e4d7da062a5ac3239214252d579a1efc8485ab68dd028a36af63bf578e" Feb 27 08:48:26 crc kubenswrapper[4612]: I0227 08:48:26.852253 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"67b9f2e4d7da062a5ac3239214252d579a1efc8485ab68dd028a36af63bf578e"} err="failed to get container status \"67b9f2e4d7da062a5ac3239214252d579a1efc8485ab68dd028a36af63bf578e\": rpc error: code = NotFound desc = could not find container \"67b9f2e4d7da062a5ac3239214252d579a1efc8485ab68dd028a36af63bf578e\": container with ID starting with 67b9f2e4d7da062a5ac3239214252d579a1efc8485ab68dd028a36af63bf578e not found: ID does not exist" Feb 27 08:48:26 crc kubenswrapper[4612]: I0227 08:48:26.852274 4612 scope.go:117] "RemoveContainer" containerID="9c616c2fd68f812eea535758f2514701d29f2f3daf88442d154c1d1f86f9d405" Feb 27 08:48:26 crc kubenswrapper[4612]: E0227 08:48:26.852613 4612 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9c616c2fd68f812eea535758f2514701d29f2f3daf88442d154c1d1f86f9d405\": container with ID starting with 9c616c2fd68f812eea535758f2514701d29f2f3daf88442d154c1d1f86f9d405 not found: ID does not exist" containerID="9c616c2fd68f812eea535758f2514701d29f2f3daf88442d154c1d1f86f9d405" Feb 27 08:48:26 crc kubenswrapper[4612]: I0227 08:48:26.852634 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9c616c2fd68f812eea535758f2514701d29f2f3daf88442d154c1d1f86f9d405"} err="failed to get container status \"9c616c2fd68f812eea535758f2514701d29f2f3daf88442d154c1d1f86f9d405\": rpc error: code = NotFound desc = could not find container \"9c616c2fd68f812eea535758f2514701d29f2f3daf88442d154c1d1f86f9d405\": container with ID starting with 9c616c2fd68f812eea535758f2514701d29f2f3daf88442d154c1d1f86f9d405 not found: ID does not exist" Feb 27 08:48:26 crc kubenswrapper[4612]: I0227 08:48:26.852648 4612 scope.go:117] "RemoveContainer" containerID="ba4c7ee554e95f9f26ca5d030ff49d6cb147c68c8d9aac0553941f5b8c9b24ec" Feb 27 08:48:26 crc kubenswrapper[4612]: E0227 08:48:26.853173 4612 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ba4c7ee554e95f9f26ca5d030ff49d6cb147c68c8d9aac0553941f5b8c9b24ec\": container with ID starting with ba4c7ee554e95f9f26ca5d030ff49d6cb147c68c8d9aac0553941f5b8c9b24ec not found: ID does not exist" containerID="ba4c7ee554e95f9f26ca5d030ff49d6cb147c68c8d9aac0553941f5b8c9b24ec" Feb 27 08:48:26 crc kubenswrapper[4612]: I0227 08:48:26.853190 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ba4c7ee554e95f9f26ca5d030ff49d6cb147c68c8d9aac0553941f5b8c9b24ec"} err="failed to get container status \"ba4c7ee554e95f9f26ca5d030ff49d6cb147c68c8d9aac0553941f5b8c9b24ec\": rpc error: code = NotFound desc = could not find container \"ba4c7ee554e95f9f26ca5d030ff49d6cb147c68c8d9aac0553941f5b8c9b24ec\": container with ID starting with ba4c7ee554e95f9f26ca5d030ff49d6cb147c68c8d9aac0553941f5b8c9b24ec not found: ID does not exist" Feb 27 08:48:26 crc kubenswrapper[4612]: I0227 08:48:26.885965 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9ebc429e-a12b-4547-9d98-d2459194642a" path="/var/lib/kubelet/pods/9ebc429e-a12b-4547-9d98-d2459194642a/volumes" Feb 27 08:48:46 crc kubenswrapper[4612]: I0227 08:48:46.027296 4612 patch_prober.go:28] interesting pod/machine-config-daemon-924vb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 27 08:48:46 crc kubenswrapper[4612]: I0227 08:48:46.027797 4612 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 27 08:49:16 crc kubenswrapper[4612]: I0227 08:49:16.026771 4612 patch_prober.go:28] interesting pod/machine-config-daemon-924vb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 27 08:49:16 crc kubenswrapper[4612]: I0227 08:49:16.027311 4612 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 27 08:49:46 crc kubenswrapper[4612]: I0227 08:49:46.027513 4612 patch_prober.go:28] interesting pod/machine-config-daemon-924vb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 27 08:49:46 crc kubenswrapper[4612]: I0227 08:49:46.028140 4612 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 27 08:49:46 crc kubenswrapper[4612]: I0227 08:49:46.028213 4612 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-924vb" Feb 27 08:49:46 crc kubenswrapper[4612]: I0227 08:49:46.029274 4612 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"2ed59616da422731358981fd9cdfe7932267d24e60d616b16a3481f816f614da"} pod="openshift-machine-config-operator/machine-config-daemon-924vb" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 27 08:49:46 crc kubenswrapper[4612]: I0227 08:49:46.029345 4612 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" containerName="machine-config-daemon" containerID="cri-o://2ed59616da422731358981fd9cdfe7932267d24e60d616b16a3481f816f614da" gracePeriod=600 Feb 27 08:49:46 crc kubenswrapper[4612]: E0227 08:49:46.156103 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-924vb_openshift-machine-config-operator(24599373-7adc-4a1b-8bb4-797bf726f43d)\"" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" Feb 27 08:49:46 crc kubenswrapper[4612]: I0227 08:49:46.386539 4612 generic.go:334] "Generic (PLEG): container finished" podID="24599373-7adc-4a1b-8bb4-797bf726f43d" containerID="2ed59616da422731358981fd9cdfe7932267d24e60d616b16a3481f816f614da" exitCode=0 Feb 27 08:49:46 crc kubenswrapper[4612]: I0227 08:49:46.386593 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-924vb" event={"ID":"24599373-7adc-4a1b-8bb4-797bf726f43d","Type":"ContainerDied","Data":"2ed59616da422731358981fd9cdfe7932267d24e60d616b16a3481f816f614da"} Feb 27 08:49:46 crc kubenswrapper[4612]: I0227 08:49:46.386632 4612 scope.go:117] "RemoveContainer" containerID="bf86b408e02a5e06a79d4d8bbbe514621c6f07302864c707bf02845f4bf9beb7" Feb 27 08:49:46 crc kubenswrapper[4612]: I0227 08:49:46.387270 4612 scope.go:117] "RemoveContainer" containerID="2ed59616da422731358981fd9cdfe7932267d24e60d616b16a3481f816f614da" Feb 27 08:49:46 crc kubenswrapper[4612]: E0227 08:49:46.387602 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-924vb_openshift-machine-config-operator(24599373-7adc-4a1b-8bb4-797bf726f43d)\"" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" Feb 27 08:49:57 crc kubenswrapper[4612]: I0227 08:49:57.852892 4612 scope.go:117] "RemoveContainer" containerID="2ed59616da422731358981fd9cdfe7932267d24e60d616b16a3481f816f614da" Feb 27 08:49:57 crc kubenswrapper[4612]: E0227 08:49:57.853673 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-924vb_openshift-machine-config-operator(24599373-7adc-4a1b-8bb4-797bf726f43d)\"" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" Feb 27 08:50:00 crc kubenswrapper[4612]: I0227 08:50:00.144879 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29536370-wp582"] Feb 27 08:50:00 crc kubenswrapper[4612]: E0227 08:50:00.146626 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dfdd99d3-8279-46f6-a853-418a80e00f19" containerName="oc" Feb 27 08:50:00 crc kubenswrapper[4612]: I0227 08:50:00.146744 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="dfdd99d3-8279-46f6-a853-418a80e00f19" containerName="oc" Feb 27 08:50:00 crc kubenswrapper[4612]: E0227 08:50:00.146832 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9ebc429e-a12b-4547-9d98-d2459194642a" containerName="registry-server" Feb 27 08:50:00 crc kubenswrapper[4612]: I0227 08:50:00.146907 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="9ebc429e-a12b-4547-9d98-d2459194642a" containerName="registry-server" Feb 27 08:50:00 crc kubenswrapper[4612]: E0227 08:50:00.146964 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9ebc429e-a12b-4547-9d98-d2459194642a" containerName="extract-content" Feb 27 08:50:00 crc kubenswrapper[4612]: I0227 08:50:00.147024 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="9ebc429e-a12b-4547-9d98-d2459194642a" containerName="extract-content" Feb 27 08:50:00 crc kubenswrapper[4612]: E0227 08:50:00.147176 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9ebc429e-a12b-4547-9d98-d2459194642a" containerName="extract-utilities" Feb 27 08:50:00 crc kubenswrapper[4612]: I0227 08:50:00.147235 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="9ebc429e-a12b-4547-9d98-d2459194642a" containerName="extract-utilities" Feb 27 08:50:00 crc kubenswrapper[4612]: I0227 08:50:00.147467 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="9ebc429e-a12b-4547-9d98-d2459194642a" containerName="registry-server" Feb 27 08:50:00 crc kubenswrapper[4612]: I0227 08:50:00.147536 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="dfdd99d3-8279-46f6-a853-418a80e00f19" containerName="oc" Feb 27 08:50:00 crc kubenswrapper[4612]: I0227 08:50:00.148306 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536370-wp582" Feb 27 08:50:00 crc kubenswrapper[4612]: I0227 08:50:00.150831 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 27 08:50:00 crc kubenswrapper[4612]: I0227 08:50:00.151266 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-5zwfx" Feb 27 08:50:00 crc kubenswrapper[4612]: I0227 08:50:00.151272 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 27 08:50:00 crc kubenswrapper[4612]: I0227 08:50:00.161641 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536370-wp582"] Feb 27 08:50:00 crc kubenswrapper[4612]: I0227 08:50:00.221243 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rfqrf\" (UniqueName: \"kubernetes.io/projected/352f7228-6eac-4b53-a13f-56a291327036-kube-api-access-rfqrf\") pod \"auto-csr-approver-29536370-wp582\" (UID: \"352f7228-6eac-4b53-a13f-56a291327036\") " pod="openshift-infra/auto-csr-approver-29536370-wp582" Feb 27 08:50:00 crc kubenswrapper[4612]: I0227 08:50:00.322961 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rfqrf\" (UniqueName: \"kubernetes.io/projected/352f7228-6eac-4b53-a13f-56a291327036-kube-api-access-rfqrf\") pod \"auto-csr-approver-29536370-wp582\" (UID: \"352f7228-6eac-4b53-a13f-56a291327036\") " pod="openshift-infra/auto-csr-approver-29536370-wp582" Feb 27 08:50:00 crc kubenswrapper[4612]: I0227 08:50:00.342826 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rfqrf\" (UniqueName: \"kubernetes.io/projected/352f7228-6eac-4b53-a13f-56a291327036-kube-api-access-rfqrf\") pod \"auto-csr-approver-29536370-wp582\" (UID: \"352f7228-6eac-4b53-a13f-56a291327036\") " pod="openshift-infra/auto-csr-approver-29536370-wp582" Feb 27 08:50:00 crc kubenswrapper[4612]: I0227 08:50:00.468741 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536370-wp582" Feb 27 08:50:00 crc kubenswrapper[4612]: I0227 08:50:00.983434 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536370-wp582"] Feb 27 08:50:01 crc kubenswrapper[4612]: I0227 08:50:01.532416 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536370-wp582" event={"ID":"352f7228-6eac-4b53-a13f-56a291327036","Type":"ContainerStarted","Data":"13cb3641995c1eaca38dc4be2339b34eacad749adf41873ab2328acb45221777"} Feb 27 08:50:02 crc kubenswrapper[4612]: I0227 08:50:02.543053 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536370-wp582" event={"ID":"352f7228-6eac-4b53-a13f-56a291327036","Type":"ContainerStarted","Data":"692e3fe94710ca549bee22efb82dd732e5aacee245610422e69ed52e893e04a0"} Feb 27 08:50:02 crc kubenswrapper[4612]: I0227 08:50:02.562004 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29536370-wp582" podStartSLOduration=1.5908471309999999 podStartE2EDuration="2.561987184s" podCreationTimestamp="2026-02-27 08:50:00 +0000 UTC" firstStartedPulling="2026-02-27 08:50:00.98873443 +0000 UTC m=+3658.842664428" lastFinishedPulling="2026-02-27 08:50:01.959874463 +0000 UTC m=+3659.813804481" observedRunningTime="2026-02-27 08:50:02.557135356 +0000 UTC m=+3660.411065344" watchObservedRunningTime="2026-02-27 08:50:02.561987184 +0000 UTC m=+3660.415917182" Feb 27 08:50:03 crc kubenswrapper[4612]: I0227 08:50:03.553184 4612 generic.go:334] "Generic (PLEG): container finished" podID="352f7228-6eac-4b53-a13f-56a291327036" containerID="692e3fe94710ca549bee22efb82dd732e5aacee245610422e69ed52e893e04a0" exitCode=0 Feb 27 08:50:03 crc kubenswrapper[4612]: I0227 08:50:03.553444 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536370-wp582" event={"ID":"352f7228-6eac-4b53-a13f-56a291327036","Type":"ContainerDied","Data":"692e3fe94710ca549bee22efb82dd732e5aacee245610422e69ed52e893e04a0"} Feb 27 08:50:05 crc kubenswrapper[4612]: I0227 08:50:05.188909 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536370-wp582" Feb 27 08:50:05 crc kubenswrapper[4612]: I0227 08:50:05.231544 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rfqrf\" (UniqueName: \"kubernetes.io/projected/352f7228-6eac-4b53-a13f-56a291327036-kube-api-access-rfqrf\") pod \"352f7228-6eac-4b53-a13f-56a291327036\" (UID: \"352f7228-6eac-4b53-a13f-56a291327036\") " Feb 27 08:50:05 crc kubenswrapper[4612]: I0227 08:50:05.244716 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/352f7228-6eac-4b53-a13f-56a291327036-kube-api-access-rfqrf" (OuterVolumeSpecName: "kube-api-access-rfqrf") pod "352f7228-6eac-4b53-a13f-56a291327036" (UID: "352f7228-6eac-4b53-a13f-56a291327036"). InnerVolumeSpecName "kube-api-access-rfqrf". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 08:50:05 crc kubenswrapper[4612]: I0227 08:50:05.333628 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rfqrf\" (UniqueName: \"kubernetes.io/projected/352f7228-6eac-4b53-a13f-56a291327036-kube-api-access-rfqrf\") on node \"crc\" DevicePath \"\"" Feb 27 08:50:05 crc kubenswrapper[4612]: I0227 08:50:05.569523 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536370-wp582" event={"ID":"352f7228-6eac-4b53-a13f-56a291327036","Type":"ContainerDied","Data":"13cb3641995c1eaca38dc4be2339b34eacad749adf41873ab2328acb45221777"} Feb 27 08:50:05 crc kubenswrapper[4612]: I0227 08:50:05.569574 4612 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="13cb3641995c1eaca38dc4be2339b34eacad749adf41873ab2328acb45221777" Feb 27 08:50:05 crc kubenswrapper[4612]: I0227 08:50:05.569624 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536370-wp582" Feb 27 08:50:05 crc kubenswrapper[4612]: I0227 08:50:05.667457 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29536364-4ztkb"] Feb 27 08:50:05 crc kubenswrapper[4612]: I0227 08:50:05.679519 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29536364-4ztkb"] Feb 27 08:50:06 crc kubenswrapper[4612]: I0227 08:50:06.867596 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="14ea3683-b0e3-4aaf-bc0f-e6e3e15bbe41" path="/var/lib/kubelet/pods/14ea3683-b0e3-4aaf-bc0f-e6e3e15bbe41/volumes" Feb 27 08:50:11 crc kubenswrapper[4612]: I0227 08:50:11.854125 4612 scope.go:117] "RemoveContainer" containerID="2ed59616da422731358981fd9cdfe7932267d24e60d616b16a3481f816f614da" Feb 27 08:50:11 crc kubenswrapper[4612]: E0227 08:50:11.855342 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-924vb_openshift-machine-config-operator(24599373-7adc-4a1b-8bb4-797bf726f43d)\"" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" Feb 27 08:50:19 crc kubenswrapper[4612]: I0227 08:50:19.194442 4612 scope.go:117] "RemoveContainer" containerID="c410fb9e5df505d0bd2d98d4d9cf5ec4aa03d9c3b71166796b145c010b5a121c" Feb 27 08:50:26 crc kubenswrapper[4612]: I0227 08:50:26.852978 4612 scope.go:117] "RemoveContainer" containerID="2ed59616da422731358981fd9cdfe7932267d24e60d616b16a3481f816f614da" Feb 27 08:50:26 crc kubenswrapper[4612]: E0227 08:50:26.853642 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-924vb_openshift-machine-config-operator(24599373-7adc-4a1b-8bb4-797bf726f43d)\"" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" Feb 27 08:50:41 crc kubenswrapper[4612]: I0227 08:50:41.853891 4612 scope.go:117] "RemoveContainer" containerID="2ed59616da422731358981fd9cdfe7932267d24e60d616b16a3481f816f614da" Feb 27 08:50:41 crc kubenswrapper[4612]: E0227 08:50:41.856050 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-924vb_openshift-machine-config-operator(24599373-7adc-4a1b-8bb4-797bf726f43d)\"" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" Feb 27 08:50:53 crc kubenswrapper[4612]: I0227 08:50:53.853028 4612 scope.go:117] "RemoveContainer" containerID="2ed59616da422731358981fd9cdfe7932267d24e60d616b16a3481f816f614da" Feb 27 08:50:53 crc kubenswrapper[4612]: E0227 08:50:53.853758 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-924vb_openshift-machine-config-operator(24599373-7adc-4a1b-8bb4-797bf726f43d)\"" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" Feb 27 08:51:06 crc kubenswrapper[4612]: I0227 08:51:06.853578 4612 scope.go:117] "RemoveContainer" containerID="2ed59616da422731358981fd9cdfe7932267d24e60d616b16a3481f816f614da" Feb 27 08:51:06 crc kubenswrapper[4612]: E0227 08:51:06.854357 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-924vb_openshift-machine-config-operator(24599373-7adc-4a1b-8bb4-797bf726f43d)\"" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" Feb 27 08:51:17 crc kubenswrapper[4612]: I0227 08:51:17.853434 4612 scope.go:117] "RemoveContainer" containerID="2ed59616da422731358981fd9cdfe7932267d24e60d616b16a3481f816f614da" Feb 27 08:51:17 crc kubenswrapper[4612]: E0227 08:51:17.854183 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-924vb_openshift-machine-config-operator(24599373-7adc-4a1b-8bb4-797bf726f43d)\"" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" Feb 27 08:51:28 crc kubenswrapper[4612]: I0227 08:51:28.097312 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-n99ts"] Feb 27 08:51:28 crc kubenswrapper[4612]: E0227 08:51:28.098059 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="352f7228-6eac-4b53-a13f-56a291327036" containerName="oc" Feb 27 08:51:28 crc kubenswrapper[4612]: I0227 08:51:28.098071 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="352f7228-6eac-4b53-a13f-56a291327036" containerName="oc" Feb 27 08:51:28 crc kubenswrapper[4612]: I0227 08:51:28.098252 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="352f7228-6eac-4b53-a13f-56a291327036" containerName="oc" Feb 27 08:51:28 crc kubenswrapper[4612]: I0227 08:51:28.099492 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-n99ts" Feb 27 08:51:28 crc kubenswrapper[4612]: I0227 08:51:28.120316 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-n99ts"] Feb 27 08:51:28 crc kubenswrapper[4612]: I0227 08:51:28.185989 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/844209be-3142-4654-93f1-821fa29ec7d4-utilities\") pod \"community-operators-n99ts\" (UID: \"844209be-3142-4654-93f1-821fa29ec7d4\") " pod="openshift-marketplace/community-operators-n99ts" Feb 27 08:51:28 crc kubenswrapper[4612]: I0227 08:51:28.186129 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/844209be-3142-4654-93f1-821fa29ec7d4-catalog-content\") pod \"community-operators-n99ts\" (UID: \"844209be-3142-4654-93f1-821fa29ec7d4\") " pod="openshift-marketplace/community-operators-n99ts" Feb 27 08:51:28 crc kubenswrapper[4612]: I0227 08:51:28.186191 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5mljf\" (UniqueName: \"kubernetes.io/projected/844209be-3142-4654-93f1-821fa29ec7d4-kube-api-access-5mljf\") pod \"community-operators-n99ts\" (UID: \"844209be-3142-4654-93f1-821fa29ec7d4\") " pod="openshift-marketplace/community-operators-n99ts" Feb 27 08:51:28 crc kubenswrapper[4612]: I0227 08:51:28.287638 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5mljf\" (UniqueName: \"kubernetes.io/projected/844209be-3142-4654-93f1-821fa29ec7d4-kube-api-access-5mljf\") pod \"community-operators-n99ts\" (UID: \"844209be-3142-4654-93f1-821fa29ec7d4\") " pod="openshift-marketplace/community-operators-n99ts" Feb 27 08:51:28 crc kubenswrapper[4612]: I0227 08:51:28.287971 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/844209be-3142-4654-93f1-821fa29ec7d4-utilities\") pod \"community-operators-n99ts\" (UID: \"844209be-3142-4654-93f1-821fa29ec7d4\") " pod="openshift-marketplace/community-operators-n99ts" Feb 27 08:51:28 crc kubenswrapper[4612]: I0227 08:51:28.288062 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/844209be-3142-4654-93f1-821fa29ec7d4-catalog-content\") pod \"community-operators-n99ts\" (UID: \"844209be-3142-4654-93f1-821fa29ec7d4\") " pod="openshift-marketplace/community-operators-n99ts" Feb 27 08:51:28 crc kubenswrapper[4612]: I0227 08:51:28.288499 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/844209be-3142-4654-93f1-821fa29ec7d4-catalog-content\") pod \"community-operators-n99ts\" (UID: \"844209be-3142-4654-93f1-821fa29ec7d4\") " pod="openshift-marketplace/community-operators-n99ts" Feb 27 08:51:28 crc kubenswrapper[4612]: I0227 08:51:28.288538 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/844209be-3142-4654-93f1-821fa29ec7d4-utilities\") pod \"community-operators-n99ts\" (UID: \"844209be-3142-4654-93f1-821fa29ec7d4\") " pod="openshift-marketplace/community-operators-n99ts" Feb 27 08:51:28 crc kubenswrapper[4612]: I0227 08:51:28.308508 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5mljf\" (UniqueName: \"kubernetes.io/projected/844209be-3142-4654-93f1-821fa29ec7d4-kube-api-access-5mljf\") pod \"community-operators-n99ts\" (UID: \"844209be-3142-4654-93f1-821fa29ec7d4\") " pod="openshift-marketplace/community-operators-n99ts" Feb 27 08:51:28 crc kubenswrapper[4612]: I0227 08:51:28.421541 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-n99ts" Feb 27 08:51:28 crc kubenswrapper[4612]: I0227 08:51:28.990589 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-n99ts"] Feb 27 08:51:29 crc kubenswrapper[4612]: I0227 08:51:29.371043 4612 generic.go:334] "Generic (PLEG): container finished" podID="844209be-3142-4654-93f1-821fa29ec7d4" containerID="0823af2cbada387fbf89c7586213e782674be197f028aae3623ebff4eac2a82e" exitCode=0 Feb 27 08:51:29 crc kubenswrapper[4612]: I0227 08:51:29.372215 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-n99ts" event={"ID":"844209be-3142-4654-93f1-821fa29ec7d4","Type":"ContainerDied","Data":"0823af2cbada387fbf89c7586213e782674be197f028aae3623ebff4eac2a82e"} Feb 27 08:51:29 crc kubenswrapper[4612]: I0227 08:51:29.372315 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-n99ts" event={"ID":"844209be-3142-4654-93f1-821fa29ec7d4","Type":"ContainerStarted","Data":"9d38bd1dbdd56a97297188ac1b7d07e7cbcbe4e0999d4158224d6965dd9e8b56"} Feb 27 08:51:31 crc kubenswrapper[4612]: I0227 08:51:31.393303 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-n99ts" event={"ID":"844209be-3142-4654-93f1-821fa29ec7d4","Type":"ContainerStarted","Data":"741e99a65ede4fceffa11b728d1121f26b576d10f885ee2b8c0681bb3b642a4c"} Feb 27 08:51:31 crc kubenswrapper[4612]: I0227 08:51:31.854847 4612 scope.go:117] "RemoveContainer" containerID="2ed59616da422731358981fd9cdfe7932267d24e60d616b16a3481f816f614da" Feb 27 08:51:31 crc kubenswrapper[4612]: E0227 08:51:31.855098 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-924vb_openshift-machine-config-operator(24599373-7adc-4a1b-8bb4-797bf726f43d)\"" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" Feb 27 08:51:32 crc kubenswrapper[4612]: I0227 08:51:32.421874 4612 generic.go:334] "Generic (PLEG): container finished" podID="844209be-3142-4654-93f1-821fa29ec7d4" containerID="741e99a65ede4fceffa11b728d1121f26b576d10f885ee2b8c0681bb3b642a4c" exitCode=0 Feb 27 08:51:32 crc kubenswrapper[4612]: I0227 08:51:32.422299 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-n99ts" event={"ID":"844209be-3142-4654-93f1-821fa29ec7d4","Type":"ContainerDied","Data":"741e99a65ede4fceffa11b728d1121f26b576d10f885ee2b8c0681bb3b642a4c"} Feb 27 08:51:33 crc kubenswrapper[4612]: I0227 08:51:33.431830 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-n99ts" event={"ID":"844209be-3142-4654-93f1-821fa29ec7d4","Type":"ContainerStarted","Data":"e310cf3548f22571651710d386dff2c5486dee36cc3ca4e7a05a6e54566e844c"} Feb 27 08:51:38 crc kubenswrapper[4612]: I0227 08:51:38.422493 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-n99ts" Feb 27 08:51:38 crc kubenswrapper[4612]: I0227 08:51:38.423420 4612 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-n99ts" Feb 27 08:51:38 crc kubenswrapper[4612]: I0227 08:51:38.486866 4612 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-n99ts" Feb 27 08:51:38 crc kubenswrapper[4612]: I0227 08:51:38.508431 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-n99ts" podStartSLOduration=7.03562084 podStartE2EDuration="10.508406239s" podCreationTimestamp="2026-02-27 08:51:28 +0000 UTC" firstStartedPulling="2026-02-27 08:51:29.373024306 +0000 UTC m=+3747.226954304" lastFinishedPulling="2026-02-27 08:51:32.845809685 +0000 UTC m=+3750.699739703" observedRunningTime="2026-02-27 08:51:33.4540293 +0000 UTC m=+3751.307959298" watchObservedRunningTime="2026-02-27 08:51:38.508406239 +0000 UTC m=+3756.362336237" Feb 27 08:51:38 crc kubenswrapper[4612]: I0227 08:51:38.544683 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-n99ts" Feb 27 08:51:38 crc kubenswrapper[4612]: I0227 08:51:38.729754 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-n99ts"] Feb 27 08:51:40 crc kubenswrapper[4612]: I0227 08:51:40.491959 4612 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-n99ts" podUID="844209be-3142-4654-93f1-821fa29ec7d4" containerName="registry-server" containerID="cri-o://e310cf3548f22571651710d386dff2c5486dee36cc3ca4e7a05a6e54566e844c" gracePeriod=2 Feb 27 08:51:41 crc kubenswrapper[4612]: I0227 08:51:41.023755 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-n99ts" Feb 27 08:51:41 crc kubenswrapper[4612]: I0227 08:51:41.041520 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/844209be-3142-4654-93f1-821fa29ec7d4-catalog-content\") pod \"844209be-3142-4654-93f1-821fa29ec7d4\" (UID: \"844209be-3142-4654-93f1-821fa29ec7d4\") " Feb 27 08:51:41 crc kubenswrapper[4612]: I0227 08:51:41.041578 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5mljf\" (UniqueName: \"kubernetes.io/projected/844209be-3142-4654-93f1-821fa29ec7d4-kube-api-access-5mljf\") pod \"844209be-3142-4654-93f1-821fa29ec7d4\" (UID: \"844209be-3142-4654-93f1-821fa29ec7d4\") " Feb 27 08:51:41 crc kubenswrapper[4612]: I0227 08:51:41.041625 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/844209be-3142-4654-93f1-821fa29ec7d4-utilities\") pod \"844209be-3142-4654-93f1-821fa29ec7d4\" (UID: \"844209be-3142-4654-93f1-821fa29ec7d4\") " Feb 27 08:51:41 crc kubenswrapper[4612]: I0227 08:51:41.042621 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/844209be-3142-4654-93f1-821fa29ec7d4-utilities" (OuterVolumeSpecName: "utilities") pod "844209be-3142-4654-93f1-821fa29ec7d4" (UID: "844209be-3142-4654-93f1-821fa29ec7d4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 08:51:41 crc kubenswrapper[4612]: I0227 08:51:41.051209 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/844209be-3142-4654-93f1-821fa29ec7d4-kube-api-access-5mljf" (OuterVolumeSpecName: "kube-api-access-5mljf") pod "844209be-3142-4654-93f1-821fa29ec7d4" (UID: "844209be-3142-4654-93f1-821fa29ec7d4"). InnerVolumeSpecName "kube-api-access-5mljf". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 08:51:41 crc kubenswrapper[4612]: I0227 08:51:41.110446 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/844209be-3142-4654-93f1-821fa29ec7d4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "844209be-3142-4654-93f1-821fa29ec7d4" (UID: "844209be-3142-4654-93f1-821fa29ec7d4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 08:51:41 crc kubenswrapper[4612]: I0227 08:51:41.145313 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5mljf\" (UniqueName: \"kubernetes.io/projected/844209be-3142-4654-93f1-821fa29ec7d4-kube-api-access-5mljf\") on node \"crc\" DevicePath \"\"" Feb 27 08:51:41 crc kubenswrapper[4612]: I0227 08:51:41.145364 4612 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/844209be-3142-4654-93f1-821fa29ec7d4-utilities\") on node \"crc\" DevicePath \"\"" Feb 27 08:51:41 crc kubenswrapper[4612]: I0227 08:51:41.145380 4612 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/844209be-3142-4654-93f1-821fa29ec7d4-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 27 08:51:41 crc kubenswrapper[4612]: I0227 08:51:41.500977 4612 generic.go:334] "Generic (PLEG): container finished" podID="844209be-3142-4654-93f1-821fa29ec7d4" containerID="e310cf3548f22571651710d386dff2c5486dee36cc3ca4e7a05a6e54566e844c" exitCode=0 Feb 27 08:51:41 crc kubenswrapper[4612]: I0227 08:51:41.501034 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-n99ts" event={"ID":"844209be-3142-4654-93f1-821fa29ec7d4","Type":"ContainerDied","Data":"e310cf3548f22571651710d386dff2c5486dee36cc3ca4e7a05a6e54566e844c"} Feb 27 08:51:41 crc kubenswrapper[4612]: I0227 08:51:41.501047 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-n99ts" Feb 27 08:51:41 crc kubenswrapper[4612]: I0227 08:51:41.501088 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-n99ts" event={"ID":"844209be-3142-4654-93f1-821fa29ec7d4","Type":"ContainerDied","Data":"9d38bd1dbdd56a97297188ac1b7d07e7cbcbe4e0999d4158224d6965dd9e8b56"} Feb 27 08:51:41 crc kubenswrapper[4612]: I0227 08:51:41.501118 4612 scope.go:117] "RemoveContainer" containerID="e310cf3548f22571651710d386dff2c5486dee36cc3ca4e7a05a6e54566e844c" Feb 27 08:51:41 crc kubenswrapper[4612]: I0227 08:51:41.527841 4612 scope.go:117] "RemoveContainer" containerID="741e99a65ede4fceffa11b728d1121f26b576d10f885ee2b8c0681bb3b642a4c" Feb 27 08:51:41 crc kubenswrapper[4612]: I0227 08:51:41.567447 4612 scope.go:117] "RemoveContainer" containerID="0823af2cbada387fbf89c7586213e782674be197f028aae3623ebff4eac2a82e" Feb 27 08:51:41 crc kubenswrapper[4612]: I0227 08:51:41.625100 4612 scope.go:117] "RemoveContainer" containerID="e310cf3548f22571651710d386dff2c5486dee36cc3ca4e7a05a6e54566e844c" Feb 27 08:51:41 crc kubenswrapper[4612]: E0227 08:51:41.626920 4612 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e310cf3548f22571651710d386dff2c5486dee36cc3ca4e7a05a6e54566e844c\": container with ID starting with e310cf3548f22571651710d386dff2c5486dee36cc3ca4e7a05a6e54566e844c not found: ID does not exist" containerID="e310cf3548f22571651710d386dff2c5486dee36cc3ca4e7a05a6e54566e844c" Feb 27 08:51:41 crc kubenswrapper[4612]: I0227 08:51:41.626963 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e310cf3548f22571651710d386dff2c5486dee36cc3ca4e7a05a6e54566e844c"} err="failed to get container status \"e310cf3548f22571651710d386dff2c5486dee36cc3ca4e7a05a6e54566e844c\": rpc error: code = NotFound desc = could not find container \"e310cf3548f22571651710d386dff2c5486dee36cc3ca4e7a05a6e54566e844c\": container with ID starting with e310cf3548f22571651710d386dff2c5486dee36cc3ca4e7a05a6e54566e844c not found: ID does not exist" Feb 27 08:51:41 crc kubenswrapper[4612]: I0227 08:51:41.626994 4612 scope.go:117] "RemoveContainer" containerID="741e99a65ede4fceffa11b728d1121f26b576d10f885ee2b8c0681bb3b642a4c" Feb 27 08:51:41 crc kubenswrapper[4612]: E0227 08:51:41.628450 4612 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"741e99a65ede4fceffa11b728d1121f26b576d10f885ee2b8c0681bb3b642a4c\": container with ID starting with 741e99a65ede4fceffa11b728d1121f26b576d10f885ee2b8c0681bb3b642a4c not found: ID does not exist" containerID="741e99a65ede4fceffa11b728d1121f26b576d10f885ee2b8c0681bb3b642a4c" Feb 27 08:51:41 crc kubenswrapper[4612]: I0227 08:51:41.628493 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"741e99a65ede4fceffa11b728d1121f26b576d10f885ee2b8c0681bb3b642a4c"} err="failed to get container status \"741e99a65ede4fceffa11b728d1121f26b576d10f885ee2b8c0681bb3b642a4c\": rpc error: code = NotFound desc = could not find container \"741e99a65ede4fceffa11b728d1121f26b576d10f885ee2b8c0681bb3b642a4c\": container with ID starting with 741e99a65ede4fceffa11b728d1121f26b576d10f885ee2b8c0681bb3b642a4c not found: ID does not exist" Feb 27 08:51:41 crc kubenswrapper[4612]: I0227 08:51:41.628519 4612 scope.go:117] "RemoveContainer" containerID="0823af2cbada387fbf89c7586213e782674be197f028aae3623ebff4eac2a82e" Feb 27 08:51:41 crc kubenswrapper[4612]: E0227 08:51:41.631219 4612 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0823af2cbada387fbf89c7586213e782674be197f028aae3623ebff4eac2a82e\": container with ID starting with 0823af2cbada387fbf89c7586213e782674be197f028aae3623ebff4eac2a82e not found: ID does not exist" containerID="0823af2cbada387fbf89c7586213e782674be197f028aae3623ebff4eac2a82e" Feb 27 08:51:41 crc kubenswrapper[4612]: I0227 08:51:41.631268 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0823af2cbada387fbf89c7586213e782674be197f028aae3623ebff4eac2a82e"} err="failed to get container status \"0823af2cbada387fbf89c7586213e782674be197f028aae3623ebff4eac2a82e\": rpc error: code = NotFound desc = could not find container \"0823af2cbada387fbf89c7586213e782674be197f028aae3623ebff4eac2a82e\": container with ID starting with 0823af2cbada387fbf89c7586213e782674be197f028aae3623ebff4eac2a82e not found: ID does not exist" Feb 27 08:51:41 crc kubenswrapper[4612]: I0227 08:51:41.634636 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-n99ts"] Feb 27 08:51:41 crc kubenswrapper[4612]: I0227 08:51:41.643680 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-n99ts"] Feb 27 08:51:42 crc kubenswrapper[4612]: I0227 08:51:42.862926 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="844209be-3142-4654-93f1-821fa29ec7d4" path="/var/lib/kubelet/pods/844209be-3142-4654-93f1-821fa29ec7d4/volumes" Feb 27 08:51:45 crc kubenswrapper[4612]: I0227 08:51:45.853476 4612 scope.go:117] "RemoveContainer" containerID="2ed59616da422731358981fd9cdfe7932267d24e60d616b16a3481f816f614da" Feb 27 08:51:45 crc kubenswrapper[4612]: E0227 08:51:45.854384 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-924vb_openshift-machine-config-operator(24599373-7adc-4a1b-8bb4-797bf726f43d)\"" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" Feb 27 08:52:00 crc kubenswrapper[4612]: I0227 08:52:00.187320 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29536372-dhs6j"] Feb 27 08:52:00 crc kubenswrapper[4612]: E0227 08:52:00.188685 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="844209be-3142-4654-93f1-821fa29ec7d4" containerName="registry-server" Feb 27 08:52:00 crc kubenswrapper[4612]: I0227 08:52:00.188721 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="844209be-3142-4654-93f1-821fa29ec7d4" containerName="registry-server" Feb 27 08:52:00 crc kubenswrapper[4612]: E0227 08:52:00.188745 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="844209be-3142-4654-93f1-821fa29ec7d4" containerName="extract-utilities" Feb 27 08:52:00 crc kubenswrapper[4612]: I0227 08:52:00.188753 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="844209be-3142-4654-93f1-821fa29ec7d4" containerName="extract-utilities" Feb 27 08:52:00 crc kubenswrapper[4612]: E0227 08:52:00.188786 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="844209be-3142-4654-93f1-821fa29ec7d4" containerName="extract-content" Feb 27 08:52:00 crc kubenswrapper[4612]: I0227 08:52:00.188793 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="844209be-3142-4654-93f1-821fa29ec7d4" containerName="extract-content" Feb 27 08:52:00 crc kubenswrapper[4612]: I0227 08:52:00.189061 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="844209be-3142-4654-93f1-821fa29ec7d4" containerName="registry-server" Feb 27 08:52:00 crc kubenswrapper[4612]: I0227 08:52:00.189971 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536372-dhs6j" Feb 27 08:52:00 crc kubenswrapper[4612]: I0227 08:52:00.192342 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 27 08:52:00 crc kubenswrapper[4612]: I0227 08:52:00.192505 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-5zwfx" Feb 27 08:52:00 crc kubenswrapper[4612]: I0227 08:52:00.192590 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 27 08:52:00 crc kubenswrapper[4612]: I0227 08:52:00.215813 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536372-dhs6j"] Feb 27 08:52:00 crc kubenswrapper[4612]: I0227 08:52:00.303899 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m8dlx\" (UniqueName: \"kubernetes.io/projected/e64df7cb-2f18-4fb7-b66d-86c993b1c8b2-kube-api-access-m8dlx\") pod \"auto-csr-approver-29536372-dhs6j\" (UID: \"e64df7cb-2f18-4fb7-b66d-86c993b1c8b2\") " pod="openshift-infra/auto-csr-approver-29536372-dhs6j" Feb 27 08:52:00 crc kubenswrapper[4612]: I0227 08:52:00.407026 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m8dlx\" (UniqueName: \"kubernetes.io/projected/e64df7cb-2f18-4fb7-b66d-86c993b1c8b2-kube-api-access-m8dlx\") pod \"auto-csr-approver-29536372-dhs6j\" (UID: \"e64df7cb-2f18-4fb7-b66d-86c993b1c8b2\") " pod="openshift-infra/auto-csr-approver-29536372-dhs6j" Feb 27 08:52:00 crc kubenswrapper[4612]: I0227 08:52:00.439047 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m8dlx\" (UniqueName: \"kubernetes.io/projected/e64df7cb-2f18-4fb7-b66d-86c993b1c8b2-kube-api-access-m8dlx\") pod \"auto-csr-approver-29536372-dhs6j\" (UID: \"e64df7cb-2f18-4fb7-b66d-86c993b1c8b2\") " pod="openshift-infra/auto-csr-approver-29536372-dhs6j" Feb 27 08:52:00 crc kubenswrapper[4612]: I0227 08:52:00.509000 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536372-dhs6j" Feb 27 08:52:00 crc kubenswrapper[4612]: I0227 08:52:00.853830 4612 scope.go:117] "RemoveContainer" containerID="2ed59616da422731358981fd9cdfe7932267d24e60d616b16a3481f816f614da" Feb 27 08:52:00 crc kubenswrapper[4612]: E0227 08:52:00.854319 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-924vb_openshift-machine-config-operator(24599373-7adc-4a1b-8bb4-797bf726f43d)\"" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" Feb 27 08:52:01 crc kubenswrapper[4612]: I0227 08:52:01.062255 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536372-dhs6j"] Feb 27 08:52:01 crc kubenswrapper[4612]: W0227 08:52:01.087833 4612 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode64df7cb_2f18_4fb7_b66d_86c993b1c8b2.slice/crio-f6cac59b65da200ce77b961d99909c8204c0254c14e12468e94b6460cc8ef208 WatchSource:0}: Error finding container f6cac59b65da200ce77b961d99909c8204c0254c14e12468e94b6460cc8ef208: Status 404 returned error can't find the container with id f6cac59b65da200ce77b961d99909c8204c0254c14e12468e94b6460cc8ef208 Feb 27 08:52:01 crc kubenswrapper[4612]: I0227 08:52:01.668870 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536372-dhs6j" event={"ID":"e64df7cb-2f18-4fb7-b66d-86c993b1c8b2","Type":"ContainerStarted","Data":"f6cac59b65da200ce77b961d99909c8204c0254c14e12468e94b6460cc8ef208"} Feb 27 08:52:02 crc kubenswrapper[4612]: I0227 08:52:02.679749 4612 generic.go:334] "Generic (PLEG): container finished" podID="e64df7cb-2f18-4fb7-b66d-86c993b1c8b2" containerID="0735a1f078fef42c96e74e75d94933547db722ce87140ee78527e7e495b48895" exitCode=0 Feb 27 08:52:02 crc kubenswrapper[4612]: I0227 08:52:02.679797 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536372-dhs6j" event={"ID":"e64df7cb-2f18-4fb7-b66d-86c993b1c8b2","Type":"ContainerDied","Data":"0735a1f078fef42c96e74e75d94933547db722ce87140ee78527e7e495b48895"} Feb 27 08:52:04 crc kubenswrapper[4612]: I0227 08:52:04.165036 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536372-dhs6j" Feb 27 08:52:04 crc kubenswrapper[4612]: I0227 08:52:04.281273 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m8dlx\" (UniqueName: \"kubernetes.io/projected/e64df7cb-2f18-4fb7-b66d-86c993b1c8b2-kube-api-access-m8dlx\") pod \"e64df7cb-2f18-4fb7-b66d-86c993b1c8b2\" (UID: \"e64df7cb-2f18-4fb7-b66d-86c993b1c8b2\") " Feb 27 08:52:04 crc kubenswrapper[4612]: I0227 08:52:04.287759 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e64df7cb-2f18-4fb7-b66d-86c993b1c8b2-kube-api-access-m8dlx" (OuterVolumeSpecName: "kube-api-access-m8dlx") pod "e64df7cb-2f18-4fb7-b66d-86c993b1c8b2" (UID: "e64df7cb-2f18-4fb7-b66d-86c993b1c8b2"). InnerVolumeSpecName "kube-api-access-m8dlx". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 08:52:04 crc kubenswrapper[4612]: I0227 08:52:04.383750 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m8dlx\" (UniqueName: \"kubernetes.io/projected/e64df7cb-2f18-4fb7-b66d-86c993b1c8b2-kube-api-access-m8dlx\") on node \"crc\" DevicePath \"\"" Feb 27 08:52:04 crc kubenswrapper[4612]: I0227 08:52:04.702105 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536372-dhs6j" event={"ID":"e64df7cb-2f18-4fb7-b66d-86c993b1c8b2","Type":"ContainerDied","Data":"f6cac59b65da200ce77b961d99909c8204c0254c14e12468e94b6460cc8ef208"} Feb 27 08:52:04 crc kubenswrapper[4612]: I0227 08:52:04.702404 4612 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f6cac59b65da200ce77b961d99909c8204c0254c14e12468e94b6460cc8ef208" Feb 27 08:52:04 crc kubenswrapper[4612]: I0227 08:52:04.702187 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536372-dhs6j" Feb 27 08:52:05 crc kubenswrapper[4612]: I0227 08:52:05.260969 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29536366-xfkw7"] Feb 27 08:52:05 crc kubenswrapper[4612]: I0227 08:52:05.271438 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29536366-xfkw7"] Feb 27 08:52:06 crc kubenswrapper[4612]: I0227 08:52:06.872005 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aa5037cd-a5d3-4edf-81e6-32d67a9d08b7" path="/var/lib/kubelet/pods/aa5037cd-a5d3-4edf-81e6-32d67a9d08b7/volumes" Feb 27 08:52:11 crc kubenswrapper[4612]: I0227 08:52:11.852757 4612 scope.go:117] "RemoveContainer" containerID="2ed59616da422731358981fd9cdfe7932267d24e60d616b16a3481f816f614da" Feb 27 08:52:11 crc kubenswrapper[4612]: E0227 08:52:11.853535 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-924vb_openshift-machine-config-operator(24599373-7adc-4a1b-8bb4-797bf726f43d)\"" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" Feb 27 08:52:19 crc kubenswrapper[4612]: I0227 08:52:19.324659 4612 scope.go:117] "RemoveContainer" containerID="11b472f6096690caa8865dd995bcfff584e2fc3df506306aaa71ec91d363c02b" Feb 27 08:52:22 crc kubenswrapper[4612]: I0227 08:52:22.860954 4612 scope.go:117] "RemoveContainer" containerID="2ed59616da422731358981fd9cdfe7932267d24e60d616b16a3481f816f614da" Feb 27 08:52:22 crc kubenswrapper[4612]: E0227 08:52:22.861791 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-924vb_openshift-machine-config-operator(24599373-7adc-4a1b-8bb4-797bf726f43d)\"" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" Feb 27 08:52:37 crc kubenswrapper[4612]: I0227 08:52:37.853220 4612 scope.go:117] "RemoveContainer" containerID="2ed59616da422731358981fd9cdfe7932267d24e60d616b16a3481f816f614da" Feb 27 08:52:37 crc kubenswrapper[4612]: E0227 08:52:37.854010 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-924vb_openshift-machine-config-operator(24599373-7adc-4a1b-8bb4-797bf726f43d)\"" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" Feb 27 08:52:51 crc kubenswrapper[4612]: I0227 08:52:51.853265 4612 scope.go:117] "RemoveContainer" containerID="2ed59616da422731358981fd9cdfe7932267d24e60d616b16a3481f816f614da" Feb 27 08:52:51 crc kubenswrapper[4612]: E0227 08:52:51.854303 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-924vb_openshift-machine-config-operator(24599373-7adc-4a1b-8bb4-797bf726f43d)\"" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" Feb 27 08:53:04 crc kubenswrapper[4612]: I0227 08:53:04.852649 4612 scope.go:117] "RemoveContainer" containerID="2ed59616da422731358981fd9cdfe7932267d24e60d616b16a3481f816f614da" Feb 27 08:53:04 crc kubenswrapper[4612]: E0227 08:53:04.853537 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-924vb_openshift-machine-config-operator(24599373-7adc-4a1b-8bb4-797bf726f43d)\"" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" Feb 27 08:53:15 crc kubenswrapper[4612]: I0227 08:53:15.854501 4612 scope.go:117] "RemoveContainer" containerID="2ed59616da422731358981fd9cdfe7932267d24e60d616b16a3481f816f614da" Feb 27 08:53:15 crc kubenswrapper[4612]: E0227 08:53:15.855778 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-924vb_openshift-machine-config-operator(24599373-7adc-4a1b-8bb4-797bf726f43d)\"" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" Feb 27 08:53:27 crc kubenswrapper[4612]: I0227 08:53:27.852910 4612 scope.go:117] "RemoveContainer" containerID="2ed59616da422731358981fd9cdfe7932267d24e60d616b16a3481f816f614da" Feb 27 08:53:27 crc kubenswrapper[4612]: E0227 08:53:27.853649 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-924vb_openshift-machine-config-operator(24599373-7adc-4a1b-8bb4-797bf726f43d)\"" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" Feb 27 08:53:41 crc kubenswrapper[4612]: I0227 08:53:41.853658 4612 scope.go:117] "RemoveContainer" containerID="2ed59616da422731358981fd9cdfe7932267d24e60d616b16a3481f816f614da" Feb 27 08:53:41 crc kubenswrapper[4612]: E0227 08:53:41.854775 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-924vb_openshift-machine-config-operator(24599373-7adc-4a1b-8bb4-797bf726f43d)\"" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" Feb 27 08:53:52 crc kubenswrapper[4612]: I0227 08:53:52.862206 4612 scope.go:117] "RemoveContainer" containerID="2ed59616da422731358981fd9cdfe7932267d24e60d616b16a3481f816f614da" Feb 27 08:53:52 crc kubenswrapper[4612]: E0227 08:53:52.862994 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-924vb_openshift-machine-config-operator(24599373-7adc-4a1b-8bb4-797bf726f43d)\"" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" Feb 27 08:54:00 crc kubenswrapper[4612]: I0227 08:54:00.162637 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29536374-dx6vb"] Feb 27 08:54:00 crc kubenswrapper[4612]: E0227 08:54:00.163950 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e64df7cb-2f18-4fb7-b66d-86c993b1c8b2" containerName="oc" Feb 27 08:54:00 crc kubenswrapper[4612]: I0227 08:54:00.163972 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="e64df7cb-2f18-4fb7-b66d-86c993b1c8b2" containerName="oc" Feb 27 08:54:00 crc kubenswrapper[4612]: I0227 08:54:00.164317 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="e64df7cb-2f18-4fb7-b66d-86c993b1c8b2" containerName="oc" Feb 27 08:54:00 crc kubenswrapper[4612]: I0227 08:54:00.165354 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536374-dx6vb" Feb 27 08:54:00 crc kubenswrapper[4612]: I0227 08:54:00.167681 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 27 08:54:00 crc kubenswrapper[4612]: I0227 08:54:00.167904 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 27 08:54:00 crc kubenswrapper[4612]: I0227 08:54:00.169607 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-5zwfx" Feb 27 08:54:00 crc kubenswrapper[4612]: I0227 08:54:00.175105 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536374-dx6vb"] Feb 27 08:54:00 crc kubenswrapper[4612]: I0227 08:54:00.345883 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qxlds\" (UniqueName: \"kubernetes.io/projected/ce53ef57-b5de-4fe6-adde-10bbec8f696d-kube-api-access-qxlds\") pod \"auto-csr-approver-29536374-dx6vb\" (UID: \"ce53ef57-b5de-4fe6-adde-10bbec8f696d\") " pod="openshift-infra/auto-csr-approver-29536374-dx6vb" Feb 27 08:54:00 crc kubenswrapper[4612]: I0227 08:54:00.447890 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qxlds\" (UniqueName: \"kubernetes.io/projected/ce53ef57-b5de-4fe6-adde-10bbec8f696d-kube-api-access-qxlds\") pod \"auto-csr-approver-29536374-dx6vb\" (UID: \"ce53ef57-b5de-4fe6-adde-10bbec8f696d\") " pod="openshift-infra/auto-csr-approver-29536374-dx6vb" Feb 27 08:54:00 crc kubenswrapper[4612]: I0227 08:54:00.475502 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qxlds\" (UniqueName: \"kubernetes.io/projected/ce53ef57-b5de-4fe6-adde-10bbec8f696d-kube-api-access-qxlds\") pod \"auto-csr-approver-29536374-dx6vb\" (UID: \"ce53ef57-b5de-4fe6-adde-10bbec8f696d\") " pod="openshift-infra/auto-csr-approver-29536374-dx6vb" Feb 27 08:54:00 crc kubenswrapper[4612]: I0227 08:54:00.488268 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536374-dx6vb" Feb 27 08:54:00 crc kubenswrapper[4612]: I0227 08:54:00.953880 4612 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 27 08:54:00 crc kubenswrapper[4612]: I0227 08:54:00.953988 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536374-dx6vb"] Feb 27 08:54:01 crc kubenswrapper[4612]: I0227 08:54:01.827298 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536374-dx6vb" event={"ID":"ce53ef57-b5de-4fe6-adde-10bbec8f696d","Type":"ContainerStarted","Data":"eabcd1dc7e90a46be98394e7e6ea4a68ddac7d7a7d16a058c23f340c2b2de5c3"} Feb 27 08:54:02 crc kubenswrapper[4612]: I0227 08:54:02.836018 4612 generic.go:334] "Generic (PLEG): container finished" podID="ce53ef57-b5de-4fe6-adde-10bbec8f696d" containerID="b82d7cb0ab80c5384a935820b9c397496afec22701011ee8fa8e4fa29ab371cf" exitCode=0 Feb 27 08:54:02 crc kubenswrapper[4612]: I0227 08:54:02.836092 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536374-dx6vb" event={"ID":"ce53ef57-b5de-4fe6-adde-10bbec8f696d","Type":"ContainerDied","Data":"b82d7cb0ab80c5384a935820b9c397496afec22701011ee8fa8e4fa29ab371cf"} Feb 27 08:54:04 crc kubenswrapper[4612]: I0227 08:54:04.276700 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536374-dx6vb" Feb 27 08:54:04 crc kubenswrapper[4612]: I0227 08:54:04.448012 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qxlds\" (UniqueName: \"kubernetes.io/projected/ce53ef57-b5de-4fe6-adde-10bbec8f696d-kube-api-access-qxlds\") pod \"ce53ef57-b5de-4fe6-adde-10bbec8f696d\" (UID: \"ce53ef57-b5de-4fe6-adde-10bbec8f696d\") " Feb 27 08:54:04 crc kubenswrapper[4612]: I0227 08:54:04.455002 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ce53ef57-b5de-4fe6-adde-10bbec8f696d-kube-api-access-qxlds" (OuterVolumeSpecName: "kube-api-access-qxlds") pod "ce53ef57-b5de-4fe6-adde-10bbec8f696d" (UID: "ce53ef57-b5de-4fe6-adde-10bbec8f696d"). InnerVolumeSpecName "kube-api-access-qxlds". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 08:54:04 crc kubenswrapper[4612]: I0227 08:54:04.550230 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qxlds\" (UniqueName: \"kubernetes.io/projected/ce53ef57-b5de-4fe6-adde-10bbec8f696d-kube-api-access-qxlds\") on node \"crc\" DevicePath \"\"" Feb 27 08:54:04 crc kubenswrapper[4612]: I0227 08:54:04.866194 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536374-dx6vb" Feb 27 08:54:04 crc kubenswrapper[4612]: I0227 08:54:04.866590 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536374-dx6vb" event={"ID":"ce53ef57-b5de-4fe6-adde-10bbec8f696d","Type":"ContainerDied","Data":"eabcd1dc7e90a46be98394e7e6ea4a68ddac7d7a7d16a058c23f340c2b2de5c3"} Feb 27 08:54:04 crc kubenswrapper[4612]: I0227 08:54:04.866628 4612 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="eabcd1dc7e90a46be98394e7e6ea4a68ddac7d7a7d16a058c23f340c2b2de5c3" Feb 27 08:54:05 crc kubenswrapper[4612]: I0227 08:54:05.374449 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29536368-ftbbd"] Feb 27 08:54:05 crc kubenswrapper[4612]: I0227 08:54:05.382069 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29536368-ftbbd"] Feb 27 08:54:05 crc kubenswrapper[4612]: I0227 08:54:05.852848 4612 scope.go:117] "RemoveContainer" containerID="2ed59616da422731358981fd9cdfe7932267d24e60d616b16a3481f816f614da" Feb 27 08:54:05 crc kubenswrapper[4612]: E0227 08:54:05.853103 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-924vb_openshift-machine-config-operator(24599373-7adc-4a1b-8bb4-797bf726f43d)\"" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" Feb 27 08:54:06 crc kubenswrapper[4612]: I0227 08:54:06.862904 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dfdd99d3-8279-46f6-a853-418a80e00f19" path="/var/lib/kubelet/pods/dfdd99d3-8279-46f6-a853-418a80e00f19/volumes" Feb 27 08:54:18 crc kubenswrapper[4612]: I0227 08:54:18.852649 4612 scope.go:117] "RemoveContainer" containerID="2ed59616da422731358981fd9cdfe7932267d24e60d616b16a3481f816f614da" Feb 27 08:54:18 crc kubenswrapper[4612]: E0227 08:54:18.853349 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-924vb_openshift-machine-config-operator(24599373-7adc-4a1b-8bb4-797bf726f43d)\"" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" Feb 27 08:54:19 crc kubenswrapper[4612]: I0227 08:54:19.460550 4612 scope.go:117] "RemoveContainer" containerID="851b5ed6baadc16ffbd744acc297674fdde486cc5393a8074492364f31670d24" Feb 27 08:54:31 crc kubenswrapper[4612]: I0227 08:54:31.853927 4612 scope.go:117] "RemoveContainer" containerID="2ed59616da422731358981fd9cdfe7932267d24e60d616b16a3481f816f614da" Feb 27 08:54:31 crc kubenswrapper[4612]: E0227 08:54:31.855149 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-924vb_openshift-machine-config-operator(24599373-7adc-4a1b-8bb4-797bf726f43d)\"" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" Feb 27 08:54:45 crc kubenswrapper[4612]: I0227 08:54:45.853854 4612 scope.go:117] "RemoveContainer" containerID="2ed59616da422731358981fd9cdfe7932267d24e60d616b16a3481f816f614da" Feb 27 08:54:45 crc kubenswrapper[4612]: E0227 08:54:45.855048 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-924vb_openshift-machine-config-operator(24599373-7adc-4a1b-8bb4-797bf726f43d)\"" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" Feb 27 08:55:00 crc kubenswrapper[4612]: I0227 08:55:00.854657 4612 scope.go:117] "RemoveContainer" containerID="2ed59616da422731358981fd9cdfe7932267d24e60d616b16a3481f816f614da" Feb 27 08:55:01 crc kubenswrapper[4612]: I0227 08:55:01.414824 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-924vb" event={"ID":"24599373-7adc-4a1b-8bb4-797bf726f43d","Type":"ContainerStarted","Data":"cebeb2cb354fa601f37e4c7cc40bcc1dcd010ada59d86d91448073fa6deac102"} Feb 27 08:56:00 crc kubenswrapper[4612]: I0227 08:56:00.163240 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29536376-p8z66"] Feb 27 08:56:00 crc kubenswrapper[4612]: E0227 08:56:00.164226 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce53ef57-b5de-4fe6-adde-10bbec8f696d" containerName="oc" Feb 27 08:56:00 crc kubenswrapper[4612]: I0227 08:56:00.164244 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce53ef57-b5de-4fe6-adde-10bbec8f696d" containerName="oc" Feb 27 08:56:00 crc kubenswrapper[4612]: I0227 08:56:00.164456 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="ce53ef57-b5de-4fe6-adde-10bbec8f696d" containerName="oc" Feb 27 08:56:00 crc kubenswrapper[4612]: I0227 08:56:00.165114 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536376-p8z66" Feb 27 08:56:00 crc kubenswrapper[4612]: I0227 08:56:00.168053 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-5zwfx" Feb 27 08:56:00 crc kubenswrapper[4612]: I0227 08:56:00.169666 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 27 08:56:00 crc kubenswrapper[4612]: I0227 08:56:00.179293 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536376-p8z66"] Feb 27 08:56:00 crc kubenswrapper[4612]: I0227 08:56:00.181837 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 27 08:56:00 crc kubenswrapper[4612]: I0227 08:56:00.212953 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t5ksn\" (UniqueName: \"kubernetes.io/projected/88ada8ea-dcbb-40c2-9599-b1fc0e3f24d6-kube-api-access-t5ksn\") pod \"auto-csr-approver-29536376-p8z66\" (UID: \"88ada8ea-dcbb-40c2-9599-b1fc0e3f24d6\") " pod="openshift-infra/auto-csr-approver-29536376-p8z66" Feb 27 08:56:00 crc kubenswrapper[4612]: I0227 08:56:00.337561 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t5ksn\" (UniqueName: \"kubernetes.io/projected/88ada8ea-dcbb-40c2-9599-b1fc0e3f24d6-kube-api-access-t5ksn\") pod \"auto-csr-approver-29536376-p8z66\" (UID: \"88ada8ea-dcbb-40c2-9599-b1fc0e3f24d6\") " pod="openshift-infra/auto-csr-approver-29536376-p8z66" Feb 27 08:56:00 crc kubenswrapper[4612]: I0227 08:56:00.372421 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t5ksn\" (UniqueName: \"kubernetes.io/projected/88ada8ea-dcbb-40c2-9599-b1fc0e3f24d6-kube-api-access-t5ksn\") pod \"auto-csr-approver-29536376-p8z66\" (UID: \"88ada8ea-dcbb-40c2-9599-b1fc0e3f24d6\") " pod="openshift-infra/auto-csr-approver-29536376-p8z66" Feb 27 08:56:00 crc kubenswrapper[4612]: I0227 08:56:00.484100 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536376-p8z66" Feb 27 08:56:00 crc kubenswrapper[4612]: I0227 08:56:00.968671 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536376-p8z66"] Feb 27 08:56:00 crc kubenswrapper[4612]: W0227 08:56:00.972490 4612 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod88ada8ea_dcbb_40c2_9599_b1fc0e3f24d6.slice/crio-87033304d0a9ba59bfebf2be31f5152ea54e4a5d4de239f72b1b94b523936edb WatchSource:0}: Error finding container 87033304d0a9ba59bfebf2be31f5152ea54e4a5d4de239f72b1b94b523936edb: Status 404 returned error can't find the container with id 87033304d0a9ba59bfebf2be31f5152ea54e4a5d4de239f72b1b94b523936edb Feb 27 08:56:01 crc kubenswrapper[4612]: I0227 08:56:01.038894 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536376-p8z66" event={"ID":"88ada8ea-dcbb-40c2-9599-b1fc0e3f24d6","Type":"ContainerStarted","Data":"87033304d0a9ba59bfebf2be31f5152ea54e4a5d4de239f72b1b94b523936edb"} Feb 27 08:56:03 crc kubenswrapper[4612]: I0227 08:56:03.057047 4612 generic.go:334] "Generic (PLEG): container finished" podID="88ada8ea-dcbb-40c2-9599-b1fc0e3f24d6" containerID="f226423c80e8686c070bc3455d4e33aab449e5154d17dfcbc3d5f3d441c5040d" exitCode=0 Feb 27 08:56:03 crc kubenswrapper[4612]: I0227 08:56:03.057104 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536376-p8z66" event={"ID":"88ada8ea-dcbb-40c2-9599-b1fc0e3f24d6","Type":"ContainerDied","Data":"f226423c80e8686c070bc3455d4e33aab449e5154d17dfcbc3d5f3d441c5040d"} Feb 27 08:56:04 crc kubenswrapper[4612]: I0227 08:56:04.446224 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536376-p8z66" Feb 27 08:56:04 crc kubenswrapper[4612]: I0227 08:56:04.521576 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t5ksn\" (UniqueName: \"kubernetes.io/projected/88ada8ea-dcbb-40c2-9599-b1fc0e3f24d6-kube-api-access-t5ksn\") pod \"88ada8ea-dcbb-40c2-9599-b1fc0e3f24d6\" (UID: \"88ada8ea-dcbb-40c2-9599-b1fc0e3f24d6\") " Feb 27 08:56:04 crc kubenswrapper[4612]: I0227 08:56:04.530411 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/88ada8ea-dcbb-40c2-9599-b1fc0e3f24d6-kube-api-access-t5ksn" (OuterVolumeSpecName: "kube-api-access-t5ksn") pod "88ada8ea-dcbb-40c2-9599-b1fc0e3f24d6" (UID: "88ada8ea-dcbb-40c2-9599-b1fc0e3f24d6"). InnerVolumeSpecName "kube-api-access-t5ksn". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 08:56:04 crc kubenswrapper[4612]: I0227 08:56:04.625033 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t5ksn\" (UniqueName: \"kubernetes.io/projected/88ada8ea-dcbb-40c2-9599-b1fc0e3f24d6-kube-api-access-t5ksn\") on node \"crc\" DevicePath \"\"" Feb 27 08:56:05 crc kubenswrapper[4612]: I0227 08:56:05.082321 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536376-p8z66" event={"ID":"88ada8ea-dcbb-40c2-9599-b1fc0e3f24d6","Type":"ContainerDied","Data":"87033304d0a9ba59bfebf2be31f5152ea54e4a5d4de239f72b1b94b523936edb"} Feb 27 08:56:05 crc kubenswrapper[4612]: I0227 08:56:05.082666 4612 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="87033304d0a9ba59bfebf2be31f5152ea54e4a5d4de239f72b1b94b523936edb" Feb 27 08:56:05 crc kubenswrapper[4612]: I0227 08:56:05.082383 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536376-p8z66" Feb 27 08:56:05 crc kubenswrapper[4612]: I0227 08:56:05.528011 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29536370-wp582"] Feb 27 08:56:05 crc kubenswrapper[4612]: I0227 08:56:05.536003 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29536370-wp582"] Feb 27 08:56:06 crc kubenswrapper[4612]: I0227 08:56:06.867176 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="352f7228-6eac-4b53-a13f-56a291327036" path="/var/lib/kubelet/pods/352f7228-6eac-4b53-a13f-56a291327036/volumes" Feb 27 08:56:19 crc kubenswrapper[4612]: I0227 08:56:19.627311 4612 scope.go:117] "RemoveContainer" containerID="692e3fe94710ca549bee22efb82dd732e5aacee245610422e69ed52e893e04a0" Feb 27 08:56:43 crc kubenswrapper[4612]: I0227 08:56:43.532523 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-l89v6"] Feb 27 08:56:43 crc kubenswrapper[4612]: E0227 08:56:43.533754 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="88ada8ea-dcbb-40c2-9599-b1fc0e3f24d6" containerName="oc" Feb 27 08:56:43 crc kubenswrapper[4612]: I0227 08:56:43.533776 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="88ada8ea-dcbb-40c2-9599-b1fc0e3f24d6" containerName="oc" Feb 27 08:56:43 crc kubenswrapper[4612]: I0227 08:56:43.534141 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="88ada8ea-dcbb-40c2-9599-b1fc0e3f24d6" containerName="oc" Feb 27 08:56:43 crc kubenswrapper[4612]: I0227 08:56:43.536222 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-l89v6" Feb 27 08:56:43 crc kubenswrapper[4612]: I0227 08:56:43.552341 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-l89v6"] Feb 27 08:56:43 crc kubenswrapper[4612]: I0227 08:56:43.652372 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9dbf4447-0cfe-4ed6-b82d-266e17dbd6d2-utilities\") pod \"redhat-marketplace-l89v6\" (UID: \"9dbf4447-0cfe-4ed6-b82d-266e17dbd6d2\") " pod="openshift-marketplace/redhat-marketplace-l89v6" Feb 27 08:56:43 crc kubenswrapper[4612]: I0227 08:56:43.652568 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w8q8t\" (UniqueName: \"kubernetes.io/projected/9dbf4447-0cfe-4ed6-b82d-266e17dbd6d2-kube-api-access-w8q8t\") pod \"redhat-marketplace-l89v6\" (UID: \"9dbf4447-0cfe-4ed6-b82d-266e17dbd6d2\") " pod="openshift-marketplace/redhat-marketplace-l89v6" Feb 27 08:56:43 crc kubenswrapper[4612]: I0227 08:56:43.652824 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9dbf4447-0cfe-4ed6-b82d-266e17dbd6d2-catalog-content\") pod \"redhat-marketplace-l89v6\" (UID: \"9dbf4447-0cfe-4ed6-b82d-266e17dbd6d2\") " pod="openshift-marketplace/redhat-marketplace-l89v6" Feb 27 08:56:43 crc kubenswrapper[4612]: I0227 08:56:43.755087 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9dbf4447-0cfe-4ed6-b82d-266e17dbd6d2-catalog-content\") pod \"redhat-marketplace-l89v6\" (UID: \"9dbf4447-0cfe-4ed6-b82d-266e17dbd6d2\") " pod="openshift-marketplace/redhat-marketplace-l89v6" Feb 27 08:56:43 crc kubenswrapper[4612]: I0227 08:56:43.755147 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9dbf4447-0cfe-4ed6-b82d-266e17dbd6d2-utilities\") pod \"redhat-marketplace-l89v6\" (UID: \"9dbf4447-0cfe-4ed6-b82d-266e17dbd6d2\") " pod="openshift-marketplace/redhat-marketplace-l89v6" Feb 27 08:56:43 crc kubenswrapper[4612]: I0227 08:56:43.755210 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w8q8t\" (UniqueName: \"kubernetes.io/projected/9dbf4447-0cfe-4ed6-b82d-266e17dbd6d2-kube-api-access-w8q8t\") pod \"redhat-marketplace-l89v6\" (UID: \"9dbf4447-0cfe-4ed6-b82d-266e17dbd6d2\") " pod="openshift-marketplace/redhat-marketplace-l89v6" Feb 27 08:56:43 crc kubenswrapper[4612]: I0227 08:56:43.755751 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9dbf4447-0cfe-4ed6-b82d-266e17dbd6d2-catalog-content\") pod \"redhat-marketplace-l89v6\" (UID: \"9dbf4447-0cfe-4ed6-b82d-266e17dbd6d2\") " pod="openshift-marketplace/redhat-marketplace-l89v6" Feb 27 08:56:43 crc kubenswrapper[4612]: I0227 08:56:43.755850 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9dbf4447-0cfe-4ed6-b82d-266e17dbd6d2-utilities\") pod \"redhat-marketplace-l89v6\" (UID: \"9dbf4447-0cfe-4ed6-b82d-266e17dbd6d2\") " pod="openshift-marketplace/redhat-marketplace-l89v6" Feb 27 08:56:43 crc kubenswrapper[4612]: I0227 08:56:43.783043 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w8q8t\" (UniqueName: \"kubernetes.io/projected/9dbf4447-0cfe-4ed6-b82d-266e17dbd6d2-kube-api-access-w8q8t\") pod \"redhat-marketplace-l89v6\" (UID: \"9dbf4447-0cfe-4ed6-b82d-266e17dbd6d2\") " pod="openshift-marketplace/redhat-marketplace-l89v6" Feb 27 08:56:43 crc kubenswrapper[4612]: I0227 08:56:43.863851 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-l89v6" Feb 27 08:56:44 crc kubenswrapper[4612]: I0227 08:56:44.443414 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-l89v6"] Feb 27 08:56:45 crc kubenswrapper[4612]: I0227 08:56:45.447241 4612 generic.go:334] "Generic (PLEG): container finished" podID="9dbf4447-0cfe-4ed6-b82d-266e17dbd6d2" containerID="ddfd368a0992a1cebbb0139b039398e19a022c2b24ab1a96b954325648fc7251" exitCode=0 Feb 27 08:56:45 crc kubenswrapper[4612]: I0227 08:56:45.447351 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-l89v6" event={"ID":"9dbf4447-0cfe-4ed6-b82d-266e17dbd6d2","Type":"ContainerDied","Data":"ddfd368a0992a1cebbb0139b039398e19a022c2b24ab1a96b954325648fc7251"} Feb 27 08:56:45 crc kubenswrapper[4612]: I0227 08:56:45.447630 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-l89v6" event={"ID":"9dbf4447-0cfe-4ed6-b82d-266e17dbd6d2","Type":"ContainerStarted","Data":"f59fcd4b0f8f235c438fb183d30229bc918874f25e2f508b3f46f31e03d42870"} Feb 27 08:56:47 crc kubenswrapper[4612]: I0227 08:56:47.467456 4612 generic.go:334] "Generic (PLEG): container finished" podID="9dbf4447-0cfe-4ed6-b82d-266e17dbd6d2" containerID="4d97cafd9d7657f40945fe8bdf7ad20da96c7b34ce23d893642e06b96035663f" exitCode=0 Feb 27 08:56:47 crc kubenswrapper[4612]: I0227 08:56:47.468047 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-l89v6" event={"ID":"9dbf4447-0cfe-4ed6-b82d-266e17dbd6d2","Type":"ContainerDied","Data":"4d97cafd9d7657f40945fe8bdf7ad20da96c7b34ce23d893642e06b96035663f"} Feb 27 08:56:48 crc kubenswrapper[4612]: I0227 08:56:48.478546 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-l89v6" event={"ID":"9dbf4447-0cfe-4ed6-b82d-266e17dbd6d2","Type":"ContainerStarted","Data":"70b704327dc8fed42216f171da3c2ef07c0781fddc4b650193560821b4b34c27"} Feb 27 08:56:48 crc kubenswrapper[4612]: I0227 08:56:48.498064 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-l89v6" podStartSLOduration=3.034686998 podStartE2EDuration="5.498047269s" podCreationTimestamp="2026-02-27 08:56:43 +0000 UTC" firstStartedPulling="2026-02-27 08:56:45.449470105 +0000 UTC m=+4063.303400103" lastFinishedPulling="2026-02-27 08:56:47.912830366 +0000 UTC m=+4065.766760374" observedRunningTime="2026-02-27 08:56:48.495076255 +0000 UTC m=+4066.349006263" watchObservedRunningTime="2026-02-27 08:56:48.498047269 +0000 UTC m=+4066.351977267" Feb 27 08:56:53 crc kubenswrapper[4612]: I0227 08:56:53.864642 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-l89v6" Feb 27 08:56:53 crc kubenswrapper[4612]: I0227 08:56:53.866088 4612 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-l89v6" Feb 27 08:56:53 crc kubenswrapper[4612]: I0227 08:56:53.944760 4612 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-l89v6" Feb 27 08:56:54 crc kubenswrapper[4612]: I0227 08:56:54.591180 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-l89v6" Feb 27 08:56:54 crc kubenswrapper[4612]: I0227 08:56:54.647716 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-l89v6"] Feb 27 08:56:56 crc kubenswrapper[4612]: I0227 08:56:56.547362 4612 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-l89v6" podUID="9dbf4447-0cfe-4ed6-b82d-266e17dbd6d2" containerName="registry-server" containerID="cri-o://70b704327dc8fed42216f171da3c2ef07c0781fddc4b650193560821b4b34c27" gracePeriod=2 Feb 27 08:56:57 crc kubenswrapper[4612]: I0227 08:56:57.797012 4612 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/swift-proxy-59ddd57cfc-qvv4h" podUID="2e631b0d-d44b-472b-90d9-6794cf71d3fd" containerName="proxy-server" probeResult="failure" output="HTTP probe failed with statuscode: 502" Feb 27 08:56:58 crc kubenswrapper[4612]: I0227 08:56:58.638351 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-l89v6" Feb 27 08:56:58 crc kubenswrapper[4612]: I0227 08:56:58.685125 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9dbf4447-0cfe-4ed6-b82d-266e17dbd6d2-catalog-content\") pod \"9dbf4447-0cfe-4ed6-b82d-266e17dbd6d2\" (UID: \"9dbf4447-0cfe-4ed6-b82d-266e17dbd6d2\") " Feb 27 08:56:58 crc kubenswrapper[4612]: I0227 08:56:58.688149 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w8q8t\" (UniqueName: \"kubernetes.io/projected/9dbf4447-0cfe-4ed6-b82d-266e17dbd6d2-kube-api-access-w8q8t\") pod \"9dbf4447-0cfe-4ed6-b82d-266e17dbd6d2\" (UID: \"9dbf4447-0cfe-4ed6-b82d-266e17dbd6d2\") " Feb 27 08:56:58 crc kubenswrapper[4612]: I0227 08:56:58.688253 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9dbf4447-0cfe-4ed6-b82d-266e17dbd6d2-utilities\") pod \"9dbf4447-0cfe-4ed6-b82d-266e17dbd6d2\" (UID: \"9dbf4447-0cfe-4ed6-b82d-266e17dbd6d2\") " Feb 27 08:56:58 crc kubenswrapper[4612]: I0227 08:56:58.689068 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9dbf4447-0cfe-4ed6-b82d-266e17dbd6d2-utilities" (OuterVolumeSpecName: "utilities") pod "9dbf4447-0cfe-4ed6-b82d-266e17dbd6d2" (UID: "9dbf4447-0cfe-4ed6-b82d-266e17dbd6d2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 08:56:58 crc kubenswrapper[4612]: I0227 08:56:58.712048 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9dbf4447-0cfe-4ed6-b82d-266e17dbd6d2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9dbf4447-0cfe-4ed6-b82d-266e17dbd6d2" (UID: "9dbf4447-0cfe-4ed6-b82d-266e17dbd6d2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 08:56:58 crc kubenswrapper[4612]: I0227 08:56:58.784714 4612 generic.go:334] "Generic (PLEG): container finished" podID="9dbf4447-0cfe-4ed6-b82d-266e17dbd6d2" containerID="70b704327dc8fed42216f171da3c2ef07c0781fddc4b650193560821b4b34c27" exitCode=0 Feb 27 08:56:58 crc kubenswrapper[4612]: I0227 08:56:58.784798 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-l89v6" event={"ID":"9dbf4447-0cfe-4ed6-b82d-266e17dbd6d2","Type":"ContainerDied","Data":"70b704327dc8fed42216f171da3c2ef07c0781fddc4b650193560821b4b34c27"} Feb 27 08:56:58 crc kubenswrapper[4612]: I0227 08:56:58.784819 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-l89v6" Feb 27 08:56:58 crc kubenswrapper[4612]: I0227 08:56:58.785161 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-l89v6" event={"ID":"9dbf4447-0cfe-4ed6-b82d-266e17dbd6d2","Type":"ContainerDied","Data":"f59fcd4b0f8f235c438fb183d30229bc918874f25e2f508b3f46f31e03d42870"} Feb 27 08:56:58 crc kubenswrapper[4612]: I0227 08:56:58.785192 4612 scope.go:117] "RemoveContainer" containerID="70b704327dc8fed42216f171da3c2ef07c0781fddc4b650193560821b4b34c27" Feb 27 08:56:58 crc kubenswrapper[4612]: I0227 08:56:58.790869 4612 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9dbf4447-0cfe-4ed6-b82d-266e17dbd6d2-utilities\") on node \"crc\" DevicePath \"\"" Feb 27 08:56:58 crc kubenswrapper[4612]: I0227 08:56:58.790957 4612 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9dbf4447-0cfe-4ed6-b82d-266e17dbd6d2-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 27 08:56:58 crc kubenswrapper[4612]: I0227 08:56:58.805410 4612 scope.go:117] "RemoveContainer" containerID="4d97cafd9d7657f40945fe8bdf7ad20da96c7b34ce23d893642e06b96035663f" Feb 27 08:56:59 crc kubenswrapper[4612]: I0227 08:56:59.185989 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9dbf4447-0cfe-4ed6-b82d-266e17dbd6d2-kube-api-access-w8q8t" (OuterVolumeSpecName: "kube-api-access-w8q8t") pod "9dbf4447-0cfe-4ed6-b82d-266e17dbd6d2" (UID: "9dbf4447-0cfe-4ed6-b82d-266e17dbd6d2"). InnerVolumeSpecName "kube-api-access-w8q8t". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 08:56:59 crc kubenswrapper[4612]: I0227 08:56:59.199793 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w8q8t\" (UniqueName: \"kubernetes.io/projected/9dbf4447-0cfe-4ed6-b82d-266e17dbd6d2-kube-api-access-w8q8t\") on node \"crc\" DevicePath \"\"" Feb 27 08:56:59 crc kubenswrapper[4612]: I0227 08:56:59.210440 4612 scope.go:117] "RemoveContainer" containerID="ddfd368a0992a1cebbb0139b039398e19a022c2b24ab1a96b954325648fc7251" Feb 27 08:56:59 crc kubenswrapper[4612]: I0227 08:56:59.397035 4612 scope.go:117] "RemoveContainer" containerID="70b704327dc8fed42216f171da3c2ef07c0781fddc4b650193560821b4b34c27" Feb 27 08:56:59 crc kubenswrapper[4612]: E0227 08:56:59.397578 4612 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"70b704327dc8fed42216f171da3c2ef07c0781fddc4b650193560821b4b34c27\": container with ID starting with 70b704327dc8fed42216f171da3c2ef07c0781fddc4b650193560821b4b34c27 not found: ID does not exist" containerID="70b704327dc8fed42216f171da3c2ef07c0781fddc4b650193560821b4b34c27" Feb 27 08:56:59 crc kubenswrapper[4612]: I0227 08:56:59.397631 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"70b704327dc8fed42216f171da3c2ef07c0781fddc4b650193560821b4b34c27"} err="failed to get container status \"70b704327dc8fed42216f171da3c2ef07c0781fddc4b650193560821b4b34c27\": rpc error: code = NotFound desc = could not find container \"70b704327dc8fed42216f171da3c2ef07c0781fddc4b650193560821b4b34c27\": container with ID starting with 70b704327dc8fed42216f171da3c2ef07c0781fddc4b650193560821b4b34c27 not found: ID does not exist" Feb 27 08:56:59 crc kubenswrapper[4612]: I0227 08:56:59.397662 4612 scope.go:117] "RemoveContainer" containerID="4d97cafd9d7657f40945fe8bdf7ad20da96c7b34ce23d893642e06b96035663f" Feb 27 08:56:59 crc kubenswrapper[4612]: E0227 08:56:59.398016 4612 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4d97cafd9d7657f40945fe8bdf7ad20da96c7b34ce23d893642e06b96035663f\": container with ID starting with 4d97cafd9d7657f40945fe8bdf7ad20da96c7b34ce23d893642e06b96035663f not found: ID does not exist" containerID="4d97cafd9d7657f40945fe8bdf7ad20da96c7b34ce23d893642e06b96035663f" Feb 27 08:56:59 crc kubenswrapper[4612]: I0227 08:56:59.398041 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4d97cafd9d7657f40945fe8bdf7ad20da96c7b34ce23d893642e06b96035663f"} err="failed to get container status \"4d97cafd9d7657f40945fe8bdf7ad20da96c7b34ce23d893642e06b96035663f\": rpc error: code = NotFound desc = could not find container \"4d97cafd9d7657f40945fe8bdf7ad20da96c7b34ce23d893642e06b96035663f\": container with ID starting with 4d97cafd9d7657f40945fe8bdf7ad20da96c7b34ce23d893642e06b96035663f not found: ID does not exist" Feb 27 08:56:59 crc kubenswrapper[4612]: I0227 08:56:59.398056 4612 scope.go:117] "RemoveContainer" containerID="ddfd368a0992a1cebbb0139b039398e19a022c2b24ab1a96b954325648fc7251" Feb 27 08:56:59 crc kubenswrapper[4612]: E0227 08:56:59.398407 4612 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ddfd368a0992a1cebbb0139b039398e19a022c2b24ab1a96b954325648fc7251\": container with ID starting with ddfd368a0992a1cebbb0139b039398e19a022c2b24ab1a96b954325648fc7251 not found: ID does not exist" containerID="ddfd368a0992a1cebbb0139b039398e19a022c2b24ab1a96b954325648fc7251" Feb 27 08:56:59 crc kubenswrapper[4612]: I0227 08:56:59.398430 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ddfd368a0992a1cebbb0139b039398e19a022c2b24ab1a96b954325648fc7251"} err="failed to get container status \"ddfd368a0992a1cebbb0139b039398e19a022c2b24ab1a96b954325648fc7251\": rpc error: code = NotFound desc = could not find container \"ddfd368a0992a1cebbb0139b039398e19a022c2b24ab1a96b954325648fc7251\": container with ID starting with ddfd368a0992a1cebbb0139b039398e19a022c2b24ab1a96b954325648fc7251 not found: ID does not exist" Feb 27 08:56:59 crc kubenswrapper[4612]: I0227 08:56:59.465285 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-l89v6"] Feb 27 08:56:59 crc kubenswrapper[4612]: I0227 08:56:59.474756 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-l89v6"] Feb 27 08:57:00 crc kubenswrapper[4612]: I0227 08:57:00.863617 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9dbf4447-0cfe-4ed6-b82d-266e17dbd6d2" path="/var/lib/kubelet/pods/9dbf4447-0cfe-4ed6-b82d-266e17dbd6d2/volumes" Feb 27 08:57:09 crc kubenswrapper[4612]: I0227 08:57:09.078393 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-pgbqx"] Feb 27 08:57:09 crc kubenswrapper[4612]: E0227 08:57:09.080445 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9dbf4447-0cfe-4ed6-b82d-266e17dbd6d2" containerName="extract-utilities" Feb 27 08:57:09 crc kubenswrapper[4612]: I0227 08:57:09.080521 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="9dbf4447-0cfe-4ed6-b82d-266e17dbd6d2" containerName="extract-utilities" Feb 27 08:57:09 crc kubenswrapper[4612]: E0227 08:57:09.080584 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9dbf4447-0cfe-4ed6-b82d-266e17dbd6d2" containerName="extract-content" Feb 27 08:57:09 crc kubenswrapper[4612]: I0227 08:57:09.080637 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="9dbf4447-0cfe-4ed6-b82d-266e17dbd6d2" containerName="extract-content" Feb 27 08:57:09 crc kubenswrapper[4612]: E0227 08:57:09.080718 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9dbf4447-0cfe-4ed6-b82d-266e17dbd6d2" containerName="registry-server" Feb 27 08:57:09 crc kubenswrapper[4612]: I0227 08:57:09.080780 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="9dbf4447-0cfe-4ed6-b82d-266e17dbd6d2" containerName="registry-server" Feb 27 08:57:09 crc kubenswrapper[4612]: I0227 08:57:09.081015 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="9dbf4447-0cfe-4ed6-b82d-266e17dbd6d2" containerName="registry-server" Feb 27 08:57:09 crc kubenswrapper[4612]: I0227 08:57:09.082351 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-pgbqx" Feb 27 08:57:09 crc kubenswrapper[4612]: I0227 08:57:09.111094 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-pgbqx"] Feb 27 08:57:09 crc kubenswrapper[4612]: I0227 08:57:09.198542 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qcsxm\" (UniqueName: \"kubernetes.io/projected/0ea808c4-f9ac-4d97-9ca6-d93728854104-kube-api-access-qcsxm\") pod \"certified-operators-pgbqx\" (UID: \"0ea808c4-f9ac-4d97-9ca6-d93728854104\") " pod="openshift-marketplace/certified-operators-pgbqx" Feb 27 08:57:09 crc kubenswrapper[4612]: I0227 08:57:09.198609 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0ea808c4-f9ac-4d97-9ca6-d93728854104-catalog-content\") pod \"certified-operators-pgbqx\" (UID: \"0ea808c4-f9ac-4d97-9ca6-d93728854104\") " pod="openshift-marketplace/certified-operators-pgbqx" Feb 27 08:57:09 crc kubenswrapper[4612]: I0227 08:57:09.198637 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0ea808c4-f9ac-4d97-9ca6-d93728854104-utilities\") pod \"certified-operators-pgbqx\" (UID: \"0ea808c4-f9ac-4d97-9ca6-d93728854104\") " pod="openshift-marketplace/certified-operators-pgbqx" Feb 27 08:57:09 crc kubenswrapper[4612]: I0227 08:57:09.300747 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qcsxm\" (UniqueName: \"kubernetes.io/projected/0ea808c4-f9ac-4d97-9ca6-d93728854104-kube-api-access-qcsxm\") pod \"certified-operators-pgbqx\" (UID: \"0ea808c4-f9ac-4d97-9ca6-d93728854104\") " pod="openshift-marketplace/certified-operators-pgbqx" Feb 27 08:57:09 crc kubenswrapper[4612]: I0227 08:57:09.300821 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0ea808c4-f9ac-4d97-9ca6-d93728854104-catalog-content\") pod \"certified-operators-pgbqx\" (UID: \"0ea808c4-f9ac-4d97-9ca6-d93728854104\") " pod="openshift-marketplace/certified-operators-pgbqx" Feb 27 08:57:09 crc kubenswrapper[4612]: I0227 08:57:09.300847 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0ea808c4-f9ac-4d97-9ca6-d93728854104-utilities\") pod \"certified-operators-pgbqx\" (UID: \"0ea808c4-f9ac-4d97-9ca6-d93728854104\") " pod="openshift-marketplace/certified-operators-pgbqx" Feb 27 08:57:09 crc kubenswrapper[4612]: I0227 08:57:09.301404 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0ea808c4-f9ac-4d97-9ca6-d93728854104-utilities\") pod \"certified-operators-pgbqx\" (UID: \"0ea808c4-f9ac-4d97-9ca6-d93728854104\") " pod="openshift-marketplace/certified-operators-pgbqx" Feb 27 08:57:09 crc kubenswrapper[4612]: I0227 08:57:09.302089 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0ea808c4-f9ac-4d97-9ca6-d93728854104-catalog-content\") pod \"certified-operators-pgbqx\" (UID: \"0ea808c4-f9ac-4d97-9ca6-d93728854104\") " pod="openshift-marketplace/certified-operators-pgbqx" Feb 27 08:57:09 crc kubenswrapper[4612]: I0227 08:57:09.322759 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qcsxm\" (UniqueName: \"kubernetes.io/projected/0ea808c4-f9ac-4d97-9ca6-d93728854104-kube-api-access-qcsxm\") pod \"certified-operators-pgbqx\" (UID: \"0ea808c4-f9ac-4d97-9ca6-d93728854104\") " pod="openshift-marketplace/certified-operators-pgbqx" Feb 27 08:57:09 crc kubenswrapper[4612]: I0227 08:57:09.416218 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-pgbqx" Feb 27 08:57:09 crc kubenswrapper[4612]: I0227 08:57:09.922144 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-pgbqx"] Feb 27 08:57:10 crc kubenswrapper[4612]: I0227 08:57:10.926587 4612 generic.go:334] "Generic (PLEG): container finished" podID="0ea808c4-f9ac-4d97-9ca6-d93728854104" containerID="13cca9605170f672105e575f3929e77aeb1d1794fcb231960134e51a0e4df94a" exitCode=0 Feb 27 08:57:10 crc kubenswrapper[4612]: I0227 08:57:10.926652 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pgbqx" event={"ID":"0ea808c4-f9ac-4d97-9ca6-d93728854104","Type":"ContainerDied","Data":"13cca9605170f672105e575f3929e77aeb1d1794fcb231960134e51a0e4df94a"} Feb 27 08:57:10 crc kubenswrapper[4612]: I0227 08:57:10.926870 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pgbqx" event={"ID":"0ea808c4-f9ac-4d97-9ca6-d93728854104","Type":"ContainerStarted","Data":"d744fb8ae67f3aa51f0e278cf2b25a8e14bb7ff6b25422354c867dec2ef7a783"} Feb 27 08:57:13 crc kubenswrapper[4612]: I0227 08:57:13.960127 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pgbqx" event={"ID":"0ea808c4-f9ac-4d97-9ca6-d93728854104","Type":"ContainerStarted","Data":"d669066cd153de33cd797e56a7035e26f887b145796d4c311cfb471340afbb04"} Feb 27 08:57:14 crc kubenswrapper[4612]: I0227 08:57:14.982505 4612 generic.go:334] "Generic (PLEG): container finished" podID="4700642e-da3a-4e53-84de-0bf4c1870ccd" containerID="9420be57aa23b5a8049b44b757f2d7a73097b7dd7c3f1abfba3c6c1c0242cf7e" exitCode=0 Feb 27 08:57:14 crc kubenswrapper[4612]: I0227 08:57:14.982636 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"4700642e-da3a-4e53-84de-0bf4c1870ccd","Type":"ContainerDied","Data":"9420be57aa23b5a8049b44b757f2d7a73097b7dd7c3f1abfba3c6c1c0242cf7e"} Feb 27 08:57:16 crc kubenswrapper[4612]: I0227 08:57:16.028225 4612 patch_prober.go:28] interesting pod/machine-config-daemon-924vb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 27 08:57:16 crc kubenswrapper[4612]: I0227 08:57:16.029352 4612 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 27 08:57:16 crc kubenswrapper[4612]: I0227 08:57:16.370511 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Feb 27 08:57:16 crc kubenswrapper[4612]: I0227 08:57:16.455088 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/4700642e-da3a-4e53-84de-0bf4c1870ccd-test-operator-ephemeral-temporary\") pod \"4700642e-da3a-4e53-84de-0bf4c1870ccd\" (UID: \"4700642e-da3a-4e53-84de-0bf4c1870ccd\") " Feb 27 08:57:16 crc kubenswrapper[4612]: I0227 08:57:16.455176 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/4700642e-da3a-4e53-84de-0bf4c1870ccd-config-data\") pod \"4700642e-da3a-4e53-84de-0bf4c1870ccd\" (UID: \"4700642e-da3a-4e53-84de-0bf4c1870ccd\") " Feb 27 08:57:16 crc kubenswrapper[4612]: I0227 08:57:16.455243 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/4700642e-da3a-4e53-84de-0bf4c1870ccd-openstack-config\") pod \"4700642e-da3a-4e53-84de-0bf4c1870ccd\" (UID: \"4700642e-da3a-4e53-84de-0bf4c1870ccd\") " Feb 27 08:57:16 crc kubenswrapper[4612]: I0227 08:57:16.455292 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4700642e-da3a-4e53-84de-0bf4c1870ccd-ssh-key\") pod \"4700642e-da3a-4e53-84de-0bf4c1870ccd\" (UID: \"4700642e-da3a-4e53-84de-0bf4c1870ccd\") " Feb 27 08:57:16 crc kubenswrapper[4612]: I0227 08:57:16.455310 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/4700642e-da3a-4e53-84de-0bf4c1870ccd-openstack-config-secret\") pod \"4700642e-da3a-4e53-84de-0bf4c1870ccd\" (UID: \"4700642e-da3a-4e53-84de-0bf4c1870ccd\") " Feb 27 08:57:16 crc kubenswrapper[4612]: I0227 08:57:16.455362 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/4700642e-da3a-4e53-84de-0bf4c1870ccd-test-operator-ephemeral-workdir\") pod \"4700642e-da3a-4e53-84de-0bf4c1870ccd\" (UID: \"4700642e-da3a-4e53-84de-0bf4c1870ccd\") " Feb 27 08:57:16 crc kubenswrapper[4612]: I0227 08:57:16.455383 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/4700642e-da3a-4e53-84de-0bf4c1870ccd-ca-certs\") pod \"4700642e-da3a-4e53-84de-0bf4c1870ccd\" (UID: \"4700642e-da3a-4e53-84de-0bf4c1870ccd\") " Feb 27 08:57:16 crc kubenswrapper[4612]: I0227 08:57:16.455403 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-logs\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"4700642e-da3a-4e53-84de-0bf4c1870ccd\" (UID: \"4700642e-da3a-4e53-84de-0bf4c1870ccd\") " Feb 27 08:57:16 crc kubenswrapper[4612]: I0227 08:57:16.455457 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-69g4w\" (UniqueName: \"kubernetes.io/projected/4700642e-da3a-4e53-84de-0bf4c1870ccd-kube-api-access-69g4w\") pod \"4700642e-da3a-4e53-84de-0bf4c1870ccd\" (UID: \"4700642e-da3a-4e53-84de-0bf4c1870ccd\") " Feb 27 08:57:16 crc kubenswrapper[4612]: I0227 08:57:16.458853 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4700642e-da3a-4e53-84de-0bf4c1870ccd-test-operator-ephemeral-temporary" (OuterVolumeSpecName: "test-operator-ephemeral-temporary") pod "4700642e-da3a-4e53-84de-0bf4c1870ccd" (UID: "4700642e-da3a-4e53-84de-0bf4c1870ccd"). InnerVolumeSpecName "test-operator-ephemeral-temporary". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 08:57:16 crc kubenswrapper[4612]: I0227 08:57:16.459158 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4700642e-da3a-4e53-84de-0bf4c1870ccd-config-data" (OuterVolumeSpecName: "config-data") pod "4700642e-da3a-4e53-84de-0bf4c1870ccd" (UID: "4700642e-da3a-4e53-84de-0bf4c1870ccd"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 08:57:16 crc kubenswrapper[4612]: I0227 08:57:16.463158 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4700642e-da3a-4e53-84de-0bf4c1870ccd-test-operator-ephemeral-workdir" (OuterVolumeSpecName: "test-operator-ephemeral-workdir") pod "4700642e-da3a-4e53-84de-0bf4c1870ccd" (UID: "4700642e-da3a-4e53-84de-0bf4c1870ccd"). InnerVolumeSpecName "test-operator-ephemeral-workdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 08:57:16 crc kubenswrapper[4612]: I0227 08:57:16.464837 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage07-crc" (OuterVolumeSpecName: "test-operator-logs") pod "4700642e-da3a-4e53-84de-0bf4c1870ccd" (UID: "4700642e-da3a-4e53-84de-0bf4c1870ccd"). InnerVolumeSpecName "local-storage07-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Feb 27 08:57:16 crc kubenswrapper[4612]: I0227 08:57:16.466626 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4700642e-da3a-4e53-84de-0bf4c1870ccd-kube-api-access-69g4w" (OuterVolumeSpecName: "kube-api-access-69g4w") pod "4700642e-da3a-4e53-84de-0bf4c1870ccd" (UID: "4700642e-da3a-4e53-84de-0bf4c1870ccd"). InnerVolumeSpecName "kube-api-access-69g4w". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 08:57:16 crc kubenswrapper[4612]: I0227 08:57:16.482855 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4700642e-da3a-4e53-84de-0bf4c1870ccd-ca-certs" (OuterVolumeSpecName: "ca-certs") pod "4700642e-da3a-4e53-84de-0bf4c1870ccd" (UID: "4700642e-da3a-4e53-84de-0bf4c1870ccd"). InnerVolumeSpecName "ca-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:57:16 crc kubenswrapper[4612]: I0227 08:57:16.494788 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4700642e-da3a-4e53-84de-0bf4c1870ccd-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "4700642e-da3a-4e53-84de-0bf4c1870ccd" (UID: "4700642e-da3a-4e53-84de-0bf4c1870ccd"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:57:16 crc kubenswrapper[4612]: I0227 08:57:16.506584 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4700642e-da3a-4e53-84de-0bf4c1870ccd-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "4700642e-da3a-4e53-84de-0bf4c1870ccd" (UID: "4700642e-da3a-4e53-84de-0bf4c1870ccd"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 08:57:16 crc kubenswrapper[4612]: I0227 08:57:16.509949 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4700642e-da3a-4e53-84de-0bf4c1870ccd-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "4700642e-da3a-4e53-84de-0bf4c1870ccd" (UID: "4700642e-da3a-4e53-84de-0bf4c1870ccd"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 08:57:16 crc kubenswrapper[4612]: I0227 08:57:16.558355 4612 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4700642e-da3a-4e53-84de-0bf4c1870ccd-ssh-key\") on node \"crc\" DevicePath \"\"" Feb 27 08:57:16 crc kubenswrapper[4612]: I0227 08:57:16.558403 4612 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/4700642e-da3a-4e53-84de-0bf4c1870ccd-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Feb 27 08:57:16 crc kubenswrapper[4612]: I0227 08:57:16.558419 4612 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/4700642e-da3a-4e53-84de-0bf4c1870ccd-test-operator-ephemeral-workdir\") on node \"crc\" DevicePath \"\"" Feb 27 08:57:16 crc kubenswrapper[4612]: I0227 08:57:16.558428 4612 reconciler_common.go:293] "Volume detached for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/4700642e-da3a-4e53-84de-0bf4c1870ccd-ca-certs\") on node \"crc\" DevicePath \"\"" Feb 27 08:57:16 crc kubenswrapper[4612]: I0227 08:57:16.559117 4612 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" " Feb 27 08:57:16 crc kubenswrapper[4612]: I0227 08:57:16.559136 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-69g4w\" (UniqueName: \"kubernetes.io/projected/4700642e-da3a-4e53-84de-0bf4c1870ccd-kube-api-access-69g4w\") on node \"crc\" DevicePath \"\"" Feb 27 08:57:16 crc kubenswrapper[4612]: I0227 08:57:16.559146 4612 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/4700642e-da3a-4e53-84de-0bf4c1870ccd-test-operator-ephemeral-temporary\") on node \"crc\" DevicePath \"\"" Feb 27 08:57:16 crc kubenswrapper[4612]: I0227 08:57:16.559155 4612 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/4700642e-da3a-4e53-84de-0bf4c1870ccd-config-data\") on node \"crc\" DevicePath \"\"" Feb 27 08:57:16 crc kubenswrapper[4612]: I0227 08:57:16.559165 4612 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/4700642e-da3a-4e53-84de-0bf4c1870ccd-openstack-config\") on node \"crc\" DevicePath \"\"" Feb 27 08:57:16 crc kubenswrapper[4612]: I0227 08:57:16.588129 4612 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage07-crc" (UniqueName: "kubernetes.io/local-volume/local-storage07-crc") on node "crc" Feb 27 08:57:16 crc kubenswrapper[4612]: I0227 08:57:16.661224 4612 reconciler_common.go:293] "Volume detached for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" DevicePath \"\"" Feb 27 08:57:17 crc kubenswrapper[4612]: I0227 08:57:17.004323 4612 generic.go:334] "Generic (PLEG): container finished" podID="0ea808c4-f9ac-4d97-9ca6-d93728854104" containerID="d669066cd153de33cd797e56a7035e26f887b145796d4c311cfb471340afbb04" exitCode=0 Feb 27 08:57:17 crc kubenswrapper[4612]: I0227 08:57:17.004375 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pgbqx" event={"ID":"0ea808c4-f9ac-4d97-9ca6-d93728854104","Type":"ContainerDied","Data":"d669066cd153de33cd797e56a7035e26f887b145796d4c311cfb471340afbb04"} Feb 27 08:57:17 crc kubenswrapper[4612]: I0227 08:57:17.006781 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"4700642e-da3a-4e53-84de-0bf4c1870ccd","Type":"ContainerDied","Data":"cc7f3ff65e1f9decbe0960576c5071da6b09e9a686a9761a1b3bc857d80ff2f0"} Feb 27 08:57:17 crc kubenswrapper[4612]: I0227 08:57:17.006805 4612 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cc7f3ff65e1f9decbe0960576c5071da6b09e9a686a9761a1b3bc857d80ff2f0" Feb 27 08:57:17 crc kubenswrapper[4612]: I0227 08:57:17.006867 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Feb 27 08:57:18 crc kubenswrapper[4612]: I0227 08:57:18.019897 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pgbqx" event={"ID":"0ea808c4-f9ac-4d97-9ca6-d93728854104","Type":"ContainerStarted","Data":"ee26726ae1da5d1d9065b6dfe8b77a95a5dd5fcfdaa82cd9f34d0d0d458166d6"} Feb 27 08:57:18 crc kubenswrapper[4612]: I0227 08:57:18.044060 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-pgbqx" podStartSLOduration=2.561178962 podStartE2EDuration="9.044037195s" podCreationTimestamp="2026-02-27 08:57:09 +0000 UTC" firstStartedPulling="2026-02-27 08:57:10.928268478 +0000 UTC m=+4088.782198486" lastFinishedPulling="2026-02-27 08:57:17.411126731 +0000 UTC m=+4095.265056719" observedRunningTime="2026-02-27 08:57:18.041457822 +0000 UTC m=+4095.895387820" watchObservedRunningTime="2026-02-27 08:57:18.044037195 +0000 UTC m=+4095.897967213" Feb 27 08:57:19 crc kubenswrapper[4612]: I0227 08:57:19.416579 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-pgbqx" Feb 27 08:57:19 crc kubenswrapper[4612]: I0227 08:57:19.416655 4612 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-pgbqx" Feb 27 08:57:20 crc kubenswrapper[4612]: I0227 08:57:20.470151 4612 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-pgbqx" podUID="0ea808c4-f9ac-4d97-9ca6-d93728854104" containerName="registry-server" probeResult="failure" output=< Feb 27 08:57:20 crc kubenswrapper[4612]: timeout: failed to connect service ":50051" within 1s Feb 27 08:57:20 crc kubenswrapper[4612]: > Feb 27 08:57:21 crc kubenswrapper[4612]: I0227 08:57:21.920667 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Feb 27 08:57:21 crc kubenswrapper[4612]: E0227 08:57:21.921459 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4700642e-da3a-4e53-84de-0bf4c1870ccd" containerName="tempest-tests-tempest-tests-runner" Feb 27 08:57:21 crc kubenswrapper[4612]: I0227 08:57:21.921499 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="4700642e-da3a-4e53-84de-0bf4c1870ccd" containerName="tempest-tests-tempest-tests-runner" Feb 27 08:57:21 crc kubenswrapper[4612]: I0227 08:57:21.922025 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="4700642e-da3a-4e53-84de-0bf4c1870ccd" containerName="tempest-tests-tempest-tests-runner" Feb 27 08:57:21 crc kubenswrapper[4612]: I0227 08:57:21.923420 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Feb 27 08:57:21 crc kubenswrapper[4612]: I0227 08:57:21.925582 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-9rfwb" Feb 27 08:57:21 crc kubenswrapper[4612]: I0227 08:57:21.955425 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Feb 27 08:57:22 crc kubenswrapper[4612]: I0227 08:57:22.078902 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-47qfl\" (UniqueName: \"kubernetes.io/projected/8aad3c36-76d6-4991-9494-0725b3e67aea-kube-api-access-47qfl\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"8aad3c36-76d6-4991-9494-0725b3e67aea\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Feb 27 08:57:22 crc kubenswrapper[4612]: I0227 08:57:22.079051 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"8aad3c36-76d6-4991-9494-0725b3e67aea\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Feb 27 08:57:22 crc kubenswrapper[4612]: I0227 08:57:22.181534 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-47qfl\" (UniqueName: \"kubernetes.io/projected/8aad3c36-76d6-4991-9494-0725b3e67aea-kube-api-access-47qfl\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"8aad3c36-76d6-4991-9494-0725b3e67aea\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Feb 27 08:57:22 crc kubenswrapper[4612]: I0227 08:57:22.181946 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"8aad3c36-76d6-4991-9494-0725b3e67aea\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Feb 27 08:57:22 crc kubenswrapper[4612]: I0227 08:57:22.183216 4612 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"8aad3c36-76d6-4991-9494-0725b3e67aea\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Feb 27 08:57:22 crc kubenswrapper[4612]: I0227 08:57:22.210463 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-47qfl\" (UniqueName: \"kubernetes.io/projected/8aad3c36-76d6-4991-9494-0725b3e67aea-kube-api-access-47qfl\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"8aad3c36-76d6-4991-9494-0725b3e67aea\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Feb 27 08:57:22 crc kubenswrapper[4612]: I0227 08:57:22.221489 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"8aad3c36-76d6-4991-9494-0725b3e67aea\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Feb 27 08:57:22 crc kubenswrapper[4612]: I0227 08:57:22.256225 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Feb 27 08:57:22 crc kubenswrapper[4612]: I0227 08:57:22.768549 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Feb 27 08:57:23 crc kubenswrapper[4612]: W0227 08:57:23.083581 4612 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8aad3c36_76d6_4991_9494_0725b3e67aea.slice/crio-3b51281a5e5995c9387442bec8ad7a90d0b96c3ad308a906a05e41ec0b07e9b5 WatchSource:0}: Error finding container 3b51281a5e5995c9387442bec8ad7a90d0b96c3ad308a906a05e41ec0b07e9b5: Status 404 returned error can't find the container with id 3b51281a5e5995c9387442bec8ad7a90d0b96c3ad308a906a05e41ec0b07e9b5 Feb 27 08:57:24 crc kubenswrapper[4612]: I0227 08:57:24.083594 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"8aad3c36-76d6-4991-9494-0725b3e67aea","Type":"ContainerStarted","Data":"3b51281a5e5995c9387442bec8ad7a90d0b96c3ad308a906a05e41ec0b07e9b5"} Feb 27 08:57:25 crc kubenswrapper[4612]: I0227 08:57:25.101266 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"8aad3c36-76d6-4991-9494-0725b3e67aea","Type":"ContainerStarted","Data":"8cfa48ef2b539dc0cf22338c63b5b612053caee3eb5142d763c0efa0adea8334"} Feb 27 08:57:25 crc kubenswrapper[4612]: I0227 08:57:25.127535 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" podStartSLOduration=3.021220758 podStartE2EDuration="4.127487437s" podCreationTimestamp="2026-02-27 08:57:21 +0000 UTC" firstStartedPulling="2026-02-27 08:57:23.088368849 +0000 UTC m=+4100.942298857" lastFinishedPulling="2026-02-27 08:57:24.194635538 +0000 UTC m=+4102.048565536" observedRunningTime="2026-02-27 08:57:25.122621699 +0000 UTC m=+4102.976551737" watchObservedRunningTime="2026-02-27 08:57:25.127487437 +0000 UTC m=+4102.981417475" Feb 27 08:57:29 crc kubenswrapper[4612]: I0227 08:57:29.464199 4612 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-pgbqx" Feb 27 08:57:29 crc kubenswrapper[4612]: I0227 08:57:29.515141 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-pgbqx" Feb 27 08:57:29 crc kubenswrapper[4612]: I0227 08:57:29.699370 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-pgbqx"] Feb 27 08:57:31 crc kubenswrapper[4612]: I0227 08:57:31.158731 4612 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-pgbqx" podUID="0ea808c4-f9ac-4d97-9ca6-d93728854104" containerName="registry-server" containerID="cri-o://ee26726ae1da5d1d9065b6dfe8b77a95a5dd5fcfdaa82cd9f34d0d0d458166d6" gracePeriod=2 Feb 27 08:57:31 crc kubenswrapper[4612]: I0227 08:57:31.689815 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-pgbqx" Feb 27 08:57:31 crc kubenswrapper[4612]: I0227 08:57:31.784966 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0ea808c4-f9ac-4d97-9ca6-d93728854104-utilities\") pod \"0ea808c4-f9ac-4d97-9ca6-d93728854104\" (UID: \"0ea808c4-f9ac-4d97-9ca6-d93728854104\") " Feb 27 08:57:31 crc kubenswrapper[4612]: I0227 08:57:31.785003 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0ea808c4-f9ac-4d97-9ca6-d93728854104-catalog-content\") pod \"0ea808c4-f9ac-4d97-9ca6-d93728854104\" (UID: \"0ea808c4-f9ac-4d97-9ca6-d93728854104\") " Feb 27 08:57:31 crc kubenswrapper[4612]: I0227 08:57:31.785263 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qcsxm\" (UniqueName: \"kubernetes.io/projected/0ea808c4-f9ac-4d97-9ca6-d93728854104-kube-api-access-qcsxm\") pod \"0ea808c4-f9ac-4d97-9ca6-d93728854104\" (UID: \"0ea808c4-f9ac-4d97-9ca6-d93728854104\") " Feb 27 08:57:31 crc kubenswrapper[4612]: I0227 08:57:31.785991 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0ea808c4-f9ac-4d97-9ca6-d93728854104-utilities" (OuterVolumeSpecName: "utilities") pod "0ea808c4-f9ac-4d97-9ca6-d93728854104" (UID: "0ea808c4-f9ac-4d97-9ca6-d93728854104"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 08:57:31 crc kubenswrapper[4612]: I0227 08:57:31.799512 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0ea808c4-f9ac-4d97-9ca6-d93728854104-kube-api-access-qcsxm" (OuterVolumeSpecName: "kube-api-access-qcsxm") pod "0ea808c4-f9ac-4d97-9ca6-d93728854104" (UID: "0ea808c4-f9ac-4d97-9ca6-d93728854104"). InnerVolumeSpecName "kube-api-access-qcsxm". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 08:57:31 crc kubenswrapper[4612]: I0227 08:57:31.848600 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0ea808c4-f9ac-4d97-9ca6-d93728854104-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0ea808c4-f9ac-4d97-9ca6-d93728854104" (UID: "0ea808c4-f9ac-4d97-9ca6-d93728854104"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 08:57:31 crc kubenswrapper[4612]: I0227 08:57:31.887322 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qcsxm\" (UniqueName: \"kubernetes.io/projected/0ea808c4-f9ac-4d97-9ca6-d93728854104-kube-api-access-qcsxm\") on node \"crc\" DevicePath \"\"" Feb 27 08:57:31 crc kubenswrapper[4612]: I0227 08:57:31.887357 4612 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0ea808c4-f9ac-4d97-9ca6-d93728854104-utilities\") on node \"crc\" DevicePath \"\"" Feb 27 08:57:31 crc kubenswrapper[4612]: I0227 08:57:31.887366 4612 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0ea808c4-f9ac-4d97-9ca6-d93728854104-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 27 08:57:32 crc kubenswrapper[4612]: I0227 08:57:32.168868 4612 generic.go:334] "Generic (PLEG): container finished" podID="0ea808c4-f9ac-4d97-9ca6-d93728854104" containerID="ee26726ae1da5d1d9065b6dfe8b77a95a5dd5fcfdaa82cd9f34d0d0d458166d6" exitCode=0 Feb 27 08:57:32 crc kubenswrapper[4612]: I0227 08:57:32.168926 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pgbqx" event={"ID":"0ea808c4-f9ac-4d97-9ca6-d93728854104","Type":"ContainerDied","Data":"ee26726ae1da5d1d9065b6dfe8b77a95a5dd5fcfdaa82cd9f34d0d0d458166d6"} Feb 27 08:57:32 crc kubenswrapper[4612]: I0227 08:57:32.168958 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pgbqx" event={"ID":"0ea808c4-f9ac-4d97-9ca6-d93728854104","Type":"ContainerDied","Data":"d744fb8ae67f3aa51f0e278cf2b25a8e14bb7ff6b25422354c867dec2ef7a783"} Feb 27 08:57:32 crc kubenswrapper[4612]: I0227 08:57:32.168979 4612 scope.go:117] "RemoveContainer" containerID="ee26726ae1da5d1d9065b6dfe8b77a95a5dd5fcfdaa82cd9f34d0d0d458166d6" Feb 27 08:57:32 crc kubenswrapper[4612]: I0227 08:57:32.169137 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-pgbqx" Feb 27 08:57:32 crc kubenswrapper[4612]: I0227 08:57:32.194808 4612 scope.go:117] "RemoveContainer" containerID="d669066cd153de33cd797e56a7035e26f887b145796d4c311cfb471340afbb04" Feb 27 08:57:32 crc kubenswrapper[4612]: I0227 08:57:32.210261 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-pgbqx"] Feb 27 08:57:32 crc kubenswrapper[4612]: I0227 08:57:32.219269 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-pgbqx"] Feb 27 08:57:32 crc kubenswrapper[4612]: I0227 08:57:32.219624 4612 scope.go:117] "RemoveContainer" containerID="13cca9605170f672105e575f3929e77aeb1d1794fcb231960134e51a0e4df94a" Feb 27 08:57:32 crc kubenswrapper[4612]: I0227 08:57:32.271136 4612 scope.go:117] "RemoveContainer" containerID="ee26726ae1da5d1d9065b6dfe8b77a95a5dd5fcfdaa82cd9f34d0d0d458166d6" Feb 27 08:57:32 crc kubenswrapper[4612]: E0227 08:57:32.271744 4612 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ee26726ae1da5d1d9065b6dfe8b77a95a5dd5fcfdaa82cd9f34d0d0d458166d6\": container with ID starting with ee26726ae1da5d1d9065b6dfe8b77a95a5dd5fcfdaa82cd9f34d0d0d458166d6 not found: ID does not exist" containerID="ee26726ae1da5d1d9065b6dfe8b77a95a5dd5fcfdaa82cd9f34d0d0d458166d6" Feb 27 08:57:32 crc kubenswrapper[4612]: I0227 08:57:32.271795 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ee26726ae1da5d1d9065b6dfe8b77a95a5dd5fcfdaa82cd9f34d0d0d458166d6"} err="failed to get container status \"ee26726ae1da5d1d9065b6dfe8b77a95a5dd5fcfdaa82cd9f34d0d0d458166d6\": rpc error: code = NotFound desc = could not find container \"ee26726ae1da5d1d9065b6dfe8b77a95a5dd5fcfdaa82cd9f34d0d0d458166d6\": container with ID starting with ee26726ae1da5d1d9065b6dfe8b77a95a5dd5fcfdaa82cd9f34d0d0d458166d6 not found: ID does not exist" Feb 27 08:57:32 crc kubenswrapper[4612]: I0227 08:57:32.271830 4612 scope.go:117] "RemoveContainer" containerID="d669066cd153de33cd797e56a7035e26f887b145796d4c311cfb471340afbb04" Feb 27 08:57:32 crc kubenswrapper[4612]: E0227 08:57:32.272208 4612 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d669066cd153de33cd797e56a7035e26f887b145796d4c311cfb471340afbb04\": container with ID starting with d669066cd153de33cd797e56a7035e26f887b145796d4c311cfb471340afbb04 not found: ID does not exist" containerID="d669066cd153de33cd797e56a7035e26f887b145796d4c311cfb471340afbb04" Feb 27 08:57:32 crc kubenswrapper[4612]: I0227 08:57:32.272259 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d669066cd153de33cd797e56a7035e26f887b145796d4c311cfb471340afbb04"} err="failed to get container status \"d669066cd153de33cd797e56a7035e26f887b145796d4c311cfb471340afbb04\": rpc error: code = NotFound desc = could not find container \"d669066cd153de33cd797e56a7035e26f887b145796d4c311cfb471340afbb04\": container with ID starting with d669066cd153de33cd797e56a7035e26f887b145796d4c311cfb471340afbb04 not found: ID does not exist" Feb 27 08:57:32 crc kubenswrapper[4612]: I0227 08:57:32.272291 4612 scope.go:117] "RemoveContainer" containerID="13cca9605170f672105e575f3929e77aeb1d1794fcb231960134e51a0e4df94a" Feb 27 08:57:32 crc kubenswrapper[4612]: E0227 08:57:32.272607 4612 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"13cca9605170f672105e575f3929e77aeb1d1794fcb231960134e51a0e4df94a\": container with ID starting with 13cca9605170f672105e575f3929e77aeb1d1794fcb231960134e51a0e4df94a not found: ID does not exist" containerID="13cca9605170f672105e575f3929e77aeb1d1794fcb231960134e51a0e4df94a" Feb 27 08:57:32 crc kubenswrapper[4612]: I0227 08:57:32.272637 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"13cca9605170f672105e575f3929e77aeb1d1794fcb231960134e51a0e4df94a"} err="failed to get container status \"13cca9605170f672105e575f3929e77aeb1d1794fcb231960134e51a0e4df94a\": rpc error: code = NotFound desc = could not find container \"13cca9605170f672105e575f3929e77aeb1d1794fcb231960134e51a0e4df94a\": container with ID starting with 13cca9605170f672105e575f3929e77aeb1d1794fcb231960134e51a0e4df94a not found: ID does not exist" Feb 27 08:57:32 crc kubenswrapper[4612]: I0227 08:57:32.868445 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0ea808c4-f9ac-4d97-9ca6-d93728854104" path="/var/lib/kubelet/pods/0ea808c4-f9ac-4d97-9ca6-d93728854104/volumes" Feb 27 08:57:45 crc kubenswrapper[4612]: I0227 08:57:45.914383 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-mchzb/must-gather-t4rbr"] Feb 27 08:57:45 crc kubenswrapper[4612]: E0227 08:57:45.916337 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0ea808c4-f9ac-4d97-9ca6-d93728854104" containerName="registry-server" Feb 27 08:57:45 crc kubenswrapper[4612]: I0227 08:57:45.916357 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="0ea808c4-f9ac-4d97-9ca6-d93728854104" containerName="registry-server" Feb 27 08:57:45 crc kubenswrapper[4612]: E0227 08:57:45.916369 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0ea808c4-f9ac-4d97-9ca6-d93728854104" containerName="extract-content" Feb 27 08:57:45 crc kubenswrapper[4612]: I0227 08:57:45.916376 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="0ea808c4-f9ac-4d97-9ca6-d93728854104" containerName="extract-content" Feb 27 08:57:45 crc kubenswrapper[4612]: E0227 08:57:45.916433 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0ea808c4-f9ac-4d97-9ca6-d93728854104" containerName="extract-utilities" Feb 27 08:57:45 crc kubenswrapper[4612]: I0227 08:57:45.916445 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="0ea808c4-f9ac-4d97-9ca6-d93728854104" containerName="extract-utilities" Feb 27 08:57:45 crc kubenswrapper[4612]: I0227 08:57:45.916645 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="0ea808c4-f9ac-4d97-9ca6-d93728854104" containerName="registry-server" Feb 27 08:57:45 crc kubenswrapper[4612]: I0227 08:57:45.917837 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mchzb/must-gather-t4rbr" Feb 27 08:57:45 crc kubenswrapper[4612]: I0227 08:57:45.928983 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-mchzb"/"default-dockercfg-47x2k" Feb 27 08:57:45 crc kubenswrapper[4612]: I0227 08:57:45.930787 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-mchzb"/"openshift-service-ca.crt" Feb 27 08:57:45 crc kubenswrapper[4612]: I0227 08:57:45.931012 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-mchzb"/"kube-root-ca.crt" Feb 27 08:57:45 crc kubenswrapper[4612]: I0227 08:57:45.966156 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-mchzb/must-gather-t4rbr"] Feb 27 08:57:45 crc kubenswrapper[4612]: I0227 08:57:45.990121 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nvrfq\" (UniqueName: \"kubernetes.io/projected/3f06074f-6e22-4526-a04b-e842577c6f2e-kube-api-access-nvrfq\") pod \"must-gather-t4rbr\" (UID: \"3f06074f-6e22-4526-a04b-e842577c6f2e\") " pod="openshift-must-gather-mchzb/must-gather-t4rbr" Feb 27 08:57:45 crc kubenswrapper[4612]: I0227 08:57:45.990176 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/3f06074f-6e22-4526-a04b-e842577c6f2e-must-gather-output\") pod \"must-gather-t4rbr\" (UID: \"3f06074f-6e22-4526-a04b-e842577c6f2e\") " pod="openshift-must-gather-mchzb/must-gather-t4rbr" Feb 27 08:57:46 crc kubenswrapper[4612]: I0227 08:57:46.030309 4612 patch_prober.go:28] interesting pod/machine-config-daemon-924vb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 27 08:57:46 crc kubenswrapper[4612]: I0227 08:57:46.030440 4612 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 27 08:57:46 crc kubenswrapper[4612]: I0227 08:57:46.091562 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/3f06074f-6e22-4526-a04b-e842577c6f2e-must-gather-output\") pod \"must-gather-t4rbr\" (UID: \"3f06074f-6e22-4526-a04b-e842577c6f2e\") " pod="openshift-must-gather-mchzb/must-gather-t4rbr" Feb 27 08:57:46 crc kubenswrapper[4612]: I0227 08:57:46.091787 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nvrfq\" (UniqueName: \"kubernetes.io/projected/3f06074f-6e22-4526-a04b-e842577c6f2e-kube-api-access-nvrfq\") pod \"must-gather-t4rbr\" (UID: \"3f06074f-6e22-4526-a04b-e842577c6f2e\") " pod="openshift-must-gather-mchzb/must-gather-t4rbr" Feb 27 08:57:46 crc kubenswrapper[4612]: I0227 08:57:46.092521 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/3f06074f-6e22-4526-a04b-e842577c6f2e-must-gather-output\") pod \"must-gather-t4rbr\" (UID: \"3f06074f-6e22-4526-a04b-e842577c6f2e\") " pod="openshift-must-gather-mchzb/must-gather-t4rbr" Feb 27 08:57:46 crc kubenswrapper[4612]: I0227 08:57:46.117461 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nvrfq\" (UniqueName: \"kubernetes.io/projected/3f06074f-6e22-4526-a04b-e842577c6f2e-kube-api-access-nvrfq\") pod \"must-gather-t4rbr\" (UID: \"3f06074f-6e22-4526-a04b-e842577c6f2e\") " pod="openshift-must-gather-mchzb/must-gather-t4rbr" Feb 27 08:57:46 crc kubenswrapper[4612]: I0227 08:57:46.232890 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mchzb/must-gather-t4rbr" Feb 27 08:57:46 crc kubenswrapper[4612]: I0227 08:57:46.756975 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-mchzb/must-gather-t4rbr"] Feb 27 08:57:47 crc kubenswrapper[4612]: W0227 08:57:47.084413 4612 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3f06074f_6e22_4526_a04b_e842577c6f2e.slice/crio-d71c8e79a9b4ccee2b3969ef383b100ebd4867bde8c5351978c97801259bb5cd WatchSource:0}: Error finding container d71c8e79a9b4ccee2b3969ef383b100ebd4867bde8c5351978c97801259bb5cd: Status 404 returned error can't find the container with id d71c8e79a9b4ccee2b3969ef383b100ebd4867bde8c5351978c97801259bb5cd Feb 27 08:57:47 crc kubenswrapper[4612]: I0227 08:57:47.322637 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mchzb/must-gather-t4rbr" event={"ID":"3f06074f-6e22-4526-a04b-e842577c6f2e","Type":"ContainerStarted","Data":"d71c8e79a9b4ccee2b3969ef383b100ebd4867bde8c5351978c97801259bb5cd"} Feb 27 08:57:56 crc kubenswrapper[4612]: I0227 08:57:56.438008 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mchzb/must-gather-t4rbr" event={"ID":"3f06074f-6e22-4526-a04b-e842577c6f2e","Type":"ContainerStarted","Data":"b9508d72b81df0d55578ad38e42b9f06bdb5f38cb474a26239d791fdd4d9848e"} Feb 27 08:57:56 crc kubenswrapper[4612]: I0227 08:57:56.438591 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mchzb/must-gather-t4rbr" event={"ID":"3f06074f-6e22-4526-a04b-e842577c6f2e","Type":"ContainerStarted","Data":"e6328b790dfbc8f52b3131fc81f90e6ddef8bfc38dc9e78acb8c6353c4afc402"} Feb 27 08:57:56 crc kubenswrapper[4612]: I0227 08:57:56.462267 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-mchzb/must-gather-t4rbr" podStartSLOduration=3.306703538 podStartE2EDuration="11.46223948s" podCreationTimestamp="2026-02-27 08:57:45 +0000 UTC" firstStartedPulling="2026-02-27 08:57:47.086946394 +0000 UTC m=+4124.940876392" lastFinishedPulling="2026-02-27 08:57:55.242482296 +0000 UTC m=+4133.096412334" observedRunningTime="2026-02-27 08:57:56.459255925 +0000 UTC m=+4134.313186003" watchObservedRunningTime="2026-02-27 08:57:56.46223948 +0000 UTC m=+4134.316169508" Feb 27 08:57:57 crc kubenswrapper[4612]: I0227 08:57:57.528951 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-pnrpr"] Feb 27 08:57:57 crc kubenswrapper[4612]: I0227 08:57:57.531577 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pnrpr" Feb 27 08:57:57 crc kubenswrapper[4612]: I0227 08:57:57.543646 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-pnrpr"] Feb 27 08:57:57 crc kubenswrapper[4612]: I0227 08:57:57.631737 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/20a06f2f-cc9e-460f-93a2-d40186958b17-utilities\") pod \"redhat-operators-pnrpr\" (UID: \"20a06f2f-cc9e-460f-93a2-d40186958b17\") " pod="openshift-marketplace/redhat-operators-pnrpr" Feb 27 08:57:57 crc kubenswrapper[4612]: I0227 08:57:57.631807 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/20a06f2f-cc9e-460f-93a2-d40186958b17-catalog-content\") pod \"redhat-operators-pnrpr\" (UID: \"20a06f2f-cc9e-460f-93a2-d40186958b17\") " pod="openshift-marketplace/redhat-operators-pnrpr" Feb 27 08:57:57 crc kubenswrapper[4612]: I0227 08:57:57.631944 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5vg7t\" (UniqueName: \"kubernetes.io/projected/20a06f2f-cc9e-460f-93a2-d40186958b17-kube-api-access-5vg7t\") pod \"redhat-operators-pnrpr\" (UID: \"20a06f2f-cc9e-460f-93a2-d40186958b17\") " pod="openshift-marketplace/redhat-operators-pnrpr" Feb 27 08:57:57 crc kubenswrapper[4612]: I0227 08:57:57.733619 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/20a06f2f-cc9e-460f-93a2-d40186958b17-utilities\") pod \"redhat-operators-pnrpr\" (UID: \"20a06f2f-cc9e-460f-93a2-d40186958b17\") " pod="openshift-marketplace/redhat-operators-pnrpr" Feb 27 08:57:57 crc kubenswrapper[4612]: I0227 08:57:57.733675 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/20a06f2f-cc9e-460f-93a2-d40186958b17-catalog-content\") pod \"redhat-operators-pnrpr\" (UID: \"20a06f2f-cc9e-460f-93a2-d40186958b17\") " pod="openshift-marketplace/redhat-operators-pnrpr" Feb 27 08:57:57 crc kubenswrapper[4612]: I0227 08:57:57.733746 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5vg7t\" (UniqueName: \"kubernetes.io/projected/20a06f2f-cc9e-460f-93a2-d40186958b17-kube-api-access-5vg7t\") pod \"redhat-operators-pnrpr\" (UID: \"20a06f2f-cc9e-460f-93a2-d40186958b17\") " pod="openshift-marketplace/redhat-operators-pnrpr" Feb 27 08:57:57 crc kubenswrapper[4612]: I0227 08:57:57.734423 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/20a06f2f-cc9e-460f-93a2-d40186958b17-utilities\") pod \"redhat-operators-pnrpr\" (UID: \"20a06f2f-cc9e-460f-93a2-d40186958b17\") " pod="openshift-marketplace/redhat-operators-pnrpr" Feb 27 08:57:57 crc kubenswrapper[4612]: I0227 08:57:57.734628 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/20a06f2f-cc9e-460f-93a2-d40186958b17-catalog-content\") pod \"redhat-operators-pnrpr\" (UID: \"20a06f2f-cc9e-460f-93a2-d40186958b17\") " pod="openshift-marketplace/redhat-operators-pnrpr" Feb 27 08:57:57 crc kubenswrapper[4612]: I0227 08:57:57.755400 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5vg7t\" (UniqueName: \"kubernetes.io/projected/20a06f2f-cc9e-460f-93a2-d40186958b17-kube-api-access-5vg7t\") pod \"redhat-operators-pnrpr\" (UID: \"20a06f2f-cc9e-460f-93a2-d40186958b17\") " pod="openshift-marketplace/redhat-operators-pnrpr" Feb 27 08:57:57 crc kubenswrapper[4612]: I0227 08:57:57.857858 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pnrpr" Feb 27 08:57:58 crc kubenswrapper[4612]: I0227 08:57:58.833452 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-pnrpr"] Feb 27 08:57:59 crc kubenswrapper[4612]: I0227 08:57:59.494999 4612 generic.go:334] "Generic (PLEG): container finished" podID="20a06f2f-cc9e-460f-93a2-d40186958b17" containerID="98f5549ad40c65600fd0ca072101fe2f6f7f518d058f2e493d58457c7bbb87ad" exitCode=0 Feb 27 08:57:59 crc kubenswrapper[4612]: I0227 08:57:59.495328 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pnrpr" event={"ID":"20a06f2f-cc9e-460f-93a2-d40186958b17","Type":"ContainerDied","Data":"98f5549ad40c65600fd0ca072101fe2f6f7f518d058f2e493d58457c7bbb87ad"} Feb 27 08:57:59 crc kubenswrapper[4612]: I0227 08:57:59.495361 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pnrpr" event={"ID":"20a06f2f-cc9e-460f-93a2-d40186958b17","Type":"ContainerStarted","Data":"406f6b525d135a8545647ad57d2bf809358836c0c0df1cfad995b2c00dc035b5"} Feb 27 08:58:00 crc kubenswrapper[4612]: I0227 08:58:00.163208 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29536378-k6phd"] Feb 27 08:58:00 crc kubenswrapper[4612]: I0227 08:58:00.164663 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536378-k6phd" Feb 27 08:58:00 crc kubenswrapper[4612]: I0227 08:58:00.169130 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-5zwfx" Feb 27 08:58:00 crc kubenswrapper[4612]: I0227 08:58:00.169449 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 27 08:58:00 crc kubenswrapper[4612]: I0227 08:58:00.170017 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 27 08:58:00 crc kubenswrapper[4612]: I0227 08:58:00.187753 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536378-k6phd"] Feb 27 08:58:00 crc kubenswrapper[4612]: I0227 08:58:00.293987 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vcw7x\" (UniqueName: \"kubernetes.io/projected/c50f2ffa-23aa-482f-87ee-d0d700753b07-kube-api-access-vcw7x\") pod \"auto-csr-approver-29536378-k6phd\" (UID: \"c50f2ffa-23aa-482f-87ee-d0d700753b07\") " pod="openshift-infra/auto-csr-approver-29536378-k6phd" Feb 27 08:58:00 crc kubenswrapper[4612]: I0227 08:58:00.395839 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vcw7x\" (UniqueName: \"kubernetes.io/projected/c50f2ffa-23aa-482f-87ee-d0d700753b07-kube-api-access-vcw7x\") pod \"auto-csr-approver-29536378-k6phd\" (UID: \"c50f2ffa-23aa-482f-87ee-d0d700753b07\") " pod="openshift-infra/auto-csr-approver-29536378-k6phd" Feb 27 08:58:00 crc kubenswrapper[4612]: I0227 08:58:00.420506 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vcw7x\" (UniqueName: \"kubernetes.io/projected/c50f2ffa-23aa-482f-87ee-d0d700753b07-kube-api-access-vcw7x\") pod \"auto-csr-approver-29536378-k6phd\" (UID: \"c50f2ffa-23aa-482f-87ee-d0d700753b07\") " pod="openshift-infra/auto-csr-approver-29536378-k6phd" Feb 27 08:58:00 crc kubenswrapper[4612]: I0227 08:58:00.505180 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pnrpr" event={"ID":"20a06f2f-cc9e-460f-93a2-d40186958b17","Type":"ContainerStarted","Data":"d39e1e208b749c107d0156055aea1fb77deebb69104511981ee9602d176db66e"} Feb 27 08:58:00 crc kubenswrapper[4612]: I0227 08:58:00.533781 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536378-k6phd" Feb 27 08:58:01 crc kubenswrapper[4612]: I0227 08:58:01.049342 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536378-k6phd"] Feb 27 08:58:01 crc kubenswrapper[4612]: W0227 08:58:01.057293 4612 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc50f2ffa_23aa_482f_87ee_d0d700753b07.slice/crio-6fbcf638792dc358f2a155b06a7888fe6bbd1a0bfe6eac0d896a0f5dec0541dd WatchSource:0}: Error finding container 6fbcf638792dc358f2a155b06a7888fe6bbd1a0bfe6eac0d896a0f5dec0541dd: Status 404 returned error can't find the container with id 6fbcf638792dc358f2a155b06a7888fe6bbd1a0bfe6eac0d896a0f5dec0541dd Feb 27 08:58:01 crc kubenswrapper[4612]: I0227 08:58:01.534951 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536378-k6phd" event={"ID":"c50f2ffa-23aa-482f-87ee-d0d700753b07","Type":"ContainerStarted","Data":"6fbcf638792dc358f2a155b06a7888fe6bbd1a0bfe6eac0d896a0f5dec0541dd"} Feb 27 08:58:02 crc kubenswrapper[4612]: I0227 08:58:02.545390 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536378-k6phd" event={"ID":"c50f2ffa-23aa-482f-87ee-d0d700753b07","Type":"ContainerStarted","Data":"65beaa35fb57fae1bd4f6c00722d4c06b4b0d5c2ede32f9b100d41e9a0fb3621"} Feb 27 08:58:02 crc kubenswrapper[4612]: I0227 08:58:02.567560 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29536378-k6phd" podStartSLOduration=1.6533761120000001 podStartE2EDuration="2.567540435s" podCreationTimestamp="2026-02-27 08:58:00 +0000 UTC" firstStartedPulling="2026-02-27 08:58:01.05923378 +0000 UTC m=+4138.913163778" lastFinishedPulling="2026-02-27 08:58:01.973398103 +0000 UTC m=+4139.827328101" observedRunningTime="2026-02-27 08:58:02.562957545 +0000 UTC m=+4140.416887543" watchObservedRunningTime="2026-02-27 08:58:02.567540435 +0000 UTC m=+4140.421470433" Feb 27 08:58:02 crc kubenswrapper[4612]: I0227 08:58:02.893491 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-mchzb/crc-debug-llhsn"] Feb 27 08:58:02 crc kubenswrapper[4612]: I0227 08:58:02.896139 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mchzb/crc-debug-llhsn" Feb 27 08:58:03 crc kubenswrapper[4612]: I0227 08:58:03.043932 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ce6584ee-7f03-46a1-a30e-741b7f3c08d2-host\") pod \"crc-debug-llhsn\" (UID: \"ce6584ee-7f03-46a1-a30e-741b7f3c08d2\") " pod="openshift-must-gather-mchzb/crc-debug-llhsn" Feb 27 08:58:03 crc kubenswrapper[4612]: I0227 08:58:03.044015 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mdlh5\" (UniqueName: \"kubernetes.io/projected/ce6584ee-7f03-46a1-a30e-741b7f3c08d2-kube-api-access-mdlh5\") pod \"crc-debug-llhsn\" (UID: \"ce6584ee-7f03-46a1-a30e-741b7f3c08d2\") " pod="openshift-must-gather-mchzb/crc-debug-llhsn" Feb 27 08:58:03 crc kubenswrapper[4612]: I0227 08:58:03.146092 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ce6584ee-7f03-46a1-a30e-741b7f3c08d2-host\") pod \"crc-debug-llhsn\" (UID: \"ce6584ee-7f03-46a1-a30e-741b7f3c08d2\") " pod="openshift-must-gather-mchzb/crc-debug-llhsn" Feb 27 08:58:03 crc kubenswrapper[4612]: I0227 08:58:03.146159 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mdlh5\" (UniqueName: \"kubernetes.io/projected/ce6584ee-7f03-46a1-a30e-741b7f3c08d2-kube-api-access-mdlh5\") pod \"crc-debug-llhsn\" (UID: \"ce6584ee-7f03-46a1-a30e-741b7f3c08d2\") " pod="openshift-must-gather-mchzb/crc-debug-llhsn" Feb 27 08:58:03 crc kubenswrapper[4612]: I0227 08:58:03.146219 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ce6584ee-7f03-46a1-a30e-741b7f3c08d2-host\") pod \"crc-debug-llhsn\" (UID: \"ce6584ee-7f03-46a1-a30e-741b7f3c08d2\") " pod="openshift-must-gather-mchzb/crc-debug-llhsn" Feb 27 08:58:03 crc kubenswrapper[4612]: I0227 08:58:03.173756 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mdlh5\" (UniqueName: \"kubernetes.io/projected/ce6584ee-7f03-46a1-a30e-741b7f3c08d2-kube-api-access-mdlh5\") pod \"crc-debug-llhsn\" (UID: \"ce6584ee-7f03-46a1-a30e-741b7f3c08d2\") " pod="openshift-must-gather-mchzb/crc-debug-llhsn" Feb 27 08:58:03 crc kubenswrapper[4612]: I0227 08:58:03.222784 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mchzb/crc-debug-llhsn" Feb 27 08:58:03 crc kubenswrapper[4612]: W0227 08:58:03.278250 4612 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podce6584ee_7f03_46a1_a30e_741b7f3c08d2.slice/crio-ac3d34b5e8a6b6a86280161484f48906ffcf0ee116673e121c3038d82ebbc88d WatchSource:0}: Error finding container ac3d34b5e8a6b6a86280161484f48906ffcf0ee116673e121c3038d82ebbc88d: Status 404 returned error can't find the container with id ac3d34b5e8a6b6a86280161484f48906ffcf0ee116673e121c3038d82ebbc88d Feb 27 08:58:03 crc kubenswrapper[4612]: I0227 08:58:03.557965 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mchzb/crc-debug-llhsn" event={"ID":"ce6584ee-7f03-46a1-a30e-741b7f3c08d2","Type":"ContainerStarted","Data":"ac3d34b5e8a6b6a86280161484f48906ffcf0ee116673e121c3038d82ebbc88d"} Feb 27 08:58:03 crc kubenswrapper[4612]: I0227 08:58:03.561084 4612 generic.go:334] "Generic (PLEG): container finished" podID="c50f2ffa-23aa-482f-87ee-d0d700753b07" containerID="65beaa35fb57fae1bd4f6c00722d4c06b4b0d5c2ede32f9b100d41e9a0fb3621" exitCode=0 Feb 27 08:58:03 crc kubenswrapper[4612]: I0227 08:58:03.561140 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536378-k6phd" event={"ID":"c50f2ffa-23aa-482f-87ee-d0d700753b07","Type":"ContainerDied","Data":"65beaa35fb57fae1bd4f6c00722d4c06b4b0d5c2ede32f9b100d41e9a0fb3621"} Feb 27 08:58:05 crc kubenswrapper[4612]: I0227 08:58:05.507167 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536378-k6phd" Feb 27 08:58:05 crc kubenswrapper[4612]: I0227 08:58:05.588136 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536378-k6phd" event={"ID":"c50f2ffa-23aa-482f-87ee-d0d700753b07","Type":"ContainerDied","Data":"6fbcf638792dc358f2a155b06a7888fe6bbd1a0bfe6eac0d896a0f5dec0541dd"} Feb 27 08:58:05 crc kubenswrapper[4612]: I0227 08:58:05.588427 4612 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6fbcf638792dc358f2a155b06a7888fe6bbd1a0bfe6eac0d896a0f5dec0541dd" Feb 27 08:58:05 crc kubenswrapper[4612]: I0227 08:58:05.588198 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536378-k6phd" Feb 27 08:58:05 crc kubenswrapper[4612]: I0227 08:58:05.592043 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vcw7x\" (UniqueName: \"kubernetes.io/projected/c50f2ffa-23aa-482f-87ee-d0d700753b07-kube-api-access-vcw7x\") pod \"c50f2ffa-23aa-482f-87ee-d0d700753b07\" (UID: \"c50f2ffa-23aa-482f-87ee-d0d700753b07\") " Feb 27 08:58:05 crc kubenswrapper[4612]: I0227 08:58:05.624899 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c50f2ffa-23aa-482f-87ee-d0d700753b07-kube-api-access-vcw7x" (OuterVolumeSpecName: "kube-api-access-vcw7x") pod "c50f2ffa-23aa-482f-87ee-d0d700753b07" (UID: "c50f2ffa-23aa-482f-87ee-d0d700753b07"). InnerVolumeSpecName "kube-api-access-vcw7x". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 08:58:05 crc kubenswrapper[4612]: I0227 08:58:05.672819 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29536372-dhs6j"] Feb 27 08:58:05 crc kubenswrapper[4612]: I0227 08:58:05.682096 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29536372-dhs6j"] Feb 27 08:58:05 crc kubenswrapper[4612]: I0227 08:58:05.694121 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vcw7x\" (UniqueName: \"kubernetes.io/projected/c50f2ffa-23aa-482f-87ee-d0d700753b07-kube-api-access-vcw7x\") on node \"crc\" DevicePath \"\"" Feb 27 08:58:06 crc kubenswrapper[4612]: I0227 08:58:06.866736 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e64df7cb-2f18-4fb7-b66d-86c993b1c8b2" path="/var/lib/kubelet/pods/e64df7cb-2f18-4fb7-b66d-86c993b1c8b2/volumes" Feb 27 08:58:09 crc kubenswrapper[4612]: I0227 08:58:09.631355 4612 generic.go:334] "Generic (PLEG): container finished" podID="20a06f2f-cc9e-460f-93a2-d40186958b17" containerID="d39e1e208b749c107d0156055aea1fb77deebb69104511981ee9602d176db66e" exitCode=0 Feb 27 08:58:09 crc kubenswrapper[4612]: I0227 08:58:09.631913 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pnrpr" event={"ID":"20a06f2f-cc9e-460f-93a2-d40186958b17","Type":"ContainerDied","Data":"d39e1e208b749c107d0156055aea1fb77deebb69104511981ee9602d176db66e"} Feb 27 08:58:16 crc kubenswrapper[4612]: I0227 08:58:16.027013 4612 patch_prober.go:28] interesting pod/machine-config-daemon-924vb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 27 08:58:16 crc kubenswrapper[4612]: I0227 08:58:16.027507 4612 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 27 08:58:16 crc kubenswrapper[4612]: I0227 08:58:16.027554 4612 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-924vb" Feb 27 08:58:16 crc kubenswrapper[4612]: I0227 08:58:16.028299 4612 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"cebeb2cb354fa601f37e4c7cc40bcc1dcd010ada59d86d91448073fa6deac102"} pod="openshift-machine-config-operator/machine-config-daemon-924vb" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 27 08:58:16 crc kubenswrapper[4612]: I0227 08:58:16.028350 4612 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" containerName="machine-config-daemon" containerID="cri-o://cebeb2cb354fa601f37e4c7cc40bcc1dcd010ada59d86d91448073fa6deac102" gracePeriod=600 Feb 27 08:58:16 crc kubenswrapper[4612]: I0227 08:58:16.702647 4612 generic.go:334] "Generic (PLEG): container finished" podID="24599373-7adc-4a1b-8bb4-797bf726f43d" containerID="cebeb2cb354fa601f37e4c7cc40bcc1dcd010ada59d86d91448073fa6deac102" exitCode=0 Feb 27 08:58:16 crc kubenswrapper[4612]: I0227 08:58:16.702716 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-924vb" event={"ID":"24599373-7adc-4a1b-8bb4-797bf726f43d","Type":"ContainerDied","Data":"cebeb2cb354fa601f37e4c7cc40bcc1dcd010ada59d86d91448073fa6deac102"} Feb 27 08:58:16 crc kubenswrapper[4612]: I0227 08:58:16.703303 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-924vb" event={"ID":"24599373-7adc-4a1b-8bb4-797bf726f43d","Type":"ContainerStarted","Data":"aa498fa4b5f3843852a1ebe0b165b06254525eeb49f5224c7bd2a0f8c2ae2480"} Feb 27 08:58:16 crc kubenswrapper[4612]: I0227 08:58:16.703365 4612 scope.go:117] "RemoveContainer" containerID="2ed59616da422731358981fd9cdfe7932267d24e60d616b16a3481f816f614da" Feb 27 08:58:16 crc kubenswrapper[4612]: I0227 08:58:16.709131 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pnrpr" event={"ID":"20a06f2f-cc9e-460f-93a2-d40186958b17","Type":"ContainerStarted","Data":"c25cd841b66fb17c777d42668eda205250ad25d4b7ab2c8a57016678cc70e7c6"} Feb 27 08:58:16 crc kubenswrapper[4612]: I0227 08:58:16.713871 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mchzb/crc-debug-llhsn" event={"ID":"ce6584ee-7f03-46a1-a30e-741b7f3c08d2","Type":"ContainerStarted","Data":"9c008fe33903e854c28f4859f9be27c7651a865e82c16ed2fe42d44c29243c67"} Feb 27 08:58:16 crc kubenswrapper[4612]: I0227 08:58:16.746226 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-pnrpr" podStartSLOduration=3.5091684819999998 podStartE2EDuration="19.746179936s" podCreationTimestamp="2026-02-27 08:57:57 +0000 UTC" firstStartedPulling="2026-02-27 08:57:59.497465747 +0000 UTC m=+4137.351395745" lastFinishedPulling="2026-02-27 08:58:15.734477201 +0000 UTC m=+4153.588407199" observedRunningTime="2026-02-27 08:58:16.743709436 +0000 UTC m=+4154.597639434" watchObservedRunningTime="2026-02-27 08:58:16.746179936 +0000 UTC m=+4154.600109954" Feb 27 08:58:16 crc kubenswrapper[4612]: I0227 08:58:16.767087 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-mchzb/crc-debug-llhsn" podStartSLOduration=2.297576415 podStartE2EDuration="14.76707109s" podCreationTimestamp="2026-02-27 08:58:02 +0000 UTC" firstStartedPulling="2026-02-27 08:58:03.287910748 +0000 UTC m=+4141.141840746" lastFinishedPulling="2026-02-27 08:58:15.757405423 +0000 UTC m=+4153.611335421" observedRunningTime="2026-02-27 08:58:16.756486849 +0000 UTC m=+4154.610416847" watchObservedRunningTime="2026-02-27 08:58:16.76707109 +0000 UTC m=+4154.621001088" Feb 27 08:58:17 crc kubenswrapper[4612]: I0227 08:58:17.858165 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-pnrpr" Feb 27 08:58:17 crc kubenswrapper[4612]: I0227 08:58:17.858501 4612 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-pnrpr" Feb 27 08:58:18 crc kubenswrapper[4612]: I0227 08:58:18.916816 4612 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-pnrpr" podUID="20a06f2f-cc9e-460f-93a2-d40186958b17" containerName="registry-server" probeResult="failure" output=< Feb 27 08:58:18 crc kubenswrapper[4612]: timeout: failed to connect service ":50051" within 1s Feb 27 08:58:18 crc kubenswrapper[4612]: > Feb 27 08:58:19 crc kubenswrapper[4612]: I0227 08:58:19.755071 4612 scope.go:117] "RemoveContainer" containerID="0735a1f078fef42c96e74e75d94933547db722ce87140ee78527e7e495b48895" Feb 27 08:58:28 crc kubenswrapper[4612]: I0227 08:58:28.907513 4612 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-pnrpr" podUID="20a06f2f-cc9e-460f-93a2-d40186958b17" containerName="registry-server" probeResult="failure" output=< Feb 27 08:58:28 crc kubenswrapper[4612]: timeout: failed to connect service ":50051" within 1s Feb 27 08:58:28 crc kubenswrapper[4612]: > Feb 27 08:58:38 crc kubenswrapper[4612]: I0227 08:58:38.907597 4612 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-pnrpr" podUID="20a06f2f-cc9e-460f-93a2-d40186958b17" containerName="registry-server" probeResult="failure" output=< Feb 27 08:58:38 crc kubenswrapper[4612]: timeout: failed to connect service ":50051" within 1s Feb 27 08:58:38 crc kubenswrapper[4612]: > Feb 27 08:58:48 crc kubenswrapper[4612]: I0227 08:58:48.901950 4612 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-pnrpr" podUID="20a06f2f-cc9e-460f-93a2-d40186958b17" containerName="registry-server" probeResult="failure" output=< Feb 27 08:58:48 crc kubenswrapper[4612]: timeout: failed to connect service ":50051" within 1s Feb 27 08:58:48 crc kubenswrapper[4612]: > Feb 27 08:58:58 crc kubenswrapper[4612]: I0227 08:58:58.922510 4612 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-pnrpr" podUID="20a06f2f-cc9e-460f-93a2-d40186958b17" containerName="registry-server" probeResult="failure" output=< Feb 27 08:58:58 crc kubenswrapper[4612]: timeout: failed to connect service ":50051" within 1s Feb 27 08:58:58 crc kubenswrapper[4612]: > Feb 27 08:59:07 crc kubenswrapper[4612]: I0227 08:59:07.927973 4612 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-pnrpr" Feb 27 08:59:08 crc kubenswrapper[4612]: I0227 08:59:07.999474 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-pnrpr" Feb 27 08:59:08 crc kubenswrapper[4612]: I0227 08:59:08.176284 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-pnrpr"] Feb 27 08:59:08 crc kubenswrapper[4612]: I0227 08:59:08.632437 4612 generic.go:334] "Generic (PLEG): container finished" podID="ce6584ee-7f03-46a1-a30e-741b7f3c08d2" containerID="9c008fe33903e854c28f4859f9be27c7651a865e82c16ed2fe42d44c29243c67" exitCode=0 Feb 27 08:59:08 crc kubenswrapper[4612]: I0227 08:59:08.632498 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mchzb/crc-debug-llhsn" event={"ID":"ce6584ee-7f03-46a1-a30e-741b7f3c08d2","Type":"ContainerDied","Data":"9c008fe33903e854c28f4859f9be27c7651a865e82c16ed2fe42d44c29243c67"} Feb 27 08:59:09 crc kubenswrapper[4612]: I0227 08:59:09.640446 4612 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-pnrpr" podUID="20a06f2f-cc9e-460f-93a2-d40186958b17" containerName="registry-server" containerID="cri-o://c25cd841b66fb17c777d42668eda205250ad25d4b7ab2c8a57016678cc70e7c6" gracePeriod=2 Feb 27 08:59:09 crc kubenswrapper[4612]: I0227 08:59:09.746997 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mchzb/crc-debug-llhsn" Feb 27 08:59:09 crc kubenswrapper[4612]: I0227 08:59:09.810180 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-mchzb/crc-debug-llhsn"] Feb 27 08:59:09 crc kubenswrapper[4612]: I0227 08:59:09.848350 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-mchzb/crc-debug-llhsn"] Feb 27 08:59:09 crc kubenswrapper[4612]: I0227 08:59:09.869801 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ce6584ee-7f03-46a1-a30e-741b7f3c08d2-host\") pod \"ce6584ee-7f03-46a1-a30e-741b7f3c08d2\" (UID: \"ce6584ee-7f03-46a1-a30e-741b7f3c08d2\") " Feb 27 08:59:09 crc kubenswrapper[4612]: I0227 08:59:09.869910 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ce6584ee-7f03-46a1-a30e-741b7f3c08d2-host" (OuterVolumeSpecName: "host") pod "ce6584ee-7f03-46a1-a30e-741b7f3c08d2" (UID: "ce6584ee-7f03-46a1-a30e-741b7f3c08d2"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 27 08:59:09 crc kubenswrapper[4612]: I0227 08:59:09.870057 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mdlh5\" (UniqueName: \"kubernetes.io/projected/ce6584ee-7f03-46a1-a30e-741b7f3c08d2-kube-api-access-mdlh5\") pod \"ce6584ee-7f03-46a1-a30e-741b7f3c08d2\" (UID: \"ce6584ee-7f03-46a1-a30e-741b7f3c08d2\") " Feb 27 08:59:09 crc kubenswrapper[4612]: I0227 08:59:09.870984 4612 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ce6584ee-7f03-46a1-a30e-741b7f3c08d2-host\") on node \"crc\" DevicePath \"\"" Feb 27 08:59:09 crc kubenswrapper[4612]: I0227 08:59:09.888493 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ce6584ee-7f03-46a1-a30e-741b7f3c08d2-kube-api-access-mdlh5" (OuterVolumeSpecName: "kube-api-access-mdlh5") pod "ce6584ee-7f03-46a1-a30e-741b7f3c08d2" (UID: "ce6584ee-7f03-46a1-a30e-741b7f3c08d2"). InnerVolumeSpecName "kube-api-access-mdlh5". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 08:59:09 crc kubenswrapper[4612]: I0227 08:59:09.973229 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mdlh5\" (UniqueName: \"kubernetes.io/projected/ce6584ee-7f03-46a1-a30e-741b7f3c08d2-kube-api-access-mdlh5\") on node \"crc\" DevicePath \"\"" Feb 27 08:59:10 crc kubenswrapper[4612]: I0227 08:59:10.463283 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pnrpr" Feb 27 08:59:10 crc kubenswrapper[4612]: I0227 08:59:10.584850 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/20a06f2f-cc9e-460f-93a2-d40186958b17-utilities\") pod \"20a06f2f-cc9e-460f-93a2-d40186958b17\" (UID: \"20a06f2f-cc9e-460f-93a2-d40186958b17\") " Feb 27 08:59:10 crc kubenswrapper[4612]: I0227 08:59:10.585204 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/20a06f2f-cc9e-460f-93a2-d40186958b17-catalog-content\") pod \"20a06f2f-cc9e-460f-93a2-d40186958b17\" (UID: \"20a06f2f-cc9e-460f-93a2-d40186958b17\") " Feb 27 08:59:10 crc kubenswrapper[4612]: I0227 08:59:10.585280 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5vg7t\" (UniqueName: \"kubernetes.io/projected/20a06f2f-cc9e-460f-93a2-d40186958b17-kube-api-access-5vg7t\") pod \"20a06f2f-cc9e-460f-93a2-d40186958b17\" (UID: \"20a06f2f-cc9e-460f-93a2-d40186958b17\") " Feb 27 08:59:10 crc kubenswrapper[4612]: I0227 08:59:10.585500 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/20a06f2f-cc9e-460f-93a2-d40186958b17-utilities" (OuterVolumeSpecName: "utilities") pod "20a06f2f-cc9e-460f-93a2-d40186958b17" (UID: "20a06f2f-cc9e-460f-93a2-d40186958b17"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 08:59:10 crc kubenswrapper[4612]: I0227 08:59:10.586139 4612 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/20a06f2f-cc9e-460f-93a2-d40186958b17-utilities\") on node \"crc\" DevicePath \"\"" Feb 27 08:59:10 crc kubenswrapper[4612]: I0227 08:59:10.590172 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20a06f2f-cc9e-460f-93a2-d40186958b17-kube-api-access-5vg7t" (OuterVolumeSpecName: "kube-api-access-5vg7t") pod "20a06f2f-cc9e-460f-93a2-d40186958b17" (UID: "20a06f2f-cc9e-460f-93a2-d40186958b17"). InnerVolumeSpecName "kube-api-access-5vg7t". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 08:59:10 crc kubenswrapper[4612]: I0227 08:59:10.649258 4612 generic.go:334] "Generic (PLEG): container finished" podID="20a06f2f-cc9e-460f-93a2-d40186958b17" containerID="c25cd841b66fb17c777d42668eda205250ad25d4b7ab2c8a57016678cc70e7c6" exitCode=0 Feb 27 08:59:10 crc kubenswrapper[4612]: I0227 08:59:10.649335 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pnrpr" event={"ID":"20a06f2f-cc9e-460f-93a2-d40186958b17","Type":"ContainerDied","Data":"c25cd841b66fb17c777d42668eda205250ad25d4b7ab2c8a57016678cc70e7c6"} Feb 27 08:59:10 crc kubenswrapper[4612]: I0227 08:59:10.649366 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pnrpr" event={"ID":"20a06f2f-cc9e-460f-93a2-d40186958b17","Type":"ContainerDied","Data":"406f6b525d135a8545647ad57d2bf809358836c0c0df1cfad995b2c00dc035b5"} Feb 27 08:59:10 crc kubenswrapper[4612]: I0227 08:59:10.649382 4612 scope.go:117] "RemoveContainer" containerID="c25cd841b66fb17c777d42668eda205250ad25d4b7ab2c8a57016678cc70e7c6" Feb 27 08:59:10 crc kubenswrapper[4612]: I0227 08:59:10.649498 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pnrpr" Feb 27 08:59:10 crc kubenswrapper[4612]: I0227 08:59:10.652916 4612 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ac3d34b5e8a6b6a86280161484f48906ffcf0ee116673e121c3038d82ebbc88d" Feb 27 08:59:10 crc kubenswrapper[4612]: I0227 08:59:10.652971 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mchzb/crc-debug-llhsn" Feb 27 08:59:10 crc kubenswrapper[4612]: I0227 08:59:10.687926 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5vg7t\" (UniqueName: \"kubernetes.io/projected/20a06f2f-cc9e-460f-93a2-d40186958b17-kube-api-access-5vg7t\") on node \"crc\" DevicePath \"\"" Feb 27 08:59:10 crc kubenswrapper[4612]: I0227 08:59:10.711907 4612 scope.go:117] "RemoveContainer" containerID="d39e1e208b749c107d0156055aea1fb77deebb69104511981ee9602d176db66e" Feb 27 08:59:10 crc kubenswrapper[4612]: I0227 08:59:10.714905 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/20a06f2f-cc9e-460f-93a2-d40186958b17-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "20a06f2f-cc9e-460f-93a2-d40186958b17" (UID: "20a06f2f-cc9e-460f-93a2-d40186958b17"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 08:59:10 crc kubenswrapper[4612]: I0227 08:59:10.743463 4612 scope.go:117] "RemoveContainer" containerID="98f5549ad40c65600fd0ca072101fe2f6f7f518d058f2e493d58457c7bbb87ad" Feb 27 08:59:10 crc kubenswrapper[4612]: I0227 08:59:10.784837 4612 scope.go:117] "RemoveContainer" containerID="c25cd841b66fb17c777d42668eda205250ad25d4b7ab2c8a57016678cc70e7c6" Feb 27 08:59:10 crc kubenswrapper[4612]: E0227 08:59:10.785318 4612 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c25cd841b66fb17c777d42668eda205250ad25d4b7ab2c8a57016678cc70e7c6\": container with ID starting with c25cd841b66fb17c777d42668eda205250ad25d4b7ab2c8a57016678cc70e7c6 not found: ID does not exist" containerID="c25cd841b66fb17c777d42668eda205250ad25d4b7ab2c8a57016678cc70e7c6" Feb 27 08:59:10 crc kubenswrapper[4612]: I0227 08:59:10.785347 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c25cd841b66fb17c777d42668eda205250ad25d4b7ab2c8a57016678cc70e7c6"} err="failed to get container status \"c25cd841b66fb17c777d42668eda205250ad25d4b7ab2c8a57016678cc70e7c6\": rpc error: code = NotFound desc = could not find container \"c25cd841b66fb17c777d42668eda205250ad25d4b7ab2c8a57016678cc70e7c6\": container with ID starting with c25cd841b66fb17c777d42668eda205250ad25d4b7ab2c8a57016678cc70e7c6 not found: ID does not exist" Feb 27 08:59:10 crc kubenswrapper[4612]: I0227 08:59:10.785367 4612 scope.go:117] "RemoveContainer" containerID="d39e1e208b749c107d0156055aea1fb77deebb69104511981ee9602d176db66e" Feb 27 08:59:10 crc kubenswrapper[4612]: E0227 08:59:10.785606 4612 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d39e1e208b749c107d0156055aea1fb77deebb69104511981ee9602d176db66e\": container with ID starting with d39e1e208b749c107d0156055aea1fb77deebb69104511981ee9602d176db66e not found: ID does not exist" containerID="d39e1e208b749c107d0156055aea1fb77deebb69104511981ee9602d176db66e" Feb 27 08:59:10 crc kubenswrapper[4612]: I0227 08:59:10.785630 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d39e1e208b749c107d0156055aea1fb77deebb69104511981ee9602d176db66e"} err="failed to get container status \"d39e1e208b749c107d0156055aea1fb77deebb69104511981ee9602d176db66e\": rpc error: code = NotFound desc = could not find container \"d39e1e208b749c107d0156055aea1fb77deebb69104511981ee9602d176db66e\": container with ID starting with d39e1e208b749c107d0156055aea1fb77deebb69104511981ee9602d176db66e not found: ID does not exist" Feb 27 08:59:10 crc kubenswrapper[4612]: I0227 08:59:10.785647 4612 scope.go:117] "RemoveContainer" containerID="98f5549ad40c65600fd0ca072101fe2f6f7f518d058f2e493d58457c7bbb87ad" Feb 27 08:59:10 crc kubenswrapper[4612]: E0227 08:59:10.786131 4612 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"98f5549ad40c65600fd0ca072101fe2f6f7f518d058f2e493d58457c7bbb87ad\": container with ID starting with 98f5549ad40c65600fd0ca072101fe2f6f7f518d058f2e493d58457c7bbb87ad not found: ID does not exist" containerID="98f5549ad40c65600fd0ca072101fe2f6f7f518d058f2e493d58457c7bbb87ad" Feb 27 08:59:10 crc kubenswrapper[4612]: I0227 08:59:10.786153 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"98f5549ad40c65600fd0ca072101fe2f6f7f518d058f2e493d58457c7bbb87ad"} err="failed to get container status \"98f5549ad40c65600fd0ca072101fe2f6f7f518d058f2e493d58457c7bbb87ad\": rpc error: code = NotFound desc = could not find container \"98f5549ad40c65600fd0ca072101fe2f6f7f518d058f2e493d58457c7bbb87ad\": container with ID starting with 98f5549ad40c65600fd0ca072101fe2f6f7f518d058f2e493d58457c7bbb87ad not found: ID does not exist" Feb 27 08:59:10 crc kubenswrapper[4612]: I0227 08:59:10.789396 4612 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/20a06f2f-cc9e-460f-93a2-d40186958b17-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 27 08:59:10 crc kubenswrapper[4612]: I0227 08:59:10.866449 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ce6584ee-7f03-46a1-a30e-741b7f3c08d2" path="/var/lib/kubelet/pods/ce6584ee-7f03-46a1-a30e-741b7f3c08d2/volumes" Feb 27 08:59:10 crc kubenswrapper[4612]: I0227 08:59:10.975249 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-pnrpr"] Feb 27 08:59:10 crc kubenswrapper[4612]: I0227 08:59:10.984586 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-pnrpr"] Feb 27 08:59:11 crc kubenswrapper[4612]: I0227 08:59:11.051073 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-mchzb/crc-debug-h2nrv"] Feb 27 08:59:11 crc kubenswrapper[4612]: E0227 08:59:11.051545 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce6584ee-7f03-46a1-a30e-741b7f3c08d2" containerName="container-00" Feb 27 08:59:11 crc kubenswrapper[4612]: I0227 08:59:11.051571 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce6584ee-7f03-46a1-a30e-741b7f3c08d2" containerName="container-00" Feb 27 08:59:11 crc kubenswrapper[4612]: E0227 08:59:11.051590 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="20a06f2f-cc9e-460f-93a2-d40186958b17" containerName="registry-server" Feb 27 08:59:11 crc kubenswrapper[4612]: I0227 08:59:11.051600 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="20a06f2f-cc9e-460f-93a2-d40186958b17" containerName="registry-server" Feb 27 08:59:11 crc kubenswrapper[4612]: E0227 08:59:11.051628 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="20a06f2f-cc9e-460f-93a2-d40186958b17" containerName="extract-content" Feb 27 08:59:11 crc kubenswrapper[4612]: I0227 08:59:11.051637 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="20a06f2f-cc9e-460f-93a2-d40186958b17" containerName="extract-content" Feb 27 08:59:11 crc kubenswrapper[4612]: E0227 08:59:11.051656 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="20a06f2f-cc9e-460f-93a2-d40186958b17" containerName="extract-utilities" Feb 27 08:59:11 crc kubenswrapper[4612]: I0227 08:59:11.051664 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="20a06f2f-cc9e-460f-93a2-d40186958b17" containerName="extract-utilities" Feb 27 08:59:11 crc kubenswrapper[4612]: E0227 08:59:11.051681 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c50f2ffa-23aa-482f-87ee-d0d700753b07" containerName="oc" Feb 27 08:59:11 crc kubenswrapper[4612]: I0227 08:59:11.051689 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="c50f2ffa-23aa-482f-87ee-d0d700753b07" containerName="oc" Feb 27 08:59:11 crc kubenswrapper[4612]: I0227 08:59:11.051985 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="ce6584ee-7f03-46a1-a30e-741b7f3c08d2" containerName="container-00" Feb 27 08:59:11 crc kubenswrapper[4612]: I0227 08:59:11.052029 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="c50f2ffa-23aa-482f-87ee-d0d700753b07" containerName="oc" Feb 27 08:59:11 crc kubenswrapper[4612]: I0227 08:59:11.052042 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="20a06f2f-cc9e-460f-93a2-d40186958b17" containerName="registry-server" Feb 27 08:59:11 crc kubenswrapper[4612]: I0227 08:59:11.052782 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mchzb/crc-debug-h2nrv" Feb 27 08:59:11 crc kubenswrapper[4612]: I0227 08:59:11.196162 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8jd7k\" (UniqueName: \"kubernetes.io/projected/74156337-72b1-4df1-9f68-fba571a2b91e-kube-api-access-8jd7k\") pod \"crc-debug-h2nrv\" (UID: \"74156337-72b1-4df1-9f68-fba571a2b91e\") " pod="openshift-must-gather-mchzb/crc-debug-h2nrv" Feb 27 08:59:11 crc kubenswrapper[4612]: I0227 08:59:11.196384 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/74156337-72b1-4df1-9f68-fba571a2b91e-host\") pod \"crc-debug-h2nrv\" (UID: \"74156337-72b1-4df1-9f68-fba571a2b91e\") " pod="openshift-must-gather-mchzb/crc-debug-h2nrv" Feb 27 08:59:11 crc kubenswrapper[4612]: I0227 08:59:11.298304 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/74156337-72b1-4df1-9f68-fba571a2b91e-host\") pod \"crc-debug-h2nrv\" (UID: \"74156337-72b1-4df1-9f68-fba571a2b91e\") " pod="openshift-must-gather-mchzb/crc-debug-h2nrv" Feb 27 08:59:11 crc kubenswrapper[4612]: I0227 08:59:11.298434 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8jd7k\" (UniqueName: \"kubernetes.io/projected/74156337-72b1-4df1-9f68-fba571a2b91e-kube-api-access-8jd7k\") pod \"crc-debug-h2nrv\" (UID: \"74156337-72b1-4df1-9f68-fba571a2b91e\") " pod="openshift-must-gather-mchzb/crc-debug-h2nrv" Feb 27 08:59:11 crc kubenswrapper[4612]: I0227 08:59:11.298443 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/74156337-72b1-4df1-9f68-fba571a2b91e-host\") pod \"crc-debug-h2nrv\" (UID: \"74156337-72b1-4df1-9f68-fba571a2b91e\") " pod="openshift-must-gather-mchzb/crc-debug-h2nrv" Feb 27 08:59:11 crc kubenswrapper[4612]: I0227 08:59:11.322354 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8jd7k\" (UniqueName: \"kubernetes.io/projected/74156337-72b1-4df1-9f68-fba571a2b91e-kube-api-access-8jd7k\") pod \"crc-debug-h2nrv\" (UID: \"74156337-72b1-4df1-9f68-fba571a2b91e\") " pod="openshift-must-gather-mchzb/crc-debug-h2nrv" Feb 27 08:59:11 crc kubenswrapper[4612]: I0227 08:59:11.369813 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mchzb/crc-debug-h2nrv" Feb 27 08:59:11 crc kubenswrapper[4612]: I0227 08:59:11.663353 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mchzb/crc-debug-h2nrv" event={"ID":"74156337-72b1-4df1-9f68-fba571a2b91e","Type":"ContainerStarted","Data":"48ebf8c43d201cdb1c751905cabb0358a44ccf8fa725e65bf0cd190df29e5d51"} Feb 27 08:59:11 crc kubenswrapper[4612]: I0227 08:59:11.663782 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mchzb/crc-debug-h2nrv" event={"ID":"74156337-72b1-4df1-9f68-fba571a2b91e","Type":"ContainerStarted","Data":"0a614a25f50bcf572443bfa9b85ccc7c6365ae1ba679f4bcc02f468a1a046a7b"} Feb 27 08:59:11 crc kubenswrapper[4612]: I0227 08:59:11.678075 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-mchzb/crc-debug-h2nrv" podStartSLOduration=0.67805409 podStartE2EDuration="678.05409ms" podCreationTimestamp="2026-02-27 08:59:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 08:59:11.676574028 +0000 UTC m=+4209.530504056" watchObservedRunningTime="2026-02-27 08:59:11.67805409 +0000 UTC m=+4209.531984098" Feb 27 08:59:12 crc kubenswrapper[4612]: I0227 08:59:12.674507 4612 generic.go:334] "Generic (PLEG): container finished" podID="74156337-72b1-4df1-9f68-fba571a2b91e" containerID="48ebf8c43d201cdb1c751905cabb0358a44ccf8fa725e65bf0cd190df29e5d51" exitCode=0 Feb 27 08:59:12 crc kubenswrapper[4612]: I0227 08:59:12.674605 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mchzb/crc-debug-h2nrv" event={"ID":"74156337-72b1-4df1-9f68-fba571a2b91e","Type":"ContainerDied","Data":"48ebf8c43d201cdb1c751905cabb0358a44ccf8fa725e65bf0cd190df29e5d51"} Feb 27 08:59:12 crc kubenswrapper[4612]: I0227 08:59:12.879997 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20a06f2f-cc9e-460f-93a2-d40186958b17" path="/var/lib/kubelet/pods/20a06f2f-cc9e-460f-93a2-d40186958b17/volumes" Feb 27 08:59:13 crc kubenswrapper[4612]: I0227 08:59:13.824742 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mchzb/crc-debug-h2nrv" Feb 27 08:59:13 crc kubenswrapper[4612]: I0227 08:59:13.859791 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-mchzb/crc-debug-h2nrv"] Feb 27 08:59:13 crc kubenswrapper[4612]: I0227 08:59:13.872233 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-mchzb/crc-debug-h2nrv"] Feb 27 08:59:13 crc kubenswrapper[4612]: I0227 08:59:13.961356 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/74156337-72b1-4df1-9f68-fba571a2b91e-host\") pod \"74156337-72b1-4df1-9f68-fba571a2b91e\" (UID: \"74156337-72b1-4df1-9f68-fba571a2b91e\") " Feb 27 08:59:13 crc kubenswrapper[4612]: I0227 08:59:13.961416 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8jd7k\" (UniqueName: \"kubernetes.io/projected/74156337-72b1-4df1-9f68-fba571a2b91e-kube-api-access-8jd7k\") pod \"74156337-72b1-4df1-9f68-fba571a2b91e\" (UID: \"74156337-72b1-4df1-9f68-fba571a2b91e\") " Feb 27 08:59:13 crc kubenswrapper[4612]: I0227 08:59:13.961500 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/74156337-72b1-4df1-9f68-fba571a2b91e-host" (OuterVolumeSpecName: "host") pod "74156337-72b1-4df1-9f68-fba571a2b91e" (UID: "74156337-72b1-4df1-9f68-fba571a2b91e"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 27 08:59:13 crc kubenswrapper[4612]: I0227 08:59:13.962351 4612 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/74156337-72b1-4df1-9f68-fba571a2b91e-host\") on node \"crc\" DevicePath \"\"" Feb 27 08:59:13 crc kubenswrapper[4612]: I0227 08:59:13.967868 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/74156337-72b1-4df1-9f68-fba571a2b91e-kube-api-access-8jd7k" (OuterVolumeSpecName: "kube-api-access-8jd7k") pod "74156337-72b1-4df1-9f68-fba571a2b91e" (UID: "74156337-72b1-4df1-9f68-fba571a2b91e"). InnerVolumeSpecName "kube-api-access-8jd7k". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 08:59:14 crc kubenswrapper[4612]: I0227 08:59:14.063890 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8jd7k\" (UniqueName: \"kubernetes.io/projected/74156337-72b1-4df1-9f68-fba571a2b91e-kube-api-access-8jd7k\") on node \"crc\" DevicePath \"\"" Feb 27 08:59:14 crc kubenswrapper[4612]: I0227 08:59:14.736939 4612 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0a614a25f50bcf572443bfa9b85ccc7c6365ae1ba679f4bcc02f468a1a046a7b" Feb 27 08:59:14 crc kubenswrapper[4612]: I0227 08:59:14.737031 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mchzb/crc-debug-h2nrv" Feb 27 08:59:14 crc kubenswrapper[4612]: I0227 08:59:14.863273 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="74156337-72b1-4df1-9f68-fba571a2b91e" path="/var/lib/kubelet/pods/74156337-72b1-4df1-9f68-fba571a2b91e/volumes" Feb 27 08:59:15 crc kubenswrapper[4612]: I0227 08:59:15.131068 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-mchzb/crc-debug-rch9k"] Feb 27 08:59:15 crc kubenswrapper[4612]: E0227 08:59:15.131434 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74156337-72b1-4df1-9f68-fba571a2b91e" containerName="container-00" Feb 27 08:59:15 crc kubenswrapper[4612]: I0227 08:59:15.131449 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="74156337-72b1-4df1-9f68-fba571a2b91e" containerName="container-00" Feb 27 08:59:15 crc kubenswrapper[4612]: I0227 08:59:15.131617 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="74156337-72b1-4df1-9f68-fba571a2b91e" containerName="container-00" Feb 27 08:59:15 crc kubenswrapper[4612]: I0227 08:59:15.132164 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mchzb/crc-debug-rch9k" Feb 27 08:59:15 crc kubenswrapper[4612]: I0227 08:59:15.204846 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fb79n\" (UniqueName: \"kubernetes.io/projected/1a951b45-d1b3-4e5c-ab0b-ec8081522be2-kube-api-access-fb79n\") pod \"crc-debug-rch9k\" (UID: \"1a951b45-d1b3-4e5c-ab0b-ec8081522be2\") " pod="openshift-must-gather-mchzb/crc-debug-rch9k" Feb 27 08:59:15 crc kubenswrapper[4612]: I0227 08:59:15.205259 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/1a951b45-d1b3-4e5c-ab0b-ec8081522be2-host\") pod \"crc-debug-rch9k\" (UID: \"1a951b45-d1b3-4e5c-ab0b-ec8081522be2\") " pod="openshift-must-gather-mchzb/crc-debug-rch9k" Feb 27 08:59:15 crc kubenswrapper[4612]: I0227 08:59:15.306665 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/1a951b45-d1b3-4e5c-ab0b-ec8081522be2-host\") pod \"crc-debug-rch9k\" (UID: \"1a951b45-d1b3-4e5c-ab0b-ec8081522be2\") " pod="openshift-must-gather-mchzb/crc-debug-rch9k" Feb 27 08:59:15 crc kubenswrapper[4612]: I0227 08:59:15.306797 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fb79n\" (UniqueName: \"kubernetes.io/projected/1a951b45-d1b3-4e5c-ab0b-ec8081522be2-kube-api-access-fb79n\") pod \"crc-debug-rch9k\" (UID: \"1a951b45-d1b3-4e5c-ab0b-ec8081522be2\") " pod="openshift-must-gather-mchzb/crc-debug-rch9k" Feb 27 08:59:15 crc kubenswrapper[4612]: I0227 08:59:15.306905 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/1a951b45-d1b3-4e5c-ab0b-ec8081522be2-host\") pod \"crc-debug-rch9k\" (UID: \"1a951b45-d1b3-4e5c-ab0b-ec8081522be2\") " pod="openshift-must-gather-mchzb/crc-debug-rch9k" Feb 27 08:59:15 crc kubenswrapper[4612]: I0227 08:59:15.331636 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fb79n\" (UniqueName: \"kubernetes.io/projected/1a951b45-d1b3-4e5c-ab0b-ec8081522be2-kube-api-access-fb79n\") pod \"crc-debug-rch9k\" (UID: \"1a951b45-d1b3-4e5c-ab0b-ec8081522be2\") " pod="openshift-must-gather-mchzb/crc-debug-rch9k" Feb 27 08:59:15 crc kubenswrapper[4612]: I0227 08:59:15.446519 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mchzb/crc-debug-rch9k" Feb 27 08:59:15 crc kubenswrapper[4612]: W0227 08:59:15.474324 4612 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1a951b45_d1b3_4e5c_ab0b_ec8081522be2.slice/crio-de7e3a3adea45c54a88e54b1d4e698a62aeed76a59528ce8a661deac8781d8ff WatchSource:0}: Error finding container de7e3a3adea45c54a88e54b1d4e698a62aeed76a59528ce8a661deac8781d8ff: Status 404 returned error can't find the container with id de7e3a3adea45c54a88e54b1d4e698a62aeed76a59528ce8a661deac8781d8ff Feb 27 08:59:15 crc kubenswrapper[4612]: I0227 08:59:15.748229 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mchzb/crc-debug-rch9k" event={"ID":"1a951b45-d1b3-4e5c-ab0b-ec8081522be2","Type":"ContainerStarted","Data":"de7e3a3adea45c54a88e54b1d4e698a62aeed76a59528ce8a661deac8781d8ff"} Feb 27 08:59:16 crc kubenswrapper[4612]: I0227 08:59:16.776763 4612 generic.go:334] "Generic (PLEG): container finished" podID="1a951b45-d1b3-4e5c-ab0b-ec8081522be2" containerID="42a1de80ebcbe985e5b3da1a0f9e47fbf9349457376f6237bef7f26cd5887b3b" exitCode=0 Feb 27 08:59:16 crc kubenswrapper[4612]: I0227 08:59:16.776966 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mchzb/crc-debug-rch9k" event={"ID":"1a951b45-d1b3-4e5c-ab0b-ec8081522be2","Type":"ContainerDied","Data":"42a1de80ebcbe985e5b3da1a0f9e47fbf9349457376f6237bef7f26cd5887b3b"} Feb 27 08:59:16 crc kubenswrapper[4612]: I0227 08:59:16.817591 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-mchzb/crc-debug-rch9k"] Feb 27 08:59:16 crc kubenswrapper[4612]: I0227 08:59:16.827418 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-mchzb/crc-debug-rch9k"] Feb 27 08:59:17 crc kubenswrapper[4612]: I0227 08:59:17.896378 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mchzb/crc-debug-rch9k" Feb 27 08:59:17 crc kubenswrapper[4612]: I0227 08:59:17.979274 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/1a951b45-d1b3-4e5c-ab0b-ec8081522be2-host\") pod \"1a951b45-d1b3-4e5c-ab0b-ec8081522be2\" (UID: \"1a951b45-d1b3-4e5c-ab0b-ec8081522be2\") " Feb 27 08:59:17 crc kubenswrapper[4612]: I0227 08:59:17.979343 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fb79n\" (UniqueName: \"kubernetes.io/projected/1a951b45-d1b3-4e5c-ab0b-ec8081522be2-kube-api-access-fb79n\") pod \"1a951b45-d1b3-4e5c-ab0b-ec8081522be2\" (UID: \"1a951b45-d1b3-4e5c-ab0b-ec8081522be2\") " Feb 27 08:59:17 crc kubenswrapper[4612]: I0227 08:59:17.979422 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1a951b45-d1b3-4e5c-ab0b-ec8081522be2-host" (OuterVolumeSpecName: "host") pod "1a951b45-d1b3-4e5c-ab0b-ec8081522be2" (UID: "1a951b45-d1b3-4e5c-ab0b-ec8081522be2"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 27 08:59:17 crc kubenswrapper[4612]: I0227 08:59:17.980134 4612 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/1a951b45-d1b3-4e5c-ab0b-ec8081522be2-host\") on node \"crc\" DevicePath \"\"" Feb 27 08:59:17 crc kubenswrapper[4612]: I0227 08:59:17.992378 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1a951b45-d1b3-4e5c-ab0b-ec8081522be2-kube-api-access-fb79n" (OuterVolumeSpecName: "kube-api-access-fb79n") pod "1a951b45-d1b3-4e5c-ab0b-ec8081522be2" (UID: "1a951b45-d1b3-4e5c-ab0b-ec8081522be2"). InnerVolumeSpecName "kube-api-access-fb79n". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 08:59:18 crc kubenswrapper[4612]: I0227 08:59:18.082337 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fb79n\" (UniqueName: \"kubernetes.io/projected/1a951b45-d1b3-4e5c-ab0b-ec8081522be2-kube-api-access-fb79n\") on node \"crc\" DevicePath \"\"" Feb 27 08:59:18 crc kubenswrapper[4612]: I0227 08:59:18.801678 4612 scope.go:117] "RemoveContainer" containerID="42a1de80ebcbe985e5b3da1a0f9e47fbf9349457376f6237bef7f26cd5887b3b" Feb 27 08:59:18 crc kubenswrapper[4612]: I0227 08:59:18.801742 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mchzb/crc-debug-rch9k" Feb 27 08:59:18 crc kubenswrapper[4612]: I0227 08:59:18.874388 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1a951b45-d1b3-4e5c-ab0b-ec8081522be2" path="/var/lib/kubelet/pods/1a951b45-d1b3-4e5c-ab0b-ec8081522be2/volumes" Feb 27 08:59:53 crc kubenswrapper[4612]: I0227 08:59:53.484040 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-7db7f7dcc8-f95qr_16621861-feac-4972-ad1c-1e365692c23e/barbican-api/0.log" Feb 27 08:59:53 crc kubenswrapper[4612]: I0227 08:59:53.652895 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-7db7f7dcc8-f95qr_16621861-feac-4972-ad1c-1e365692c23e/barbican-api-log/0.log" Feb 27 08:59:53 crc kubenswrapper[4612]: I0227 08:59:53.772629 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-745b984886-tzpzt_e2579b13-034e-474a-bbe6-2062a710f380/barbican-keystone-listener/0.log" Feb 27 08:59:53 crc kubenswrapper[4612]: I0227 08:59:53.784841 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-745b984886-tzpzt_e2579b13-034e-474a-bbe6-2062a710f380/barbican-keystone-listener-log/0.log" Feb 27 08:59:53 crc kubenswrapper[4612]: I0227 08:59:53.938431 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-7558bddf6c-j9f2z_74bec848-d5b3-43d0-97d7-eda229642bb5/barbican-worker/0.log" Feb 27 08:59:54 crc kubenswrapper[4612]: I0227 08:59:54.485607 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-7558bddf6c-j9f2z_74bec848-d5b3-43d0-97d7-eda229642bb5/barbican-worker-log/0.log" Feb 27 08:59:54 crc kubenswrapper[4612]: I0227 08:59:54.577082 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-k4jmw_9caaee00-687f-47de-8336-859a249751f7/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Feb 27 08:59:54 crc kubenswrapper[4612]: I0227 08:59:54.751286 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_16ab2b55-65e7-4d63-b521-0c6421ba52ea/ceilometer-central-agent/0.log" Feb 27 08:59:55 crc kubenswrapper[4612]: I0227 08:59:55.703436 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_16ab2b55-65e7-4d63-b521-0c6421ba52ea/ceilometer-notification-agent/0.log" Feb 27 08:59:55 crc kubenswrapper[4612]: I0227 08:59:55.740564 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_16ab2b55-65e7-4d63-b521-0c6421ba52ea/proxy-httpd/0.log" Feb 27 08:59:55 crc kubenswrapper[4612]: I0227 08:59:55.793311 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_16ab2b55-65e7-4d63-b521-0c6421ba52ea/sg-core/0.log" Feb 27 08:59:56 crc kubenswrapper[4612]: I0227 08:59:56.038095 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_9ba31a95-0fce-4eb6-98be-6292b7146e21/cinder-api-log/0.log" Feb 27 08:59:56 crc kubenswrapper[4612]: I0227 08:59:56.135281 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_9ba31a95-0fce-4eb6-98be-6292b7146e21/cinder-api/0.log" Feb 27 08:59:56 crc kubenswrapper[4612]: I0227 08:59:56.838474 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_0cf1b618-59b9-4688-85b9-1a7f6d2af657/cinder-scheduler/0.log" Feb 27 08:59:56 crc kubenswrapper[4612]: I0227 08:59:56.841905 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-v4fds_c6f08c4d-80c1-49fb-b8ac-6d25b9d7f143/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Feb 27 08:59:56 crc kubenswrapper[4612]: I0227 08:59:56.997699 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_0cf1b618-59b9-4688-85b9-1a7f6d2af657/probe/0.log" Feb 27 08:59:57 crc kubenswrapper[4612]: I0227 08:59:57.164155 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-w9sl2_68ded60c-e2ee-48fb-986b-cebf9e447b82/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Feb 27 08:59:57 crc kubenswrapper[4612]: I0227 08:59:57.269130 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-798f84bf5f-g5zjp_3c836155-a5e3-4181-9460-4147825f36ab/init/0.log" Feb 27 08:59:57 crc kubenswrapper[4612]: I0227 08:59:57.578593 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-798f84bf5f-g5zjp_3c836155-a5e3-4181-9460-4147825f36ab/init/0.log" Feb 27 08:59:57 crc kubenswrapper[4612]: I0227 08:59:57.607913 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-t4m6d_8439abb9-0c80-44f9-9ec4-ef0c99a50182/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Feb 27 08:59:57 crc kubenswrapper[4612]: I0227 08:59:57.668123 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-798f84bf5f-g5zjp_3c836155-a5e3-4181-9460-4147825f36ab/dnsmasq-dns/0.log" Feb 27 08:59:57 crc kubenswrapper[4612]: I0227 08:59:57.827746 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_58945c32-1660-410d-9a49-b0a38829f907/glance-httpd/0.log" Feb 27 08:59:57 crc kubenswrapper[4612]: I0227 08:59:57.952903 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_58945c32-1660-410d-9a49-b0a38829f907/glance-log/0.log" Feb 27 08:59:58 crc kubenswrapper[4612]: I0227 08:59:58.103961 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_f4ea14ec-7119-4952-a548-8afa0ad086bb/glance-httpd/0.log" Feb 27 08:59:58 crc kubenswrapper[4612]: I0227 08:59:58.276783 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_f4ea14ec-7119-4952-a548-8afa0ad086bb/glance-log/0.log" Feb 27 08:59:58 crc kubenswrapper[4612]: I0227 08:59:58.467052 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-77948db5bb-97qh6_b72b541d-4661-44a7-a121-c2a8aef6db11/horizon/1.log" Feb 27 08:59:58 crc kubenswrapper[4612]: I0227 08:59:58.720282 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-77948db5bb-97qh6_b72b541d-4661-44a7-a121-c2a8aef6db11/horizon/0.log" Feb 27 08:59:59 crc kubenswrapper[4612]: I0227 08:59:59.550313 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-7fwd7_301bc35d-5ee7-4a2e-af9b-47a57e6c5123/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Feb 27 08:59:59 crc kubenswrapper[4612]: I0227 08:59:59.566226 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-frrmr_4f248d51-18f9-4d20-b64b-d23765054818/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Feb 27 08:59:59 crc kubenswrapper[4612]: I0227 08:59:59.594753 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-77948db5bb-97qh6_b72b541d-4661-44a7-a121-c2a8aef6db11/horizon-log/0.log" Feb 27 08:59:59 crc kubenswrapper[4612]: I0227 08:59:59.810940 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_7a3af89c-deaa-4d32-b00b-ae3ca1de8ee9/kube-state-metrics/0.log" Feb 27 09:00:00 crc kubenswrapper[4612]: I0227 09:00:00.168627 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29536380-52gnm"] Feb 27 09:00:00 crc kubenswrapper[4612]: E0227 09:00:00.170057 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a951b45-d1b3-4e5c-ab0b-ec8081522be2" containerName="container-00" Feb 27 09:00:00 crc kubenswrapper[4612]: I0227 09:00:00.170081 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a951b45-d1b3-4e5c-ab0b-ec8081522be2" containerName="container-00" Feb 27 09:00:00 crc kubenswrapper[4612]: I0227 09:00:00.170317 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="1a951b45-d1b3-4e5c-ab0b-ec8081522be2" containerName="container-00" Feb 27 09:00:00 crc kubenswrapper[4612]: I0227 09:00:00.170966 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29536380-52gnm" Feb 27 09:00:00 crc kubenswrapper[4612]: I0227 09:00:00.179363 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Feb 27 09:00:00 crc kubenswrapper[4612]: I0227 09:00:00.179611 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Feb 27 09:00:00 crc kubenswrapper[4612]: I0227 09:00:00.182635 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29536380-wkqc6"] Feb 27 09:00:00 crc kubenswrapper[4612]: I0227 09:00:00.184162 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536380-wkqc6" Feb 27 09:00:00 crc kubenswrapper[4612]: I0227 09:00:00.212526 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 27 09:00:00 crc kubenswrapper[4612]: I0227 09:00:00.213367 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-5zwfx" Feb 27 09:00:00 crc kubenswrapper[4612]: I0227 09:00:00.213917 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 27 09:00:00 crc kubenswrapper[4612]: I0227 09:00:00.241564 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29536380-52gnm"] Feb 27 09:00:00 crc kubenswrapper[4612]: I0227 09:00:00.245378 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lkbc9\" (UniqueName: \"kubernetes.io/projected/e1e5befa-cd14-4bb6-82be-547ace37b1d7-kube-api-access-lkbc9\") pod \"collect-profiles-29536380-52gnm\" (UID: \"e1e5befa-cd14-4bb6-82be-547ace37b1d7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29536380-52gnm" Feb 27 09:00:00 crc kubenswrapper[4612]: I0227 09:00:00.245422 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zg877\" (UniqueName: \"kubernetes.io/projected/45882240-d427-4e23-9415-100cbdcedb42-kube-api-access-zg877\") pod \"auto-csr-approver-29536380-wkqc6\" (UID: \"45882240-d427-4e23-9415-100cbdcedb42\") " pod="openshift-infra/auto-csr-approver-29536380-wkqc6" Feb 27 09:00:00 crc kubenswrapper[4612]: I0227 09:00:00.245555 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e1e5befa-cd14-4bb6-82be-547ace37b1d7-config-volume\") pod \"collect-profiles-29536380-52gnm\" (UID: \"e1e5befa-cd14-4bb6-82be-547ace37b1d7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29536380-52gnm" Feb 27 09:00:00 crc kubenswrapper[4612]: I0227 09:00:00.245581 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e1e5befa-cd14-4bb6-82be-547ace37b1d7-secret-volume\") pod \"collect-profiles-29536380-52gnm\" (UID: \"e1e5befa-cd14-4bb6-82be-547ace37b1d7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29536380-52gnm" Feb 27 09:00:00 crc kubenswrapper[4612]: I0227 09:00:00.294305 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536380-wkqc6"] Feb 27 09:00:00 crc kubenswrapper[4612]: I0227 09:00:00.321678 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-9c6fdd6-lvn8p_71940e05-ecfa-46de-aefe-c213add5362c/keystone-api/0.log" Feb 27 09:00:00 crc kubenswrapper[4612]: I0227 09:00:00.349518 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e1e5befa-cd14-4bb6-82be-547ace37b1d7-config-volume\") pod \"collect-profiles-29536380-52gnm\" (UID: \"e1e5befa-cd14-4bb6-82be-547ace37b1d7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29536380-52gnm" Feb 27 09:00:00 crc kubenswrapper[4612]: I0227 09:00:00.349569 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e1e5befa-cd14-4bb6-82be-547ace37b1d7-secret-volume\") pod \"collect-profiles-29536380-52gnm\" (UID: \"e1e5befa-cd14-4bb6-82be-547ace37b1d7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29536380-52gnm" Feb 27 09:00:00 crc kubenswrapper[4612]: I0227 09:00:00.349632 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lkbc9\" (UniqueName: \"kubernetes.io/projected/e1e5befa-cd14-4bb6-82be-547ace37b1d7-kube-api-access-lkbc9\") pod \"collect-profiles-29536380-52gnm\" (UID: \"e1e5befa-cd14-4bb6-82be-547ace37b1d7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29536380-52gnm" Feb 27 09:00:00 crc kubenswrapper[4612]: I0227 09:00:00.349665 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zg877\" (UniqueName: \"kubernetes.io/projected/45882240-d427-4e23-9415-100cbdcedb42-kube-api-access-zg877\") pod \"auto-csr-approver-29536380-wkqc6\" (UID: \"45882240-d427-4e23-9415-100cbdcedb42\") " pod="openshift-infra/auto-csr-approver-29536380-wkqc6" Feb 27 09:00:00 crc kubenswrapper[4612]: I0227 09:00:00.364876 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e1e5befa-cd14-4bb6-82be-547ace37b1d7-config-volume\") pod \"collect-profiles-29536380-52gnm\" (UID: \"e1e5befa-cd14-4bb6-82be-547ace37b1d7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29536380-52gnm" Feb 27 09:00:00 crc kubenswrapper[4612]: I0227 09:00:00.390227 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lkbc9\" (UniqueName: \"kubernetes.io/projected/e1e5befa-cd14-4bb6-82be-547ace37b1d7-kube-api-access-lkbc9\") pod \"collect-profiles-29536380-52gnm\" (UID: \"e1e5befa-cd14-4bb6-82be-547ace37b1d7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29536380-52gnm" Feb 27 09:00:00 crc kubenswrapper[4612]: I0227 09:00:00.399506 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zg877\" (UniqueName: \"kubernetes.io/projected/45882240-d427-4e23-9415-100cbdcedb42-kube-api-access-zg877\") pod \"auto-csr-approver-29536380-wkqc6\" (UID: \"45882240-d427-4e23-9415-100cbdcedb42\") " pod="openshift-infra/auto-csr-approver-29536380-wkqc6" Feb 27 09:00:00 crc kubenswrapper[4612]: I0227 09:00:00.400233 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e1e5befa-cd14-4bb6-82be-547ace37b1d7-secret-volume\") pod \"collect-profiles-29536380-52gnm\" (UID: \"e1e5befa-cd14-4bb6-82be-547ace37b1d7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29536380-52gnm" Feb 27 09:00:00 crc kubenswrapper[4612]: I0227 09:00:00.514757 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29536380-52gnm" Feb 27 09:00:00 crc kubenswrapper[4612]: I0227 09:00:00.553839 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536380-wkqc6" Feb 27 09:00:00 crc kubenswrapper[4612]: I0227 09:00:00.592180 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-27c56_f4308354-0bda-4fb3-976f-cc2fa471dcb4/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Feb 27 09:00:01 crc kubenswrapper[4612]: I0227 09:00:01.172424 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29536380-52gnm"] Feb 27 09:00:01 crc kubenswrapper[4612]: I0227 09:00:01.213108 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_d19bfe90-c5ca-4089-b4f0-bf3eccff30c8/memcached/0.log" Feb 27 09:00:01 crc kubenswrapper[4612]: I0227 09:00:01.402641 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536380-wkqc6"] Feb 27 09:00:01 crc kubenswrapper[4612]: I0227 09:00:01.411978 4612 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 27 09:00:01 crc kubenswrapper[4612]: I0227 09:00:01.419744 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-x52ht_a213a1c0-9409-487d-834a-a60e11fd8a66/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Feb 27 09:00:01 crc kubenswrapper[4612]: I0227 09:00:01.655975 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-cc6d746d7-b8wbb_41eb38cf-80fa-408e-b16c-686ef58a143f/neutron-httpd/0.log" Feb 27 09:00:01 crc kubenswrapper[4612]: I0227 09:00:01.660184 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-cc6d746d7-b8wbb_41eb38cf-80fa-408e-b16c-686ef58a143f/neutron-api/0.log" Feb 27 09:00:01 crc kubenswrapper[4612]: I0227 09:00:01.887239 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29536380-52gnm" event={"ID":"e1e5befa-cd14-4bb6-82be-547ace37b1d7","Type":"ContainerStarted","Data":"bc81ae2292e348be960ba79d18774b23c04cdf00eeb2d15c4ba5588fd5b7cfb7"} Feb 27 09:00:01 crc kubenswrapper[4612]: I0227 09:00:01.887281 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29536380-52gnm" event={"ID":"e1e5befa-cd14-4bb6-82be-547ace37b1d7","Type":"ContainerStarted","Data":"ad59c7344ae1fb1efe903067eb091adb20898cbeaf3302389d0d7a607d9e9830"} Feb 27 09:00:01 crc kubenswrapper[4612]: I0227 09:00:01.889122 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536380-wkqc6" event={"ID":"45882240-d427-4e23-9415-100cbdcedb42","Type":"ContainerStarted","Data":"d2edb1c26968147782b91586f4b48187e203dab06bb91a5559b184e37703fcd3"} Feb 27 09:00:02 crc kubenswrapper[4612]: I0227 09:00:02.257428 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_602ab3aa-60d5-45ff-9890-647a6f8ffb6c/nova-cell0-conductor-conductor/0.log" Feb 27 09:00:02 crc kubenswrapper[4612]: I0227 09:00:02.293661 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_fb906a42-d4ee-4e3e-b4c2-8924890a9e26/nova-cell1-conductor-conductor/0.log" Feb 27 09:00:02 crc kubenswrapper[4612]: I0227 09:00:02.701181 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_211046a5-967c-44f0-8320-ddcc223dda97/nova-cell1-novncproxy-novncproxy/0.log" Feb 27 09:00:02 crc kubenswrapper[4612]: I0227 09:00:02.757574 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-bnlnx_24758178-1be1-49f2-a15e-196426ec0dcd/nova-edpm-deployment-openstack-edpm-ipam/0.log" Feb 27 09:00:02 crc kubenswrapper[4612]: I0227 09:00:02.938288 4612 generic.go:334] "Generic (PLEG): container finished" podID="e1e5befa-cd14-4bb6-82be-547ace37b1d7" containerID="bc81ae2292e348be960ba79d18774b23c04cdf00eeb2d15c4ba5588fd5b7cfb7" exitCode=0 Feb 27 09:00:02 crc kubenswrapper[4612]: I0227 09:00:02.938468 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29536380-52gnm" event={"ID":"e1e5befa-cd14-4bb6-82be-547ace37b1d7","Type":"ContainerDied","Data":"bc81ae2292e348be960ba79d18774b23c04cdf00eeb2d15c4ba5588fd5b7cfb7"} Feb 27 09:00:02 crc kubenswrapper[4612]: I0227 09:00:02.946288 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_e492f61b-94d4-4c4e-ac47-168fdb1cd997/nova-api-log/0.log" Feb 27 09:00:02 crc kubenswrapper[4612]: I0227 09:00:02.957174 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_e492f61b-94d4-4c4e-ac47-168fdb1cd997/nova-api-api/0.log" Feb 27 09:00:03 crc kubenswrapper[4612]: I0227 09:00:03.089529 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_24c836ae-ba59-4432-acee-fd17b60c8236/nova-metadata-log/0.log" Feb 27 09:00:03 crc kubenswrapper[4612]: I0227 09:00:03.332072 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29536380-52gnm" Feb 27 09:00:03 crc kubenswrapper[4612]: I0227 09:00:03.386763 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e1e5befa-cd14-4bb6-82be-547ace37b1d7-config-volume\") pod \"e1e5befa-cd14-4bb6-82be-547ace37b1d7\" (UID: \"e1e5befa-cd14-4bb6-82be-547ace37b1d7\") " Feb 27 09:00:03 crc kubenswrapper[4612]: I0227 09:00:03.386980 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lkbc9\" (UniqueName: \"kubernetes.io/projected/e1e5befa-cd14-4bb6-82be-547ace37b1d7-kube-api-access-lkbc9\") pod \"e1e5befa-cd14-4bb6-82be-547ace37b1d7\" (UID: \"e1e5befa-cd14-4bb6-82be-547ace37b1d7\") " Feb 27 09:00:03 crc kubenswrapper[4612]: I0227 09:00:03.387049 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e1e5befa-cd14-4bb6-82be-547ace37b1d7-secret-volume\") pod \"e1e5befa-cd14-4bb6-82be-547ace37b1d7\" (UID: \"e1e5befa-cd14-4bb6-82be-547ace37b1d7\") " Feb 27 09:00:03 crc kubenswrapper[4612]: I0227 09:00:03.387739 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e1e5befa-cd14-4bb6-82be-547ace37b1d7-config-volume" (OuterVolumeSpecName: "config-volume") pod "e1e5befa-cd14-4bb6-82be-547ace37b1d7" (UID: "e1e5befa-cd14-4bb6-82be-547ace37b1d7"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 09:00:03 crc kubenswrapper[4612]: I0227 09:00:03.395703 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e1e5befa-cd14-4bb6-82be-547ace37b1d7-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "e1e5befa-cd14-4bb6-82be-547ace37b1d7" (UID: "e1e5befa-cd14-4bb6-82be-547ace37b1d7"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 09:00:03 crc kubenswrapper[4612]: I0227 09:00:03.411222 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e1e5befa-cd14-4bb6-82be-547ace37b1d7-kube-api-access-lkbc9" (OuterVolumeSpecName: "kube-api-access-lkbc9") pod "e1e5befa-cd14-4bb6-82be-547ace37b1d7" (UID: "e1e5befa-cd14-4bb6-82be-547ace37b1d7"). InnerVolumeSpecName "kube-api-access-lkbc9". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 09:00:03 crc kubenswrapper[4612]: I0227 09:00:03.496850 4612 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e1e5befa-cd14-4bb6-82be-547ace37b1d7-config-volume\") on node \"crc\" DevicePath \"\"" Feb 27 09:00:03 crc kubenswrapper[4612]: I0227 09:00:03.496894 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lkbc9\" (UniqueName: \"kubernetes.io/projected/e1e5befa-cd14-4bb6-82be-547ace37b1d7-kube-api-access-lkbc9\") on node \"crc\" DevicePath \"\"" Feb 27 09:00:03 crc kubenswrapper[4612]: I0227 09:00:03.496905 4612 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e1e5befa-cd14-4bb6-82be-547ace37b1d7-secret-volume\") on node \"crc\" DevicePath \"\"" Feb 27 09:00:03 crc kubenswrapper[4612]: I0227 09:00:03.519945 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_e02a40f2-3de1-4f01-84b0-2fe56acdd396/mysql-bootstrap/0.log" Feb 27 09:00:03 crc kubenswrapper[4612]: I0227 09:00:03.725855 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_43270ba1-f597-4ca7-bc1e-741d7b915940/nova-scheduler-scheduler/0.log" Feb 27 09:00:03 crc kubenswrapper[4612]: I0227 09:00:03.742652 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_e02a40f2-3de1-4f01-84b0-2fe56acdd396/mysql-bootstrap/0.log" Feb 27 09:00:03 crc kubenswrapper[4612]: I0227 09:00:03.778610 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_e02a40f2-3de1-4f01-84b0-2fe56acdd396/galera/0.log" Feb 27 09:00:03 crc kubenswrapper[4612]: I0227 09:00:03.947893 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29536380-52gnm" event={"ID":"e1e5befa-cd14-4bb6-82be-547ace37b1d7","Type":"ContainerDied","Data":"ad59c7344ae1fb1efe903067eb091adb20898cbeaf3302389d0d7a607d9e9830"} Feb 27 09:00:03 crc kubenswrapper[4612]: I0227 09:00:03.948205 4612 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ad59c7344ae1fb1efe903067eb091adb20898cbeaf3302389d0d7a607d9e9830" Feb 27 09:00:03 crc kubenswrapper[4612]: I0227 09:00:03.948266 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29536380-52gnm" Feb 27 09:00:04 crc kubenswrapper[4612]: I0227 09:00:04.021482 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_f189e394-0c97-48a4-9c55-4bd6710923f1/mysql-bootstrap/0.log" Feb 27 09:00:04 crc kubenswrapper[4612]: I0227 09:00:04.163393 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_24c836ae-ba59-4432-acee-fd17b60c8236/nova-metadata-metadata/0.log" Feb 27 09:00:04 crc kubenswrapper[4612]: I0227 09:00:04.208862 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_f189e394-0c97-48a4-9c55-4bd6710923f1/mysql-bootstrap/0.log" Feb 27 09:00:04 crc kubenswrapper[4612]: I0227 09:00:04.299119 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_f189e394-0c97-48a4-9c55-4bd6710923f1/galera/0.log" Feb 27 09:00:04 crc kubenswrapper[4612]: I0227 09:00:04.307047 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_9fd44447-aa5f-436f-8aaa-bf3b53d99e64/openstackclient/0.log" Feb 27 09:00:04 crc kubenswrapper[4612]: I0227 09:00:04.425496 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29536335-7bn5c"] Feb 27 09:00:04 crc kubenswrapper[4612]: I0227 09:00:04.439498 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29536335-7bn5c"] Feb 27 09:00:04 crc kubenswrapper[4612]: I0227 09:00:04.679421 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-m6w89_99cc2aa0-2437-49fe-a7e0-1df45cecd195/openstack-network-exporter/0.log" Feb 27 09:00:04 crc kubenswrapper[4612]: I0227 09:00:04.847392 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-x255x_20c3f793-4286-4067-a372-37b2de08fa8e/ovsdb-server-init/0.log" Feb 27 09:00:04 crc kubenswrapper[4612]: I0227 09:00:04.862434 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4ce49ed5-459c-42b9-bc5b-55c3f9ec5b45" path="/var/lib/kubelet/pods/4ce49ed5-459c-42b9-bc5b-55c3f9ec5b45/volumes" Feb 27 09:00:05 crc kubenswrapper[4612]: I0227 09:00:05.175265 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-x255x_20c3f793-4286-4067-a372-37b2de08fa8e/ovsdb-server-init/0.log" Feb 27 09:00:05 crc kubenswrapper[4612]: I0227 09:00:05.227342 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-x255x_20c3f793-4286-4067-a372-37b2de08fa8e/ovsdb-server/0.log" Feb 27 09:00:05 crc kubenswrapper[4612]: I0227 09:00:05.238592 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-x255x_20c3f793-4286-4067-a372-37b2de08fa8e/ovs-vswitchd/0.log" Feb 27 09:00:05 crc kubenswrapper[4612]: I0227 09:00:05.341931 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-z7f9r_1ef66a51-46e7-4b76-a70b-b475d47822fd/ovn-controller/0.log" Feb 27 09:00:05 crc kubenswrapper[4612]: I0227 09:00:05.522531 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-xhcrn_dfdf6a2b-0ca5-4ea9-b87c-fd7a61f4981c/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Feb 27 09:00:05 crc kubenswrapper[4612]: I0227 09:00:05.598396 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_44e7bca8-b600-4ae0-8205-28f110f82dae/openstack-network-exporter/0.log" Feb 27 09:00:05 crc kubenswrapper[4612]: I0227 09:00:05.690411 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_44e7bca8-b600-4ae0-8205-28f110f82dae/ovn-northd/0.log" Feb 27 09:00:05 crc kubenswrapper[4612]: I0227 09:00:05.797181 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_1336636f-bc7a-45d4-a2ee-886a6f64ecd3/openstack-network-exporter/0.log" Feb 27 09:00:05 crc kubenswrapper[4612]: I0227 09:00:05.888325 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_1336636f-bc7a-45d4-a2ee-886a6f64ecd3/ovsdbserver-nb/0.log" Feb 27 09:00:05 crc kubenswrapper[4612]: I0227 09:00:05.969682 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536380-wkqc6" event={"ID":"45882240-d427-4e23-9415-100cbdcedb42","Type":"ContainerStarted","Data":"265cc3d5c79c7905f996507b86bb705ce59119077ceb548e02cc90ec3327c871"} Feb 27 09:00:05 crc kubenswrapper[4612]: I0227 09:00:05.989419 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29536380-wkqc6" podStartSLOduration=2.644306862 podStartE2EDuration="5.989397773s" podCreationTimestamp="2026-02-27 09:00:00 +0000 UTC" firstStartedPulling="2026-02-27 09:00:01.411542081 +0000 UTC m=+4259.265472069" lastFinishedPulling="2026-02-27 09:00:04.756632972 +0000 UTC m=+4262.610562980" observedRunningTime="2026-02-27 09:00:05.982707603 +0000 UTC m=+4263.836637601" watchObservedRunningTime="2026-02-27 09:00:05.989397773 +0000 UTC m=+4263.843327771" Feb 27 09:00:06 crc kubenswrapper[4612]: I0227 09:00:06.057558 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_ddac95c9-8d81-420b-836e-d1902ce8beb5/openstack-network-exporter/0.log" Feb 27 09:00:06 crc kubenswrapper[4612]: I0227 09:00:06.076281 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_ddac95c9-8d81-420b-836e-d1902ce8beb5/ovsdbserver-sb/0.log" Feb 27 09:00:06 crc kubenswrapper[4612]: I0227 09:00:06.158951 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-7d477cc4b6-5qtkx_5fb13c78-0dab-4785-90bc-21d5dabe81f6/placement-api/0.log" Feb 27 09:00:06 crc kubenswrapper[4612]: I0227 09:00:06.285267 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_40399d37-c675-448e-8b7e-eb50b5ed5629/setup-container/0.log" Feb 27 09:00:06 crc kubenswrapper[4612]: I0227 09:00:06.410855 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-7d477cc4b6-5qtkx_5fb13c78-0dab-4785-90bc-21d5dabe81f6/placement-log/0.log" Feb 27 09:00:06 crc kubenswrapper[4612]: I0227 09:00:06.663876 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_40399d37-c675-448e-8b7e-eb50b5ed5629/setup-container/0.log" Feb 27 09:00:06 crc kubenswrapper[4612]: I0227 09:00:06.666762 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_40399d37-c675-448e-8b7e-eb50b5ed5629/rabbitmq/0.log" Feb 27 09:00:06 crc kubenswrapper[4612]: I0227 09:00:06.705732 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_ba8e827f-294b-4532-8636-c81eeb511ba4/setup-container/0.log" Feb 27 09:00:06 crc kubenswrapper[4612]: I0227 09:00:06.965578 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_ba8e827f-294b-4532-8636-c81eeb511ba4/setup-container/0.log" Feb 27 09:00:06 crc kubenswrapper[4612]: I0227 09:00:06.993175 4612 generic.go:334] "Generic (PLEG): container finished" podID="45882240-d427-4e23-9415-100cbdcedb42" containerID="265cc3d5c79c7905f996507b86bb705ce59119077ceb548e02cc90ec3327c871" exitCode=0 Feb 27 09:00:06 crc kubenswrapper[4612]: I0227 09:00:06.993217 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536380-wkqc6" event={"ID":"45882240-d427-4e23-9415-100cbdcedb42","Type":"ContainerDied","Data":"265cc3d5c79c7905f996507b86bb705ce59119077ceb548e02cc90ec3327c871"} Feb 27 09:00:06 crc kubenswrapper[4612]: I0227 09:00:06.998400 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_ba8e827f-294b-4532-8636-c81eeb511ba4/rabbitmq/0.log" Feb 27 09:00:07 crc kubenswrapper[4612]: I0227 09:00:07.037160 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-cl6sd_ac2d7b40-cd87-4726-8225-6c9ce8afb4d7/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Feb 27 09:00:07 crc kubenswrapper[4612]: I0227 09:00:07.184439 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-vz4qv_9995bc24-fa2b-4916-8e71-627b11a39de7/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Feb 27 09:00:07 crc kubenswrapper[4612]: I0227 09:00:07.259721 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-g99jn_ab3a1ccc-dae6-40e3-859e-e4c0997b9071/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Feb 27 09:00:07 crc kubenswrapper[4612]: I0227 09:00:07.371973 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-psmxx_340ef584-802e-4696-b1b2-c7c8be815482/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Feb 27 09:00:07 crc kubenswrapper[4612]: I0227 09:00:07.544944 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-2g86c_d99b6ff5-a143-40ce-bee0-d4300122aff2/ssh-known-hosts-edpm-deployment/0.log" Feb 27 09:00:07 crc kubenswrapper[4612]: I0227 09:00:07.680667 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-59ddd57cfc-qvv4h_2e631b0d-d44b-472b-90d9-6794cf71d3fd/proxy-httpd/0.log" Feb 27 09:00:07 crc kubenswrapper[4612]: I0227 09:00:07.701017 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-59ddd57cfc-qvv4h_2e631b0d-d44b-472b-90d9-6794cf71d3fd/proxy-server/0.log" Feb 27 09:00:08 crc kubenswrapper[4612]: I0227 09:00:08.163535 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-gp8mt_76db835b-ff4b-4d34-9578-38c5e3a605ef/swift-ring-rebalance/0.log" Feb 27 09:00:08 crc kubenswrapper[4612]: I0227 09:00:08.312281 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_77288860-cf3c-47e6-a752-5b41aa1967fa/account-auditor/0.log" Feb 27 09:00:08 crc kubenswrapper[4612]: I0227 09:00:08.378525 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_77288860-cf3c-47e6-a752-5b41aa1967fa/account-reaper/0.log" Feb 27 09:00:08 crc kubenswrapper[4612]: I0227 09:00:08.429506 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536380-wkqc6" Feb 27 09:00:08 crc kubenswrapper[4612]: I0227 09:00:08.451495 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_77288860-cf3c-47e6-a752-5b41aa1967fa/account-replicator/0.log" Feb 27 09:00:08 crc kubenswrapper[4612]: I0227 09:00:08.483027 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zg877\" (UniqueName: \"kubernetes.io/projected/45882240-d427-4e23-9415-100cbdcedb42-kube-api-access-zg877\") pod \"45882240-d427-4e23-9415-100cbdcedb42\" (UID: \"45882240-d427-4e23-9415-100cbdcedb42\") " Feb 27 09:00:08 crc kubenswrapper[4612]: I0227 09:00:08.504353 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/45882240-d427-4e23-9415-100cbdcedb42-kube-api-access-zg877" (OuterVolumeSpecName: "kube-api-access-zg877") pod "45882240-d427-4e23-9415-100cbdcedb42" (UID: "45882240-d427-4e23-9415-100cbdcedb42"). InnerVolumeSpecName "kube-api-access-zg877". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 09:00:08 crc kubenswrapper[4612]: I0227 09:00:08.585716 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zg877\" (UniqueName: \"kubernetes.io/projected/45882240-d427-4e23-9415-100cbdcedb42-kube-api-access-zg877\") on node \"crc\" DevicePath \"\"" Feb 27 09:00:08 crc kubenswrapper[4612]: I0227 09:00:08.595487 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_77288860-cf3c-47e6-a752-5b41aa1967fa/account-server/0.log" Feb 27 09:00:08 crc kubenswrapper[4612]: I0227 09:00:08.612826 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_77288860-cf3c-47e6-a752-5b41aa1967fa/container-auditor/0.log" Feb 27 09:00:08 crc kubenswrapper[4612]: I0227 09:00:08.676344 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_77288860-cf3c-47e6-a752-5b41aa1967fa/container-replicator/0.log" Feb 27 09:00:08 crc kubenswrapper[4612]: I0227 09:00:08.771780 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_77288860-cf3c-47e6-a752-5b41aa1967fa/container-updater/0.log" Feb 27 09:00:08 crc kubenswrapper[4612]: I0227 09:00:08.774479 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_77288860-cf3c-47e6-a752-5b41aa1967fa/container-server/0.log" Feb 27 09:00:08 crc kubenswrapper[4612]: I0227 09:00:08.879841 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_77288860-cf3c-47e6-a752-5b41aa1967fa/object-expirer/0.log" Feb 27 09:00:08 crc kubenswrapper[4612]: I0227 09:00:08.899419 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_77288860-cf3c-47e6-a752-5b41aa1967fa/object-auditor/0.log" Feb 27 09:00:08 crc kubenswrapper[4612]: I0227 09:00:08.958547 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_77288860-cf3c-47e6-a752-5b41aa1967fa/object-replicator/0.log" Feb 27 09:00:09 crc kubenswrapper[4612]: I0227 09:00:09.011295 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536380-wkqc6" event={"ID":"45882240-d427-4e23-9415-100cbdcedb42","Type":"ContainerDied","Data":"d2edb1c26968147782b91586f4b48187e203dab06bb91a5559b184e37703fcd3"} Feb 27 09:00:09 crc kubenswrapper[4612]: I0227 09:00:09.011330 4612 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d2edb1c26968147782b91586f4b48187e203dab06bb91a5559b184e37703fcd3" Feb 27 09:00:09 crc kubenswrapper[4612]: I0227 09:00:09.011368 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536380-wkqc6" Feb 27 09:00:09 crc kubenswrapper[4612]: I0227 09:00:09.054207 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29536374-dx6vb"] Feb 27 09:00:09 crc kubenswrapper[4612]: I0227 09:00:09.063680 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_77288860-cf3c-47e6-a752-5b41aa1967fa/object-server/0.log" Feb 27 09:00:09 crc kubenswrapper[4612]: I0227 09:00:09.068866 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29536374-dx6vb"] Feb 27 09:00:09 crc kubenswrapper[4612]: I0227 09:00:09.132647 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_77288860-cf3c-47e6-a752-5b41aa1967fa/swift-recon-cron/0.log" Feb 27 09:00:09 crc kubenswrapper[4612]: I0227 09:00:09.137130 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_77288860-cf3c-47e6-a752-5b41aa1967fa/rsync/0.log" Feb 27 09:00:09 crc kubenswrapper[4612]: I0227 09:00:09.172712 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_77288860-cf3c-47e6-a752-5b41aa1967fa/object-updater/0.log" Feb 27 09:00:09 crc kubenswrapper[4612]: I0227 09:00:09.376754 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-qthr6_f69e6377-1d38-4128-b6ad-af91cf8f7793/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Feb 27 09:00:09 crc kubenswrapper[4612]: I0227 09:00:09.440301 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_4700642e-da3a-4e53-84de-0bf4c1870ccd/tempest-tests-tempest-tests-runner/0.log" Feb 27 09:00:09 crc kubenswrapper[4612]: I0227 09:00:09.583031 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_8aad3c36-76d6-4991-9494-0725b3e67aea/test-operator-logs-container/0.log" Feb 27 09:00:09 crc kubenswrapper[4612]: I0227 09:00:09.712382 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-vfpr6_ff510046-7399-4fdd-9880-67f6a51b61db/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Feb 27 09:00:10 crc kubenswrapper[4612]: I0227 09:00:10.863258 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ce53ef57-b5de-4fe6-adde-10bbec8f696d" path="/var/lib/kubelet/pods/ce53ef57-b5de-4fe6-adde-10bbec8f696d/volumes" Feb 27 09:00:16 crc kubenswrapper[4612]: I0227 09:00:16.027667 4612 patch_prober.go:28] interesting pod/machine-config-daemon-924vb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 27 09:00:16 crc kubenswrapper[4612]: I0227 09:00:16.028273 4612 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 27 09:00:19 crc kubenswrapper[4612]: I0227 09:00:19.941236 4612 scope.go:117] "RemoveContainer" containerID="b82d7cb0ab80c5384a935820b9c397496afec22701011ee8fa8e4fa29ab371cf" Feb 27 09:00:20 crc kubenswrapper[4612]: I0227 09:00:20.016336 4612 scope.go:117] "RemoveContainer" containerID="c749d475d5e71926c5a54a3210fb4522047e7ef2624c67d3201f8856bfcb8b67" Feb 27 09:00:40 crc kubenswrapper[4612]: I0227 09:00:40.555064 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_1e2105768d85fd59d354c5be43e261a5f6290b57040396b6c3a32dc181djh99_dfad17e6-8a21-44be-8328-2a156da54578/util/0.log" Feb 27 09:00:40 crc kubenswrapper[4612]: I0227 09:00:40.801360 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_1e2105768d85fd59d354c5be43e261a5f6290b57040396b6c3a32dc181djh99_dfad17e6-8a21-44be-8328-2a156da54578/pull/0.log" Feb 27 09:00:40 crc kubenswrapper[4612]: I0227 09:00:40.873503 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_1e2105768d85fd59d354c5be43e261a5f6290b57040396b6c3a32dc181djh99_dfad17e6-8a21-44be-8328-2a156da54578/util/0.log" Feb 27 09:00:40 crc kubenswrapper[4612]: I0227 09:00:40.883311 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_1e2105768d85fd59d354c5be43e261a5f6290b57040396b6c3a32dc181djh99_dfad17e6-8a21-44be-8328-2a156da54578/pull/0.log" Feb 27 09:00:41 crc kubenswrapper[4612]: I0227 09:00:41.129083 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_1e2105768d85fd59d354c5be43e261a5f6290b57040396b6c3a32dc181djh99_dfad17e6-8a21-44be-8328-2a156da54578/extract/0.log" Feb 27 09:00:41 crc kubenswrapper[4612]: I0227 09:00:41.129938 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_1e2105768d85fd59d354c5be43e261a5f6290b57040396b6c3a32dc181djh99_dfad17e6-8a21-44be-8328-2a156da54578/pull/0.log" Feb 27 09:00:41 crc kubenswrapper[4612]: I0227 09:00:41.181171 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_1e2105768d85fd59d354c5be43e261a5f6290b57040396b6c3a32dc181djh99_dfad17e6-8a21-44be-8328-2a156da54578/util/0.log" Feb 27 09:00:41 crc kubenswrapper[4612]: I0227 09:00:41.751389 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-6d8bf5c495-t95hp_63d519c2-d2f9-4349-ad82-0399e472d78a/manager/0.log" Feb 27 09:00:42 crc kubenswrapper[4612]: I0227 09:00:42.153831 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-784b5bb6c5-g4p5v_71c482bc-7818-4193-a80b-323682cdcfb3/manager/0.log" Feb 27 09:00:42 crc kubenswrapper[4612]: I0227 09:00:42.378372 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-69f49c598c-7f9bj_6065186a-5435-4db3-8adc-698bec4358ef/manager/0.log" Feb 27 09:00:42 crc kubenswrapper[4612]: I0227 09:00:42.604978 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-5b9b8895d5-t4rct_1903598c-2428-4e90-ab81-9a47fa5af9ef/manager/0.log" Feb 27 09:00:43 crc kubenswrapper[4612]: I0227 09:00:43.279165 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-554564d7fc-8kpwf_bd687bfa-c843-4d11-80f9-cedcdce41e3d/manager/0.log" Feb 27 09:00:43 crc kubenswrapper[4612]: I0227 09:00:43.381477 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-79d975b745-plm9k_98296d45-b487-47c3-a44a-fa63ba52426b/manager/0.log" Feb 27 09:00:43 crc kubenswrapper[4612]: I0227 09:00:43.576070 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-55d77d7b5c-462m7_36fc734e-4b3e-46bb-b914-896b1ec5c2ee/manager/0.log" Feb 27 09:00:43 crc kubenswrapper[4612]: I0227 09:00:43.833944 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-67d996989d-4dhhw_b60d4209-cd53-4cfe-99f7-7e7351041857/manager/0.log" Feb 27 09:00:43 crc kubenswrapper[4612]: I0227 09:00:43.845376 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-b4d948c87-2hm5l_259df4bd-af1b-4c61-a5bd-0ca28ac531d4/manager/0.log" Feb 27 09:00:44 crc kubenswrapper[4612]: I0227 09:00:44.294538 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-6994f66f48-qs4l7_f62fea1f-71de-4872-9baa-42ad2448d5f5/manager/0.log" Feb 27 09:00:44 crc kubenswrapper[4612]: I0227 09:00:44.484861 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-6bd4687957-2qmzb_f292b2e2-2b32-4e4e-bfca-d5db60f0b358/manager/0.log" Feb 27 09:00:44 crc kubenswrapper[4612]: I0227 09:00:44.653279 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-567668f5cf-tjqwp_d84adbda-b53c-46c7-9196-682910a0009c/manager/0.log" Feb 27 09:00:44 crc kubenswrapper[4612]: I0227 09:00:44.833935 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-659dc6bbfc-vlczg_bd923861-d2ad-4a06-b0c7-613605afefbb/manager/0.log" Feb 27 09:00:44 crc kubenswrapper[4612]: I0227 09:00:44.989203 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-7c6767dc9clp9dx_cf5ef40d-5e5c-4b2c-9570-63e552800b4e/manager/0.log" Feb 27 09:00:45 crc kubenswrapper[4612]: I0227 09:00:45.260626 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-init-8667ddcbf9-7stg7_a8e2b07a-a4cf-4a98-afeb-2a7554b687ee/operator/0.log" Feb 27 09:00:45 crc kubenswrapper[4612]: I0227 09:00:45.477877 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-f9cqc_5ed1acaa-e783-4d8d-bac0-adf0ab4a3b0f/registry-server/0.log" Feb 27 09:00:45 crc kubenswrapper[4612]: I0227 09:00:45.619581 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-5955d8c787-q74ht_e0304d11-114b-4aa3-86ca-1747c60b5bfc/manager/0.log" Feb 27 09:00:45 crc kubenswrapper[4612]: I0227 09:00:45.802660 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-8497b45c89-dvq8j_8414b96c-7e5b-4824-998e-549a0c43d9d5/manager/0.log" Feb 27 09:00:46 crc kubenswrapper[4612]: I0227 09:00:46.029897 4612 patch_prober.go:28] interesting pod/machine-config-daemon-924vb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 27 09:00:46 crc kubenswrapper[4612]: I0227 09:00:46.029980 4612 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 27 09:00:46 crc kubenswrapper[4612]: I0227 09:00:46.333528 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-8s6n2_57ba67ae-3e41-4e2b-9ee6-30b2643805dd/operator/0.log" Feb 27 09:00:46 crc kubenswrapper[4612]: I0227 09:00:46.428341 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-68f46476f-r8lqs_713a994e-490d-47b4-86ab-1203a6b806fc/manager/0.log" Feb 27 09:00:47 crc kubenswrapper[4612]: I0227 09:00:47.042420 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-589c568786-qclqp_ee3f21f6-2f91-45c6-8c80-5fb8f89a41ce/manager/0.log" Feb 27 09:00:47 crc kubenswrapper[4612]: I0227 09:00:47.140084 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5dc6794d5b-s2jnh_ad4fca59-6a87-4ebc-8681-d5a3a40f0b82/manager/0.log" Feb 27 09:00:47 crc kubenswrapper[4612]: I0227 09:00:47.193894 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-69f9884c8f-v4kkz_41aed211-3a1f-4f7c-86ca-53e445dcd585/manager/0.log" Feb 27 09:00:47 crc kubenswrapper[4612]: I0227 09:00:47.302254 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-bccc79885-zdtzv_a4619f7b-7242-459b-8558-f84c2893ee55/manager/0.log" Feb 27 09:00:50 crc kubenswrapper[4612]: I0227 09:00:50.779202 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-868647ff47-mlgz7_2ddc15ea-b7c1-44e7-b795-392890e9c14e/manager/0.log" Feb 27 09:01:00 crc kubenswrapper[4612]: I0227 09:01:00.159999 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29536381-vc87r"] Feb 27 09:01:00 crc kubenswrapper[4612]: E0227 09:01:00.160741 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="45882240-d427-4e23-9415-100cbdcedb42" containerName="oc" Feb 27 09:01:00 crc kubenswrapper[4612]: I0227 09:01:00.160753 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="45882240-d427-4e23-9415-100cbdcedb42" containerName="oc" Feb 27 09:01:00 crc kubenswrapper[4612]: E0227 09:01:00.160769 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e1e5befa-cd14-4bb6-82be-547ace37b1d7" containerName="collect-profiles" Feb 27 09:01:00 crc kubenswrapper[4612]: I0227 09:01:00.160774 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="e1e5befa-cd14-4bb6-82be-547ace37b1d7" containerName="collect-profiles" Feb 27 09:01:00 crc kubenswrapper[4612]: I0227 09:01:00.160963 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="45882240-d427-4e23-9415-100cbdcedb42" containerName="oc" Feb 27 09:01:00 crc kubenswrapper[4612]: I0227 09:01:00.160977 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="e1e5befa-cd14-4bb6-82be-547ace37b1d7" containerName="collect-profiles" Feb 27 09:01:00 crc kubenswrapper[4612]: I0227 09:01:00.161560 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29536381-vc87r" Feb 27 09:01:00 crc kubenswrapper[4612]: I0227 09:01:00.169784 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29536381-vc87r"] Feb 27 09:01:00 crc kubenswrapper[4612]: I0227 09:01:00.248462 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p546f\" (UniqueName: \"kubernetes.io/projected/7aa90d53-2716-4f3f-aacf-fe667ce25f5d-kube-api-access-p546f\") pod \"keystone-cron-29536381-vc87r\" (UID: \"7aa90d53-2716-4f3f-aacf-fe667ce25f5d\") " pod="openstack/keystone-cron-29536381-vc87r" Feb 27 09:01:00 crc kubenswrapper[4612]: I0227 09:01:00.248892 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7aa90d53-2716-4f3f-aacf-fe667ce25f5d-combined-ca-bundle\") pod \"keystone-cron-29536381-vc87r\" (UID: \"7aa90d53-2716-4f3f-aacf-fe667ce25f5d\") " pod="openstack/keystone-cron-29536381-vc87r" Feb 27 09:01:00 crc kubenswrapper[4612]: I0227 09:01:00.249112 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/7aa90d53-2716-4f3f-aacf-fe667ce25f5d-fernet-keys\") pod \"keystone-cron-29536381-vc87r\" (UID: \"7aa90d53-2716-4f3f-aacf-fe667ce25f5d\") " pod="openstack/keystone-cron-29536381-vc87r" Feb 27 09:01:00 crc kubenswrapper[4612]: I0227 09:01:00.249182 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7aa90d53-2716-4f3f-aacf-fe667ce25f5d-config-data\") pod \"keystone-cron-29536381-vc87r\" (UID: \"7aa90d53-2716-4f3f-aacf-fe667ce25f5d\") " pod="openstack/keystone-cron-29536381-vc87r" Feb 27 09:01:00 crc kubenswrapper[4612]: I0227 09:01:00.351180 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p546f\" (UniqueName: \"kubernetes.io/projected/7aa90d53-2716-4f3f-aacf-fe667ce25f5d-kube-api-access-p546f\") pod \"keystone-cron-29536381-vc87r\" (UID: \"7aa90d53-2716-4f3f-aacf-fe667ce25f5d\") " pod="openstack/keystone-cron-29536381-vc87r" Feb 27 09:01:00 crc kubenswrapper[4612]: I0227 09:01:00.351234 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7aa90d53-2716-4f3f-aacf-fe667ce25f5d-combined-ca-bundle\") pod \"keystone-cron-29536381-vc87r\" (UID: \"7aa90d53-2716-4f3f-aacf-fe667ce25f5d\") " pod="openstack/keystone-cron-29536381-vc87r" Feb 27 09:01:00 crc kubenswrapper[4612]: I0227 09:01:00.351340 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/7aa90d53-2716-4f3f-aacf-fe667ce25f5d-fernet-keys\") pod \"keystone-cron-29536381-vc87r\" (UID: \"7aa90d53-2716-4f3f-aacf-fe667ce25f5d\") " pod="openstack/keystone-cron-29536381-vc87r" Feb 27 09:01:00 crc kubenswrapper[4612]: I0227 09:01:00.351388 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7aa90d53-2716-4f3f-aacf-fe667ce25f5d-config-data\") pod \"keystone-cron-29536381-vc87r\" (UID: \"7aa90d53-2716-4f3f-aacf-fe667ce25f5d\") " pod="openstack/keystone-cron-29536381-vc87r" Feb 27 09:01:00 crc kubenswrapper[4612]: I0227 09:01:00.368811 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7aa90d53-2716-4f3f-aacf-fe667ce25f5d-combined-ca-bundle\") pod \"keystone-cron-29536381-vc87r\" (UID: \"7aa90d53-2716-4f3f-aacf-fe667ce25f5d\") " pod="openstack/keystone-cron-29536381-vc87r" Feb 27 09:01:00 crc kubenswrapper[4612]: I0227 09:01:00.368891 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/7aa90d53-2716-4f3f-aacf-fe667ce25f5d-fernet-keys\") pod \"keystone-cron-29536381-vc87r\" (UID: \"7aa90d53-2716-4f3f-aacf-fe667ce25f5d\") " pod="openstack/keystone-cron-29536381-vc87r" Feb 27 09:01:00 crc kubenswrapper[4612]: I0227 09:01:00.369359 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7aa90d53-2716-4f3f-aacf-fe667ce25f5d-config-data\") pod \"keystone-cron-29536381-vc87r\" (UID: \"7aa90d53-2716-4f3f-aacf-fe667ce25f5d\") " pod="openstack/keystone-cron-29536381-vc87r" Feb 27 09:01:00 crc kubenswrapper[4612]: I0227 09:01:00.373127 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p546f\" (UniqueName: \"kubernetes.io/projected/7aa90d53-2716-4f3f-aacf-fe667ce25f5d-kube-api-access-p546f\") pod \"keystone-cron-29536381-vc87r\" (UID: \"7aa90d53-2716-4f3f-aacf-fe667ce25f5d\") " pod="openstack/keystone-cron-29536381-vc87r" Feb 27 09:01:00 crc kubenswrapper[4612]: I0227 09:01:00.508358 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29536381-vc87r" Feb 27 09:01:00 crc kubenswrapper[4612]: I0227 09:01:00.990822 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29536381-vc87r"] Feb 27 09:01:01 crc kubenswrapper[4612]: I0227 09:01:01.475193 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29536381-vc87r" event={"ID":"7aa90d53-2716-4f3f-aacf-fe667ce25f5d","Type":"ContainerStarted","Data":"6aa221cdf2cc2c49b0e1da279ea4556e40695d8a821ebbdbd223b5e3b0bf6e30"} Feb 27 09:01:01 crc kubenswrapper[4612]: I0227 09:01:01.475231 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29536381-vc87r" event={"ID":"7aa90d53-2716-4f3f-aacf-fe667ce25f5d","Type":"ContainerStarted","Data":"1b9de4349331b9b4a6e828803a4335bbdc99eaa5fb06bd0ed64a56e40d574112"} Feb 27 09:01:01 crc kubenswrapper[4612]: I0227 09:01:01.496841 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29536381-vc87r" podStartSLOduration=1.496820152 podStartE2EDuration="1.496820152s" podCreationTimestamp="2026-02-27 09:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 09:01:01.488816155 +0000 UTC m=+4319.342746153" watchObservedRunningTime="2026-02-27 09:01:01.496820152 +0000 UTC m=+4319.350750140" Feb 27 09:01:05 crc kubenswrapper[4612]: I0227 09:01:05.507506 4612 generic.go:334] "Generic (PLEG): container finished" podID="7aa90d53-2716-4f3f-aacf-fe667ce25f5d" containerID="6aa221cdf2cc2c49b0e1da279ea4556e40695d8a821ebbdbd223b5e3b0bf6e30" exitCode=0 Feb 27 09:01:05 crc kubenswrapper[4612]: I0227 09:01:05.507600 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29536381-vc87r" event={"ID":"7aa90d53-2716-4f3f-aacf-fe667ce25f5d","Type":"ContainerDied","Data":"6aa221cdf2cc2c49b0e1da279ea4556e40695d8a821ebbdbd223b5e3b0bf6e30"} Feb 27 09:01:06 crc kubenswrapper[4612]: I0227 09:01:06.906629 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29536381-vc87r" Feb 27 09:01:06 crc kubenswrapper[4612]: I0227 09:01:06.977582 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7aa90d53-2716-4f3f-aacf-fe667ce25f5d-config-data\") pod \"7aa90d53-2716-4f3f-aacf-fe667ce25f5d\" (UID: \"7aa90d53-2716-4f3f-aacf-fe667ce25f5d\") " Feb 27 09:01:06 crc kubenswrapper[4612]: I0227 09:01:06.977819 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7aa90d53-2716-4f3f-aacf-fe667ce25f5d-combined-ca-bundle\") pod \"7aa90d53-2716-4f3f-aacf-fe667ce25f5d\" (UID: \"7aa90d53-2716-4f3f-aacf-fe667ce25f5d\") " Feb 27 09:01:06 crc kubenswrapper[4612]: I0227 09:01:06.978007 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/7aa90d53-2716-4f3f-aacf-fe667ce25f5d-fernet-keys\") pod \"7aa90d53-2716-4f3f-aacf-fe667ce25f5d\" (UID: \"7aa90d53-2716-4f3f-aacf-fe667ce25f5d\") " Feb 27 09:01:06 crc kubenswrapper[4612]: I0227 09:01:06.978166 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p546f\" (UniqueName: \"kubernetes.io/projected/7aa90d53-2716-4f3f-aacf-fe667ce25f5d-kube-api-access-p546f\") pod \"7aa90d53-2716-4f3f-aacf-fe667ce25f5d\" (UID: \"7aa90d53-2716-4f3f-aacf-fe667ce25f5d\") " Feb 27 09:01:06 crc kubenswrapper[4612]: I0227 09:01:06.986182 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7aa90d53-2716-4f3f-aacf-fe667ce25f5d-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "7aa90d53-2716-4f3f-aacf-fe667ce25f5d" (UID: "7aa90d53-2716-4f3f-aacf-fe667ce25f5d"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 09:01:06 crc kubenswrapper[4612]: I0227 09:01:06.986528 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7aa90d53-2716-4f3f-aacf-fe667ce25f5d-kube-api-access-p546f" (OuterVolumeSpecName: "kube-api-access-p546f") pod "7aa90d53-2716-4f3f-aacf-fe667ce25f5d" (UID: "7aa90d53-2716-4f3f-aacf-fe667ce25f5d"). InnerVolumeSpecName "kube-api-access-p546f". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 09:01:07 crc kubenswrapper[4612]: I0227 09:01:07.031851 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7aa90d53-2716-4f3f-aacf-fe667ce25f5d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7aa90d53-2716-4f3f-aacf-fe667ce25f5d" (UID: "7aa90d53-2716-4f3f-aacf-fe667ce25f5d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 09:01:07 crc kubenswrapper[4612]: I0227 09:01:07.070075 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7aa90d53-2716-4f3f-aacf-fe667ce25f5d-config-data" (OuterVolumeSpecName: "config-data") pod "7aa90d53-2716-4f3f-aacf-fe667ce25f5d" (UID: "7aa90d53-2716-4f3f-aacf-fe667ce25f5d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 09:01:07 crc kubenswrapper[4612]: I0227 09:01:07.080786 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p546f\" (UniqueName: \"kubernetes.io/projected/7aa90d53-2716-4f3f-aacf-fe667ce25f5d-kube-api-access-p546f\") on node \"crc\" DevicePath \"\"" Feb 27 09:01:07 crc kubenswrapper[4612]: I0227 09:01:07.080817 4612 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7aa90d53-2716-4f3f-aacf-fe667ce25f5d-config-data\") on node \"crc\" DevicePath \"\"" Feb 27 09:01:07 crc kubenswrapper[4612]: I0227 09:01:07.080827 4612 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7aa90d53-2716-4f3f-aacf-fe667ce25f5d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 27 09:01:07 crc kubenswrapper[4612]: I0227 09:01:07.080834 4612 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/7aa90d53-2716-4f3f-aacf-fe667ce25f5d-fernet-keys\") on node \"crc\" DevicePath \"\"" Feb 27 09:01:07 crc kubenswrapper[4612]: I0227 09:01:07.527455 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29536381-vc87r" event={"ID":"7aa90d53-2716-4f3f-aacf-fe667ce25f5d","Type":"ContainerDied","Data":"1b9de4349331b9b4a6e828803a4335bbdc99eaa5fb06bd0ed64a56e40d574112"} Feb 27 09:01:07 crc kubenswrapper[4612]: I0227 09:01:07.527730 4612 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1b9de4349331b9b4a6e828803a4335bbdc99eaa5fb06bd0ed64a56e40d574112" Feb 27 09:01:07 crc kubenswrapper[4612]: I0227 09:01:07.527500 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29536381-vc87r" Feb 27 09:01:09 crc kubenswrapper[4612]: I0227 09:01:09.962745 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-zpshm_d951f066-9e39-4e0a-9fa3-da9c868552d1/control-plane-machine-set-operator/0.log" Feb 27 09:01:10 crc kubenswrapper[4612]: I0227 09:01:10.097925 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-q6h7h_094b001c-e80c-4d07-a161-e00514667462/kube-rbac-proxy/0.log" Feb 27 09:01:10 crc kubenswrapper[4612]: I0227 09:01:10.208542 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-q6h7h_094b001c-e80c-4d07-a161-e00514667462/machine-api-operator/0.log" Feb 27 09:01:16 crc kubenswrapper[4612]: I0227 09:01:16.089488 4612 patch_prober.go:28] interesting pod/machine-config-daemon-924vb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 27 09:01:16 crc kubenswrapper[4612]: I0227 09:01:16.089989 4612 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 27 09:01:16 crc kubenswrapper[4612]: I0227 09:01:16.090029 4612 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-924vb" Feb 27 09:01:16 crc kubenswrapper[4612]: I0227 09:01:16.090746 4612 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"aa498fa4b5f3843852a1ebe0b165b06254525eeb49f5224c7bd2a0f8c2ae2480"} pod="openshift-machine-config-operator/machine-config-daemon-924vb" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 27 09:01:16 crc kubenswrapper[4612]: I0227 09:01:16.090801 4612 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" containerName="machine-config-daemon" containerID="cri-o://aa498fa4b5f3843852a1ebe0b165b06254525eeb49f5224c7bd2a0f8c2ae2480" gracePeriod=600 Feb 27 09:01:16 crc kubenswrapper[4612]: E0227 09:01:16.303302 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-924vb_openshift-machine-config-operator(24599373-7adc-4a1b-8bb4-797bf726f43d)\"" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" Feb 27 09:01:17 crc kubenswrapper[4612]: I0227 09:01:17.135332 4612 generic.go:334] "Generic (PLEG): container finished" podID="24599373-7adc-4a1b-8bb4-797bf726f43d" containerID="aa498fa4b5f3843852a1ebe0b165b06254525eeb49f5224c7bd2a0f8c2ae2480" exitCode=0 Feb 27 09:01:17 crc kubenswrapper[4612]: I0227 09:01:17.135373 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-924vb" event={"ID":"24599373-7adc-4a1b-8bb4-797bf726f43d","Type":"ContainerDied","Data":"aa498fa4b5f3843852a1ebe0b165b06254525eeb49f5224c7bd2a0f8c2ae2480"} Feb 27 09:01:17 crc kubenswrapper[4612]: I0227 09:01:17.135443 4612 scope.go:117] "RemoveContainer" containerID="cebeb2cb354fa601f37e4c7cc40bcc1dcd010ada59d86d91448073fa6deac102" Feb 27 09:01:17 crc kubenswrapper[4612]: I0227 09:01:17.136302 4612 scope.go:117] "RemoveContainer" containerID="aa498fa4b5f3843852a1ebe0b165b06254525eeb49f5224c7bd2a0f8c2ae2480" Feb 27 09:01:17 crc kubenswrapper[4612]: E0227 09:01:17.136787 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-924vb_openshift-machine-config-operator(24599373-7adc-4a1b-8bb4-797bf726f43d)\"" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" Feb 27 09:01:25 crc kubenswrapper[4612]: I0227 09:01:25.575248 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-858654f9db-lq868_095685d4-89ee-421e-ba88-a2df75c3a160/cert-manager-controller/0.log" Feb 27 09:01:25 crc kubenswrapper[4612]: I0227 09:01:25.707652 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-cf98fcc89-gk4r5_4df7eb22-d428-4e88-bdc5-0bab2b4806a3/cert-manager-cainjector/0.log" Feb 27 09:01:25 crc kubenswrapper[4612]: I0227 09:01:25.812442 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-687f57d79b-krsbj_5bba164a-78bf-4a9a-9147-729a1a91740c/cert-manager-webhook/0.log" Feb 27 09:01:29 crc kubenswrapper[4612]: I0227 09:01:29.853370 4612 scope.go:117] "RemoveContainer" containerID="aa498fa4b5f3843852a1ebe0b165b06254525eeb49f5224c7bd2a0f8c2ae2480" Feb 27 09:01:29 crc kubenswrapper[4612]: E0227 09:01:29.854204 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-924vb_openshift-machine-config-operator(24599373-7adc-4a1b-8bb4-797bf726f43d)\"" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" Feb 27 09:01:40 crc kubenswrapper[4612]: I0227 09:01:40.690161 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-5dcbbd79cf-4cfq5_a82cc418-4f77-4788-addb-25341d42d9fa/nmstate-console-plugin/0.log" Feb 27 09:01:40 crc kubenswrapper[4612]: I0227 09:01:40.839750 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-26zsp_5553e6c7-b45f-40c9-99ce-abd46767d31a/nmstate-handler/0.log" Feb 27 09:01:40 crc kubenswrapper[4612]: I0227 09:01:40.910423 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-69594cc75-czssc_d058174c-69c3-49cf-b846-af074b709972/kube-rbac-proxy/0.log" Feb 27 09:01:40 crc kubenswrapper[4612]: I0227 09:01:40.985705 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-69594cc75-czssc_d058174c-69c3-49cf-b846-af074b709972/nmstate-metrics/0.log" Feb 27 09:01:41 crc kubenswrapper[4612]: I0227 09:01:41.106081 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-75c5dccd6c-p6nbk_49fa2d59-6289-4edc-a6d8-2c3f7733a536/nmstate-operator/0.log" Feb 27 09:01:41 crc kubenswrapper[4612]: I0227 09:01:41.228464 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-786f45cff4-m8llg_b141dc9e-912e-4096-97fc-5df6e0ae245c/nmstate-webhook/0.log" Feb 27 09:01:42 crc kubenswrapper[4612]: I0227 09:01:42.867870 4612 scope.go:117] "RemoveContainer" containerID="aa498fa4b5f3843852a1ebe0b165b06254525eeb49f5224c7bd2a0f8c2ae2480" Feb 27 09:01:42 crc kubenswrapper[4612]: E0227 09:01:42.879317 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-924vb_openshift-machine-config-operator(24599373-7adc-4a1b-8bb4-797bf726f43d)\"" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" Feb 27 09:01:54 crc kubenswrapper[4612]: I0227 09:01:54.853662 4612 scope.go:117] "RemoveContainer" containerID="aa498fa4b5f3843852a1ebe0b165b06254525eeb49f5224c7bd2a0f8c2ae2480" Feb 27 09:01:54 crc kubenswrapper[4612]: E0227 09:01:54.854618 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-924vb_openshift-machine-config-operator(24599373-7adc-4a1b-8bb4-797bf726f43d)\"" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" Feb 27 09:02:00 crc kubenswrapper[4612]: I0227 09:02:00.160063 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29536382-2jd9v"] Feb 27 09:02:00 crc kubenswrapper[4612]: E0227 09:02:00.160892 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7aa90d53-2716-4f3f-aacf-fe667ce25f5d" containerName="keystone-cron" Feb 27 09:02:00 crc kubenswrapper[4612]: I0227 09:02:00.160906 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="7aa90d53-2716-4f3f-aacf-fe667ce25f5d" containerName="keystone-cron" Feb 27 09:02:00 crc kubenswrapper[4612]: I0227 09:02:00.161082 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="7aa90d53-2716-4f3f-aacf-fe667ce25f5d" containerName="keystone-cron" Feb 27 09:02:00 crc kubenswrapper[4612]: I0227 09:02:00.161651 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536382-2jd9v" Feb 27 09:02:00 crc kubenswrapper[4612]: I0227 09:02:00.163560 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-5zwfx" Feb 27 09:02:00 crc kubenswrapper[4612]: I0227 09:02:00.163762 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 27 09:02:00 crc kubenswrapper[4612]: I0227 09:02:00.163897 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 27 09:02:00 crc kubenswrapper[4612]: I0227 09:02:00.171752 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536382-2jd9v"] Feb 27 09:02:00 crc kubenswrapper[4612]: I0227 09:02:00.293967 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q95gz\" (UniqueName: \"kubernetes.io/projected/80a082d2-95dd-4d63-965f-8d79c871ad0e-kube-api-access-q95gz\") pod \"auto-csr-approver-29536382-2jd9v\" (UID: \"80a082d2-95dd-4d63-965f-8d79c871ad0e\") " pod="openshift-infra/auto-csr-approver-29536382-2jd9v" Feb 27 09:02:00 crc kubenswrapper[4612]: I0227 09:02:00.395506 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q95gz\" (UniqueName: \"kubernetes.io/projected/80a082d2-95dd-4d63-965f-8d79c871ad0e-kube-api-access-q95gz\") pod \"auto-csr-approver-29536382-2jd9v\" (UID: \"80a082d2-95dd-4d63-965f-8d79c871ad0e\") " pod="openshift-infra/auto-csr-approver-29536382-2jd9v" Feb 27 09:02:00 crc kubenswrapper[4612]: I0227 09:02:00.416585 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q95gz\" (UniqueName: \"kubernetes.io/projected/80a082d2-95dd-4d63-965f-8d79c871ad0e-kube-api-access-q95gz\") pod \"auto-csr-approver-29536382-2jd9v\" (UID: \"80a082d2-95dd-4d63-965f-8d79c871ad0e\") " pod="openshift-infra/auto-csr-approver-29536382-2jd9v" Feb 27 09:02:00 crc kubenswrapper[4612]: I0227 09:02:00.520728 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536382-2jd9v" Feb 27 09:02:00 crc kubenswrapper[4612]: I0227 09:02:00.996136 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536382-2jd9v"] Feb 27 09:02:01 crc kubenswrapper[4612]: I0227 09:02:01.022227 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536382-2jd9v" event={"ID":"80a082d2-95dd-4d63-965f-8d79c871ad0e","Type":"ContainerStarted","Data":"13d43b4ff366f6a7f41b760c4ef871fe8b289e14bd56b87d349ac6844d5380ca"} Feb 27 09:02:03 crc kubenswrapper[4612]: I0227 09:02:03.053802 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536382-2jd9v" event={"ID":"80a082d2-95dd-4d63-965f-8d79c871ad0e","Type":"ContainerStarted","Data":"89ca1cead2b1d034bd87ad1cb0b15782f88020d83a8b54fbedef5cfbb6857046"} Feb 27 09:02:03 crc kubenswrapper[4612]: I0227 09:02:03.078973 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29536382-2jd9v" podStartSLOduration=2.074030095 podStartE2EDuration="3.078951757s" podCreationTimestamp="2026-02-27 09:02:00 +0000 UTC" firstStartedPulling="2026-02-27 09:02:00.995606082 +0000 UTC m=+4378.849536080" lastFinishedPulling="2026-02-27 09:02:02.000527724 +0000 UTC m=+4379.854457742" observedRunningTime="2026-02-27 09:02:03.068806948 +0000 UTC m=+4380.922736966" watchObservedRunningTime="2026-02-27 09:02:03.078951757 +0000 UTC m=+4380.932881765" Feb 27 09:02:04 crc kubenswrapper[4612]: I0227 09:02:04.062209 4612 generic.go:334] "Generic (PLEG): container finished" podID="80a082d2-95dd-4d63-965f-8d79c871ad0e" containerID="89ca1cead2b1d034bd87ad1cb0b15782f88020d83a8b54fbedef5cfbb6857046" exitCode=0 Feb 27 09:02:04 crc kubenswrapper[4612]: I0227 09:02:04.062294 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536382-2jd9v" event={"ID":"80a082d2-95dd-4d63-965f-8d79c871ad0e","Type":"ContainerDied","Data":"89ca1cead2b1d034bd87ad1cb0b15782f88020d83a8b54fbedef5cfbb6857046"} Feb 27 09:02:05 crc kubenswrapper[4612]: I0227 09:02:05.381939 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536382-2jd9v" Feb 27 09:02:05 crc kubenswrapper[4612]: I0227 09:02:05.487678 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q95gz\" (UniqueName: \"kubernetes.io/projected/80a082d2-95dd-4d63-965f-8d79c871ad0e-kube-api-access-q95gz\") pod \"80a082d2-95dd-4d63-965f-8d79c871ad0e\" (UID: \"80a082d2-95dd-4d63-965f-8d79c871ad0e\") " Feb 27 09:02:05 crc kubenswrapper[4612]: I0227 09:02:05.502024 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/80a082d2-95dd-4d63-965f-8d79c871ad0e-kube-api-access-q95gz" (OuterVolumeSpecName: "kube-api-access-q95gz") pod "80a082d2-95dd-4d63-965f-8d79c871ad0e" (UID: "80a082d2-95dd-4d63-965f-8d79c871ad0e"). InnerVolumeSpecName "kube-api-access-q95gz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 09:02:05 crc kubenswrapper[4612]: I0227 09:02:05.589734 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q95gz\" (UniqueName: \"kubernetes.io/projected/80a082d2-95dd-4d63-965f-8d79c871ad0e-kube-api-access-q95gz\") on node \"crc\" DevicePath \"\"" Feb 27 09:02:05 crc kubenswrapper[4612]: I0227 09:02:05.940420 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29536376-p8z66"] Feb 27 09:02:05 crc kubenswrapper[4612]: I0227 09:02:05.951400 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29536376-p8z66"] Feb 27 09:02:06 crc kubenswrapper[4612]: I0227 09:02:06.082552 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536382-2jd9v" event={"ID":"80a082d2-95dd-4d63-965f-8d79c871ad0e","Type":"ContainerDied","Data":"13d43b4ff366f6a7f41b760c4ef871fe8b289e14bd56b87d349ac6844d5380ca"} Feb 27 09:02:06 crc kubenswrapper[4612]: I0227 09:02:06.082599 4612 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="13d43b4ff366f6a7f41b760c4ef871fe8b289e14bd56b87d349ac6844d5380ca" Feb 27 09:02:06 crc kubenswrapper[4612]: I0227 09:02:06.082611 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536382-2jd9v" Feb 27 09:02:06 crc kubenswrapper[4612]: I0227 09:02:06.862857 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="88ada8ea-dcbb-40c2-9599-b1fc0e3f24d6" path="/var/lib/kubelet/pods/88ada8ea-dcbb-40c2-9599-b1fc0e3f24d6/volumes" Feb 27 09:02:07 crc kubenswrapper[4612]: I0227 09:02:07.852548 4612 scope.go:117] "RemoveContainer" containerID="aa498fa4b5f3843852a1ebe0b165b06254525eeb49f5224c7bd2a0f8c2ae2480" Feb 27 09:02:07 crc kubenswrapper[4612]: E0227 09:02:07.853109 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-924vb_openshift-machine-config-operator(24599373-7adc-4a1b-8bb4-797bf726f43d)\"" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" Feb 27 09:02:11 crc kubenswrapper[4612]: I0227 09:02:11.375530 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-86ddb6bd46-gk8pf_2b49ec1b-e464-4e71-badd-04acfa13ef75/kube-rbac-proxy/0.log" Feb 27 09:02:11 crc kubenswrapper[4612]: I0227 09:02:11.553986 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-86ddb6bd46-gk8pf_2b49ec1b-e464-4e71-badd-04acfa13ef75/controller/0.log" Feb 27 09:02:11 crc kubenswrapper[4612]: I0227 09:02:11.643850 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-8drht_8fe63fdd-8b6f-4ecf-81d2-959ae060d567/cp-frr-files/0.log" Feb 27 09:02:11 crc kubenswrapper[4612]: I0227 09:02:11.863527 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-8drht_8fe63fdd-8b6f-4ecf-81d2-959ae060d567/cp-reloader/0.log" Feb 27 09:02:11 crc kubenswrapper[4612]: I0227 09:02:11.878430 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-8drht_8fe63fdd-8b6f-4ecf-81d2-959ae060d567/cp-metrics/0.log" Feb 27 09:02:11 crc kubenswrapper[4612]: I0227 09:02:11.890273 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-8drht_8fe63fdd-8b6f-4ecf-81d2-959ae060d567/cp-reloader/0.log" Feb 27 09:02:11 crc kubenswrapper[4612]: I0227 09:02:11.937020 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-8drht_8fe63fdd-8b6f-4ecf-81d2-959ae060d567/cp-frr-files/0.log" Feb 27 09:02:12 crc kubenswrapper[4612]: I0227 09:02:12.085932 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-8drht_8fe63fdd-8b6f-4ecf-81d2-959ae060d567/cp-reloader/0.log" Feb 27 09:02:12 crc kubenswrapper[4612]: I0227 09:02:12.135436 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-8drht_8fe63fdd-8b6f-4ecf-81d2-959ae060d567/cp-frr-files/0.log" Feb 27 09:02:12 crc kubenswrapper[4612]: I0227 09:02:12.165263 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-8drht_8fe63fdd-8b6f-4ecf-81d2-959ae060d567/cp-metrics/0.log" Feb 27 09:02:12 crc kubenswrapper[4612]: I0227 09:02:12.196181 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-8drht_8fe63fdd-8b6f-4ecf-81d2-959ae060d567/cp-metrics/0.log" Feb 27 09:02:12 crc kubenswrapper[4612]: I0227 09:02:12.402145 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-8drht_8fe63fdd-8b6f-4ecf-81d2-959ae060d567/cp-reloader/0.log" Feb 27 09:02:12 crc kubenswrapper[4612]: I0227 09:02:12.417966 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-8drht_8fe63fdd-8b6f-4ecf-81d2-959ae060d567/cp-frr-files/0.log" Feb 27 09:02:12 crc kubenswrapper[4612]: I0227 09:02:12.423001 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-8drht_8fe63fdd-8b6f-4ecf-81d2-959ae060d567/cp-metrics/0.log" Feb 27 09:02:12 crc kubenswrapper[4612]: I0227 09:02:12.495772 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-8drht_8fe63fdd-8b6f-4ecf-81d2-959ae060d567/controller/0.log" Feb 27 09:02:12 crc kubenswrapper[4612]: I0227 09:02:12.653170 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-8drht_8fe63fdd-8b6f-4ecf-81d2-959ae060d567/kube-rbac-proxy/0.log" Feb 27 09:02:12 crc kubenswrapper[4612]: I0227 09:02:12.768612 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-8drht_8fe63fdd-8b6f-4ecf-81d2-959ae060d567/frr-metrics/0.log" Feb 27 09:02:12 crc kubenswrapper[4612]: I0227 09:02:12.839169 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-8drht_8fe63fdd-8b6f-4ecf-81d2-959ae060d567/kube-rbac-proxy-frr/0.log" Feb 27 09:02:13 crc kubenswrapper[4612]: I0227 09:02:13.019970 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-8drht_8fe63fdd-8b6f-4ecf-81d2-959ae060d567/reloader/0.log" Feb 27 09:02:13 crc kubenswrapper[4612]: I0227 09:02:13.132077 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-7f989f654f-x2k2n_161b64e1-f414-4519-9fc1-c564d1cc0316/frr-k8s-webhook-server/0.log" Feb 27 09:02:13 crc kubenswrapper[4612]: I0227 09:02:13.393389 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-78d5b849b-c2rb5_29cfdfc3-d4a9-44df-b46e-638a9707a9f1/manager/0.log" Feb 27 09:02:13 crc kubenswrapper[4612]: I0227 09:02:13.625862 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-6f9ff8875f-jphph_62df4a96-aac6-4169-bbc4-d41ea85b8722/webhook-server/0.log" Feb 27 09:02:13 crc kubenswrapper[4612]: I0227 09:02:13.803561 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-zlmtg_9b1f5f78-3c58-42fb-9a8f-0166eab4f500/kube-rbac-proxy/0.log" Feb 27 09:02:14 crc kubenswrapper[4612]: I0227 09:02:14.003100 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-8drht_8fe63fdd-8b6f-4ecf-81d2-959ae060d567/frr/0.log" Feb 27 09:02:14 crc kubenswrapper[4612]: I0227 09:02:14.358122 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-zlmtg_9b1f5f78-3c58-42fb-9a8f-0166eab4f500/speaker/0.log" Feb 27 09:02:19 crc kubenswrapper[4612]: I0227 09:02:19.852813 4612 scope.go:117] "RemoveContainer" containerID="aa498fa4b5f3843852a1ebe0b165b06254525eeb49f5224c7bd2a0f8c2ae2480" Feb 27 09:02:19 crc kubenswrapper[4612]: E0227 09:02:19.853411 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-924vb_openshift-machine-config-operator(24599373-7adc-4a1b-8bb4-797bf726f43d)\"" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" Feb 27 09:02:20 crc kubenswrapper[4612]: I0227 09:02:20.131627 4612 scope.go:117] "RemoveContainer" containerID="f226423c80e8686c070bc3455d4e33aab449e5154d17dfcbc3d5f3d441c5040d" Feb 27 09:02:29 crc kubenswrapper[4612]: I0227 09:02:29.722099 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82g84wq_d7ee024d-7259-4702-bc99-fb938a71345f/util/0.log" Feb 27 09:02:30 crc kubenswrapper[4612]: I0227 09:02:30.439277 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82g84wq_d7ee024d-7259-4702-bc99-fb938a71345f/util/0.log" Feb 27 09:02:30 crc kubenswrapper[4612]: I0227 09:02:30.473359 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82g84wq_d7ee024d-7259-4702-bc99-fb938a71345f/pull/0.log" Feb 27 09:02:30 crc kubenswrapper[4612]: I0227 09:02:30.517208 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82g84wq_d7ee024d-7259-4702-bc99-fb938a71345f/pull/0.log" Feb 27 09:02:30 crc kubenswrapper[4612]: I0227 09:02:30.662963 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82g84wq_d7ee024d-7259-4702-bc99-fb938a71345f/extract/0.log" Feb 27 09:02:30 crc kubenswrapper[4612]: I0227 09:02:30.742033 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82g84wq_d7ee024d-7259-4702-bc99-fb938a71345f/util/0.log" Feb 27 09:02:30 crc kubenswrapper[4612]: I0227 09:02:30.744288 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82g84wq_d7ee024d-7259-4702-bc99-fb938a71345f/pull/0.log" Feb 27 09:02:30 crc kubenswrapper[4612]: I0227 09:02:30.912030 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-ln97g_4a605168-76bd-4605-9a60-5ea574a0b2c1/extract-utilities/0.log" Feb 27 09:02:31 crc kubenswrapper[4612]: I0227 09:02:31.070153 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-ln97g_4a605168-76bd-4605-9a60-5ea574a0b2c1/extract-utilities/0.log" Feb 27 09:02:31 crc kubenswrapper[4612]: I0227 09:02:31.086416 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-ln97g_4a605168-76bd-4605-9a60-5ea574a0b2c1/extract-content/0.log" Feb 27 09:02:31 crc kubenswrapper[4612]: I0227 09:02:31.141770 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-ln97g_4a605168-76bd-4605-9a60-5ea574a0b2c1/extract-content/0.log" Feb 27 09:02:31 crc kubenswrapper[4612]: I0227 09:02:31.256463 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-ln97g_4a605168-76bd-4605-9a60-5ea574a0b2c1/extract-utilities/0.log" Feb 27 09:02:31 crc kubenswrapper[4612]: I0227 09:02:31.281971 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-ln97g_4a605168-76bd-4605-9a60-5ea574a0b2c1/extract-content/0.log" Feb 27 09:02:31 crc kubenswrapper[4612]: I0227 09:02:31.517658 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-ck8cc_f2b71124-3e61-48fc-a493-60c759124936/extract-utilities/0.log" Feb 27 09:02:31 crc kubenswrapper[4612]: I0227 09:02:31.840018 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-ln97g_4a605168-76bd-4605-9a60-5ea574a0b2c1/registry-server/0.log" Feb 27 09:02:31 crc kubenswrapper[4612]: I0227 09:02:31.857301 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-ck8cc_f2b71124-3e61-48fc-a493-60c759124936/extract-utilities/0.log" Feb 27 09:02:31 crc kubenswrapper[4612]: I0227 09:02:31.924032 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-ck8cc_f2b71124-3e61-48fc-a493-60c759124936/extract-content/0.log" Feb 27 09:02:31 crc kubenswrapper[4612]: I0227 09:02:31.934864 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-ck8cc_f2b71124-3e61-48fc-a493-60c759124936/extract-content/0.log" Feb 27 09:02:32 crc kubenswrapper[4612]: I0227 09:02:32.124805 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-ck8cc_f2b71124-3e61-48fc-a493-60c759124936/extract-utilities/0.log" Feb 27 09:02:32 crc kubenswrapper[4612]: I0227 09:02:32.193248 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-ck8cc_f2b71124-3e61-48fc-a493-60c759124936/extract-content/0.log" Feb 27 09:02:32 crc kubenswrapper[4612]: I0227 09:02:32.488976 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4nq7w2_f3841018-8eb3-4d1d-b78a-c023bbb050cd/util/0.log" Feb 27 09:02:32 crc kubenswrapper[4612]: I0227 09:02:32.783274 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-ck8cc_f2b71124-3e61-48fc-a493-60c759124936/registry-server/0.log" Feb 27 09:02:32 crc kubenswrapper[4612]: I0227 09:02:32.803201 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4nq7w2_f3841018-8eb3-4d1d-b78a-c023bbb050cd/util/0.log" Feb 27 09:02:32 crc kubenswrapper[4612]: I0227 09:02:32.832520 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4nq7w2_f3841018-8eb3-4d1d-b78a-c023bbb050cd/pull/0.log" Feb 27 09:02:32 crc kubenswrapper[4612]: I0227 09:02:32.862981 4612 scope.go:117] "RemoveContainer" containerID="aa498fa4b5f3843852a1ebe0b165b06254525eeb49f5224c7bd2a0f8c2ae2480" Feb 27 09:02:32 crc kubenswrapper[4612]: E0227 09:02:32.863983 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-924vb_openshift-machine-config-operator(24599373-7adc-4a1b-8bb4-797bf726f43d)\"" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" Feb 27 09:02:32 crc kubenswrapper[4612]: I0227 09:02:32.895569 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4nq7w2_f3841018-8eb3-4d1d-b78a-c023bbb050cd/pull/0.log" Feb 27 09:02:33 crc kubenswrapper[4612]: I0227 09:02:33.120114 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4nq7w2_f3841018-8eb3-4d1d-b78a-c023bbb050cd/extract/0.log" Feb 27 09:02:33 crc kubenswrapper[4612]: I0227 09:02:33.185721 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4nq7w2_f3841018-8eb3-4d1d-b78a-c023bbb050cd/pull/0.log" Feb 27 09:02:33 crc kubenswrapper[4612]: I0227 09:02:33.208210 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4nq7w2_f3841018-8eb3-4d1d-b78a-c023bbb050cd/util/0.log" Feb 27 09:02:33 crc kubenswrapper[4612]: I0227 09:02:33.360049 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-88rfh_6f0a0683-7e38-410b-8c31-137e9d3a841a/marketplace-operator/0.log" Feb 27 09:02:33 crc kubenswrapper[4612]: I0227 09:02:33.429829 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-8dbvm_eb785264-58b2-4d87-89f3-8dc86f784419/extract-utilities/0.log" Feb 27 09:02:33 crc kubenswrapper[4612]: I0227 09:02:33.637229 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-8dbvm_eb785264-58b2-4d87-89f3-8dc86f784419/extract-utilities/0.log" Feb 27 09:02:33 crc kubenswrapper[4612]: I0227 09:02:33.679804 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-8dbvm_eb785264-58b2-4d87-89f3-8dc86f784419/extract-content/0.log" Feb 27 09:02:33 crc kubenswrapper[4612]: I0227 09:02:33.696721 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-8dbvm_eb785264-58b2-4d87-89f3-8dc86f784419/extract-content/0.log" Feb 27 09:02:33 crc kubenswrapper[4612]: I0227 09:02:33.850520 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-8dbvm_eb785264-58b2-4d87-89f3-8dc86f784419/extract-utilities/0.log" Feb 27 09:02:33 crc kubenswrapper[4612]: I0227 09:02:33.893891 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-8dbvm_eb785264-58b2-4d87-89f3-8dc86f784419/extract-content/0.log" Feb 27 09:02:33 crc kubenswrapper[4612]: I0227 09:02:33.919048 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-x56jg_cba13562-dcb0-4896-94eb-5d5246ffa1c2/extract-utilities/0.log" Feb 27 09:02:34 crc kubenswrapper[4612]: I0227 09:02:34.051506 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-8dbvm_eb785264-58b2-4d87-89f3-8dc86f784419/registry-server/0.log" Feb 27 09:02:34 crc kubenswrapper[4612]: I0227 09:02:34.178053 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-x56jg_cba13562-dcb0-4896-94eb-5d5246ffa1c2/extract-content/0.log" Feb 27 09:02:34 crc kubenswrapper[4612]: I0227 09:02:34.192858 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-x56jg_cba13562-dcb0-4896-94eb-5d5246ffa1c2/extract-content/0.log" Feb 27 09:02:34 crc kubenswrapper[4612]: I0227 09:02:34.218433 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-x56jg_cba13562-dcb0-4896-94eb-5d5246ffa1c2/extract-utilities/0.log" Feb 27 09:02:34 crc kubenswrapper[4612]: I0227 09:02:34.444686 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-x56jg_cba13562-dcb0-4896-94eb-5d5246ffa1c2/extract-utilities/0.log" Feb 27 09:02:34 crc kubenswrapper[4612]: I0227 09:02:34.485058 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-x56jg_cba13562-dcb0-4896-94eb-5d5246ffa1c2/extract-content/0.log" Feb 27 09:02:34 crc kubenswrapper[4612]: I0227 09:02:34.949885 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-x56jg_cba13562-dcb0-4896-94eb-5d5246ffa1c2/registry-server/0.log" Feb 27 09:02:45 crc kubenswrapper[4612]: I0227 09:02:45.852850 4612 scope.go:117] "RemoveContainer" containerID="aa498fa4b5f3843852a1ebe0b165b06254525eeb49f5224c7bd2a0f8c2ae2480" Feb 27 09:02:45 crc kubenswrapper[4612]: E0227 09:02:45.854726 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-924vb_openshift-machine-config-operator(24599373-7adc-4a1b-8bb4-797bf726f43d)\"" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" Feb 27 09:02:53 crc kubenswrapper[4612]: I0227 09:02:53.152788 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-rw4z2"] Feb 27 09:02:53 crc kubenswrapper[4612]: E0227 09:02:53.153608 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="80a082d2-95dd-4d63-965f-8d79c871ad0e" containerName="oc" Feb 27 09:02:53 crc kubenswrapper[4612]: I0227 09:02:53.153635 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="80a082d2-95dd-4d63-965f-8d79c871ad0e" containerName="oc" Feb 27 09:02:53 crc kubenswrapper[4612]: I0227 09:02:53.153840 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="80a082d2-95dd-4d63-965f-8d79c871ad0e" containerName="oc" Feb 27 09:02:53 crc kubenswrapper[4612]: I0227 09:02:53.155222 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rw4z2" Feb 27 09:02:53 crc kubenswrapper[4612]: I0227 09:02:53.175156 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-rw4z2"] Feb 27 09:02:53 crc kubenswrapper[4612]: I0227 09:02:53.224958 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0257c0d6-34d2-476b-a4d9-f0f7d14aa3da-utilities\") pod \"community-operators-rw4z2\" (UID: \"0257c0d6-34d2-476b-a4d9-f0f7d14aa3da\") " pod="openshift-marketplace/community-operators-rw4z2" Feb 27 09:02:53 crc kubenswrapper[4612]: I0227 09:02:53.225197 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0257c0d6-34d2-476b-a4d9-f0f7d14aa3da-catalog-content\") pod \"community-operators-rw4z2\" (UID: \"0257c0d6-34d2-476b-a4d9-f0f7d14aa3da\") " pod="openshift-marketplace/community-operators-rw4z2" Feb 27 09:02:53 crc kubenswrapper[4612]: I0227 09:02:53.225502 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qjv4h\" (UniqueName: \"kubernetes.io/projected/0257c0d6-34d2-476b-a4d9-f0f7d14aa3da-kube-api-access-qjv4h\") pod \"community-operators-rw4z2\" (UID: \"0257c0d6-34d2-476b-a4d9-f0f7d14aa3da\") " pod="openshift-marketplace/community-operators-rw4z2" Feb 27 09:02:53 crc kubenswrapper[4612]: I0227 09:02:53.326780 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qjv4h\" (UniqueName: \"kubernetes.io/projected/0257c0d6-34d2-476b-a4d9-f0f7d14aa3da-kube-api-access-qjv4h\") pod \"community-operators-rw4z2\" (UID: \"0257c0d6-34d2-476b-a4d9-f0f7d14aa3da\") " pod="openshift-marketplace/community-operators-rw4z2" Feb 27 09:02:53 crc kubenswrapper[4612]: I0227 09:02:53.326848 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0257c0d6-34d2-476b-a4d9-f0f7d14aa3da-utilities\") pod \"community-operators-rw4z2\" (UID: \"0257c0d6-34d2-476b-a4d9-f0f7d14aa3da\") " pod="openshift-marketplace/community-operators-rw4z2" Feb 27 09:02:53 crc kubenswrapper[4612]: I0227 09:02:53.326916 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0257c0d6-34d2-476b-a4d9-f0f7d14aa3da-catalog-content\") pod \"community-operators-rw4z2\" (UID: \"0257c0d6-34d2-476b-a4d9-f0f7d14aa3da\") " pod="openshift-marketplace/community-operators-rw4z2" Feb 27 09:02:53 crc kubenswrapper[4612]: I0227 09:02:53.327435 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0257c0d6-34d2-476b-a4d9-f0f7d14aa3da-catalog-content\") pod \"community-operators-rw4z2\" (UID: \"0257c0d6-34d2-476b-a4d9-f0f7d14aa3da\") " pod="openshift-marketplace/community-operators-rw4z2" Feb 27 09:02:53 crc kubenswrapper[4612]: I0227 09:02:53.328006 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0257c0d6-34d2-476b-a4d9-f0f7d14aa3da-utilities\") pod \"community-operators-rw4z2\" (UID: \"0257c0d6-34d2-476b-a4d9-f0f7d14aa3da\") " pod="openshift-marketplace/community-operators-rw4z2" Feb 27 09:02:53 crc kubenswrapper[4612]: I0227 09:02:53.345874 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qjv4h\" (UniqueName: \"kubernetes.io/projected/0257c0d6-34d2-476b-a4d9-f0f7d14aa3da-kube-api-access-qjv4h\") pod \"community-operators-rw4z2\" (UID: \"0257c0d6-34d2-476b-a4d9-f0f7d14aa3da\") " pod="openshift-marketplace/community-operators-rw4z2" Feb 27 09:02:53 crc kubenswrapper[4612]: I0227 09:02:53.475446 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rw4z2" Feb 27 09:02:54 crc kubenswrapper[4612]: I0227 09:02:54.054011 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-rw4z2"] Feb 27 09:02:54 crc kubenswrapper[4612]: I0227 09:02:54.527156 4612 generic.go:334] "Generic (PLEG): container finished" podID="0257c0d6-34d2-476b-a4d9-f0f7d14aa3da" containerID="eec9cffd852d9ee8075fad702bbfaa5751a501b4f7ec1f245e759d7464a6bb2b" exitCode=0 Feb 27 09:02:54 crc kubenswrapper[4612]: I0227 09:02:54.527324 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rw4z2" event={"ID":"0257c0d6-34d2-476b-a4d9-f0f7d14aa3da","Type":"ContainerDied","Data":"eec9cffd852d9ee8075fad702bbfaa5751a501b4f7ec1f245e759d7464a6bb2b"} Feb 27 09:02:54 crc kubenswrapper[4612]: I0227 09:02:54.527390 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rw4z2" event={"ID":"0257c0d6-34d2-476b-a4d9-f0f7d14aa3da","Type":"ContainerStarted","Data":"ef7c66d07cb174d9c5078906e41ff34506f8824c4e8379fe93b46db006a72abe"} Feb 27 09:02:55 crc kubenswrapper[4612]: I0227 09:02:55.539707 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rw4z2" event={"ID":"0257c0d6-34d2-476b-a4d9-f0f7d14aa3da","Type":"ContainerStarted","Data":"dc723a009f2e8710de0c6d90384039b3d1ff6bdcca8fa56b8506bc3b7e7132cb"} Feb 27 09:02:57 crc kubenswrapper[4612]: I0227 09:02:57.852843 4612 scope.go:117] "RemoveContainer" containerID="aa498fa4b5f3843852a1ebe0b165b06254525eeb49f5224c7bd2a0f8c2ae2480" Feb 27 09:02:57 crc kubenswrapper[4612]: E0227 09:02:57.853743 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-924vb_openshift-machine-config-operator(24599373-7adc-4a1b-8bb4-797bf726f43d)\"" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" Feb 27 09:02:58 crc kubenswrapper[4612]: I0227 09:02:58.563604 4612 generic.go:334] "Generic (PLEG): container finished" podID="0257c0d6-34d2-476b-a4d9-f0f7d14aa3da" containerID="dc723a009f2e8710de0c6d90384039b3d1ff6bdcca8fa56b8506bc3b7e7132cb" exitCode=0 Feb 27 09:02:58 crc kubenswrapper[4612]: I0227 09:02:58.563643 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rw4z2" event={"ID":"0257c0d6-34d2-476b-a4d9-f0f7d14aa3da","Type":"ContainerDied","Data":"dc723a009f2e8710de0c6d90384039b3d1ff6bdcca8fa56b8506bc3b7e7132cb"} Feb 27 09:02:59 crc kubenswrapper[4612]: I0227 09:02:59.574487 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rw4z2" event={"ID":"0257c0d6-34d2-476b-a4d9-f0f7d14aa3da","Type":"ContainerStarted","Data":"b3c5c426046a1613f65061f1317b3f9d63223e767990c7dbfa09841a98841dda"} Feb 27 09:02:59 crc kubenswrapper[4612]: I0227 09:02:59.616850 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-rw4z2" podStartSLOduration=2.103362145 podStartE2EDuration="6.616830427s" podCreationTimestamp="2026-02-27 09:02:53 +0000 UTC" firstStartedPulling="2026-02-27 09:02:54.529191081 +0000 UTC m=+4432.383121079" lastFinishedPulling="2026-02-27 09:02:59.042659363 +0000 UTC m=+4436.896589361" observedRunningTime="2026-02-27 09:02:59.598240008 +0000 UTC m=+4437.452170007" watchObservedRunningTime="2026-02-27 09:02:59.616830427 +0000 UTC m=+4437.470760425" Feb 27 09:03:03 crc kubenswrapper[4612]: I0227 09:03:03.476855 4612 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-rw4z2" Feb 27 09:03:03 crc kubenswrapper[4612]: I0227 09:03:03.477459 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-rw4z2" Feb 27 09:03:04 crc kubenswrapper[4612]: I0227 09:03:04.539201 4612 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-rw4z2" podUID="0257c0d6-34d2-476b-a4d9-f0f7d14aa3da" containerName="registry-server" probeResult="failure" output=< Feb 27 09:03:04 crc kubenswrapper[4612]: timeout: failed to connect service ":50051" within 1s Feb 27 09:03:04 crc kubenswrapper[4612]: > Feb 27 09:03:08 crc kubenswrapper[4612]: I0227 09:03:08.853392 4612 scope.go:117] "RemoveContainer" containerID="aa498fa4b5f3843852a1ebe0b165b06254525eeb49f5224c7bd2a0f8c2ae2480" Feb 27 09:03:08 crc kubenswrapper[4612]: E0227 09:03:08.854032 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-924vb_openshift-machine-config-operator(24599373-7adc-4a1b-8bb4-797bf726f43d)\"" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" Feb 27 09:03:14 crc kubenswrapper[4612]: I0227 09:03:14.532793 4612 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-rw4z2" podUID="0257c0d6-34d2-476b-a4d9-f0f7d14aa3da" containerName="registry-server" probeResult="failure" output=< Feb 27 09:03:14 crc kubenswrapper[4612]: timeout: failed to connect service ":50051" within 1s Feb 27 09:03:14 crc kubenswrapper[4612]: > Feb 27 09:03:19 crc kubenswrapper[4612]: I0227 09:03:19.853710 4612 scope.go:117] "RemoveContainer" containerID="aa498fa4b5f3843852a1ebe0b165b06254525eeb49f5224c7bd2a0f8c2ae2480" Feb 27 09:03:19 crc kubenswrapper[4612]: E0227 09:03:19.854672 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-924vb_openshift-machine-config-operator(24599373-7adc-4a1b-8bb4-797bf726f43d)\"" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" Feb 27 09:03:23 crc kubenswrapper[4612]: I0227 09:03:23.559539 4612 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-rw4z2" Feb 27 09:03:23 crc kubenswrapper[4612]: I0227 09:03:23.642924 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-rw4z2" Feb 27 09:03:24 crc kubenswrapper[4612]: I0227 09:03:24.362165 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-rw4z2"] Feb 27 09:03:24 crc kubenswrapper[4612]: I0227 09:03:24.783930 4612 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-rw4z2" podUID="0257c0d6-34d2-476b-a4d9-f0f7d14aa3da" containerName="registry-server" containerID="cri-o://b3c5c426046a1613f65061f1317b3f9d63223e767990c7dbfa09841a98841dda" gracePeriod=2 Feb 27 09:03:25 crc kubenswrapper[4612]: I0227 09:03:25.355741 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rw4z2" Feb 27 09:03:25 crc kubenswrapper[4612]: I0227 09:03:25.518375 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qjv4h\" (UniqueName: \"kubernetes.io/projected/0257c0d6-34d2-476b-a4d9-f0f7d14aa3da-kube-api-access-qjv4h\") pod \"0257c0d6-34d2-476b-a4d9-f0f7d14aa3da\" (UID: \"0257c0d6-34d2-476b-a4d9-f0f7d14aa3da\") " Feb 27 09:03:25 crc kubenswrapper[4612]: I0227 09:03:25.518563 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0257c0d6-34d2-476b-a4d9-f0f7d14aa3da-utilities\") pod \"0257c0d6-34d2-476b-a4d9-f0f7d14aa3da\" (UID: \"0257c0d6-34d2-476b-a4d9-f0f7d14aa3da\") " Feb 27 09:03:25 crc kubenswrapper[4612]: I0227 09:03:25.518823 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0257c0d6-34d2-476b-a4d9-f0f7d14aa3da-catalog-content\") pod \"0257c0d6-34d2-476b-a4d9-f0f7d14aa3da\" (UID: \"0257c0d6-34d2-476b-a4d9-f0f7d14aa3da\") " Feb 27 09:03:25 crc kubenswrapper[4612]: I0227 09:03:25.519481 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0257c0d6-34d2-476b-a4d9-f0f7d14aa3da-utilities" (OuterVolumeSpecName: "utilities") pod "0257c0d6-34d2-476b-a4d9-f0f7d14aa3da" (UID: "0257c0d6-34d2-476b-a4d9-f0f7d14aa3da"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 09:03:25 crc kubenswrapper[4612]: I0227 09:03:25.534110 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0257c0d6-34d2-476b-a4d9-f0f7d14aa3da-kube-api-access-qjv4h" (OuterVolumeSpecName: "kube-api-access-qjv4h") pod "0257c0d6-34d2-476b-a4d9-f0f7d14aa3da" (UID: "0257c0d6-34d2-476b-a4d9-f0f7d14aa3da"). InnerVolumeSpecName "kube-api-access-qjv4h". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 09:03:25 crc kubenswrapper[4612]: I0227 09:03:25.579185 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0257c0d6-34d2-476b-a4d9-f0f7d14aa3da-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0257c0d6-34d2-476b-a4d9-f0f7d14aa3da" (UID: "0257c0d6-34d2-476b-a4d9-f0f7d14aa3da"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 09:03:25 crc kubenswrapper[4612]: I0227 09:03:25.621717 4612 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0257c0d6-34d2-476b-a4d9-f0f7d14aa3da-utilities\") on node \"crc\" DevicePath \"\"" Feb 27 09:03:25 crc kubenswrapper[4612]: I0227 09:03:25.621748 4612 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0257c0d6-34d2-476b-a4d9-f0f7d14aa3da-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 27 09:03:25 crc kubenswrapper[4612]: I0227 09:03:25.621759 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qjv4h\" (UniqueName: \"kubernetes.io/projected/0257c0d6-34d2-476b-a4d9-f0f7d14aa3da-kube-api-access-qjv4h\") on node \"crc\" DevicePath \"\"" Feb 27 09:03:25 crc kubenswrapper[4612]: I0227 09:03:25.801154 4612 generic.go:334] "Generic (PLEG): container finished" podID="0257c0d6-34d2-476b-a4d9-f0f7d14aa3da" containerID="b3c5c426046a1613f65061f1317b3f9d63223e767990c7dbfa09841a98841dda" exitCode=0 Feb 27 09:03:25 crc kubenswrapper[4612]: I0227 09:03:25.801201 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rw4z2" event={"ID":"0257c0d6-34d2-476b-a4d9-f0f7d14aa3da","Type":"ContainerDied","Data":"b3c5c426046a1613f65061f1317b3f9d63223e767990c7dbfa09841a98841dda"} Feb 27 09:03:25 crc kubenswrapper[4612]: I0227 09:03:25.801229 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rw4z2" event={"ID":"0257c0d6-34d2-476b-a4d9-f0f7d14aa3da","Type":"ContainerDied","Data":"ef7c66d07cb174d9c5078906e41ff34506f8824c4e8379fe93b46db006a72abe"} Feb 27 09:03:25 crc kubenswrapper[4612]: I0227 09:03:25.801245 4612 scope.go:117] "RemoveContainer" containerID="b3c5c426046a1613f65061f1317b3f9d63223e767990c7dbfa09841a98841dda" Feb 27 09:03:25 crc kubenswrapper[4612]: I0227 09:03:25.801202 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rw4z2" Feb 27 09:03:25 crc kubenswrapper[4612]: I0227 09:03:25.845960 4612 scope.go:117] "RemoveContainer" containerID="dc723a009f2e8710de0c6d90384039b3d1ff6bdcca8fa56b8506bc3b7e7132cb" Feb 27 09:03:25 crc kubenswrapper[4612]: I0227 09:03:25.854776 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-rw4z2"] Feb 27 09:03:25 crc kubenswrapper[4612]: I0227 09:03:25.868896 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-rw4z2"] Feb 27 09:03:25 crc kubenswrapper[4612]: I0227 09:03:25.875097 4612 scope.go:117] "RemoveContainer" containerID="eec9cffd852d9ee8075fad702bbfaa5751a501b4f7ec1f245e759d7464a6bb2b" Feb 27 09:03:25 crc kubenswrapper[4612]: I0227 09:03:25.924784 4612 scope.go:117] "RemoveContainer" containerID="b3c5c426046a1613f65061f1317b3f9d63223e767990c7dbfa09841a98841dda" Feb 27 09:03:25 crc kubenswrapper[4612]: E0227 09:03:25.925480 4612 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b3c5c426046a1613f65061f1317b3f9d63223e767990c7dbfa09841a98841dda\": container with ID starting with b3c5c426046a1613f65061f1317b3f9d63223e767990c7dbfa09841a98841dda not found: ID does not exist" containerID="b3c5c426046a1613f65061f1317b3f9d63223e767990c7dbfa09841a98841dda" Feb 27 09:03:25 crc kubenswrapper[4612]: I0227 09:03:25.925516 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b3c5c426046a1613f65061f1317b3f9d63223e767990c7dbfa09841a98841dda"} err="failed to get container status \"b3c5c426046a1613f65061f1317b3f9d63223e767990c7dbfa09841a98841dda\": rpc error: code = NotFound desc = could not find container \"b3c5c426046a1613f65061f1317b3f9d63223e767990c7dbfa09841a98841dda\": container with ID starting with b3c5c426046a1613f65061f1317b3f9d63223e767990c7dbfa09841a98841dda not found: ID does not exist" Feb 27 09:03:25 crc kubenswrapper[4612]: I0227 09:03:25.925540 4612 scope.go:117] "RemoveContainer" containerID="dc723a009f2e8710de0c6d90384039b3d1ff6bdcca8fa56b8506bc3b7e7132cb" Feb 27 09:03:25 crc kubenswrapper[4612]: E0227 09:03:25.926203 4612 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dc723a009f2e8710de0c6d90384039b3d1ff6bdcca8fa56b8506bc3b7e7132cb\": container with ID starting with dc723a009f2e8710de0c6d90384039b3d1ff6bdcca8fa56b8506bc3b7e7132cb not found: ID does not exist" containerID="dc723a009f2e8710de0c6d90384039b3d1ff6bdcca8fa56b8506bc3b7e7132cb" Feb 27 09:03:25 crc kubenswrapper[4612]: I0227 09:03:25.926292 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dc723a009f2e8710de0c6d90384039b3d1ff6bdcca8fa56b8506bc3b7e7132cb"} err="failed to get container status \"dc723a009f2e8710de0c6d90384039b3d1ff6bdcca8fa56b8506bc3b7e7132cb\": rpc error: code = NotFound desc = could not find container \"dc723a009f2e8710de0c6d90384039b3d1ff6bdcca8fa56b8506bc3b7e7132cb\": container with ID starting with dc723a009f2e8710de0c6d90384039b3d1ff6bdcca8fa56b8506bc3b7e7132cb not found: ID does not exist" Feb 27 09:03:25 crc kubenswrapper[4612]: I0227 09:03:25.926340 4612 scope.go:117] "RemoveContainer" containerID="eec9cffd852d9ee8075fad702bbfaa5751a501b4f7ec1f245e759d7464a6bb2b" Feb 27 09:03:25 crc kubenswrapper[4612]: E0227 09:03:25.926899 4612 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eec9cffd852d9ee8075fad702bbfaa5751a501b4f7ec1f245e759d7464a6bb2b\": container with ID starting with eec9cffd852d9ee8075fad702bbfaa5751a501b4f7ec1f245e759d7464a6bb2b not found: ID does not exist" containerID="eec9cffd852d9ee8075fad702bbfaa5751a501b4f7ec1f245e759d7464a6bb2b" Feb 27 09:03:25 crc kubenswrapper[4612]: I0227 09:03:25.926927 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eec9cffd852d9ee8075fad702bbfaa5751a501b4f7ec1f245e759d7464a6bb2b"} err="failed to get container status \"eec9cffd852d9ee8075fad702bbfaa5751a501b4f7ec1f245e759d7464a6bb2b\": rpc error: code = NotFound desc = could not find container \"eec9cffd852d9ee8075fad702bbfaa5751a501b4f7ec1f245e759d7464a6bb2b\": container with ID starting with eec9cffd852d9ee8075fad702bbfaa5751a501b4f7ec1f245e759d7464a6bb2b not found: ID does not exist" Feb 27 09:03:26 crc kubenswrapper[4612]: I0227 09:03:26.895026 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0257c0d6-34d2-476b-a4d9-f0f7d14aa3da" path="/var/lib/kubelet/pods/0257c0d6-34d2-476b-a4d9-f0f7d14aa3da/volumes" Feb 27 09:03:31 crc kubenswrapper[4612]: I0227 09:03:31.853253 4612 scope.go:117] "RemoveContainer" containerID="aa498fa4b5f3843852a1ebe0b165b06254525eeb49f5224c7bd2a0f8c2ae2480" Feb 27 09:03:31 crc kubenswrapper[4612]: E0227 09:03:31.855594 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-924vb_openshift-machine-config-operator(24599373-7adc-4a1b-8bb4-797bf726f43d)\"" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" Feb 27 09:03:42 crc kubenswrapper[4612]: I0227 09:03:42.887014 4612 scope.go:117] "RemoveContainer" containerID="aa498fa4b5f3843852a1ebe0b165b06254525eeb49f5224c7bd2a0f8c2ae2480" Feb 27 09:03:42 crc kubenswrapper[4612]: E0227 09:03:42.888230 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-924vb_openshift-machine-config-operator(24599373-7adc-4a1b-8bb4-797bf726f43d)\"" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" Feb 27 09:03:53 crc kubenswrapper[4612]: I0227 09:03:53.853065 4612 scope.go:117] "RemoveContainer" containerID="aa498fa4b5f3843852a1ebe0b165b06254525eeb49f5224c7bd2a0f8c2ae2480" Feb 27 09:03:53 crc kubenswrapper[4612]: E0227 09:03:53.853735 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-924vb_openshift-machine-config-operator(24599373-7adc-4a1b-8bb4-797bf726f43d)\"" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" Feb 27 09:04:00 crc kubenswrapper[4612]: I0227 09:04:00.152415 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29536384-d7rj9"] Feb 27 09:04:00 crc kubenswrapper[4612]: E0227 09:04:00.153245 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0257c0d6-34d2-476b-a4d9-f0f7d14aa3da" containerName="extract-content" Feb 27 09:04:00 crc kubenswrapper[4612]: I0227 09:04:00.153256 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="0257c0d6-34d2-476b-a4d9-f0f7d14aa3da" containerName="extract-content" Feb 27 09:04:00 crc kubenswrapper[4612]: E0227 09:04:00.153279 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0257c0d6-34d2-476b-a4d9-f0f7d14aa3da" containerName="registry-server" Feb 27 09:04:00 crc kubenswrapper[4612]: I0227 09:04:00.153285 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="0257c0d6-34d2-476b-a4d9-f0f7d14aa3da" containerName="registry-server" Feb 27 09:04:00 crc kubenswrapper[4612]: E0227 09:04:00.153298 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0257c0d6-34d2-476b-a4d9-f0f7d14aa3da" containerName="extract-utilities" Feb 27 09:04:00 crc kubenswrapper[4612]: I0227 09:04:00.153304 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="0257c0d6-34d2-476b-a4d9-f0f7d14aa3da" containerName="extract-utilities" Feb 27 09:04:00 crc kubenswrapper[4612]: I0227 09:04:00.153490 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="0257c0d6-34d2-476b-a4d9-f0f7d14aa3da" containerName="registry-server" Feb 27 09:04:00 crc kubenswrapper[4612]: I0227 09:04:00.154065 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536384-d7rj9" Feb 27 09:04:00 crc kubenswrapper[4612]: I0227 09:04:00.156410 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 27 09:04:00 crc kubenswrapper[4612]: I0227 09:04:00.156867 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-5zwfx" Feb 27 09:04:00 crc kubenswrapper[4612]: I0227 09:04:00.176354 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536384-d7rj9"] Feb 27 09:04:00 crc kubenswrapper[4612]: I0227 09:04:00.183507 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 27 09:04:00 crc kubenswrapper[4612]: I0227 09:04:00.256788 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zzzm8\" (UniqueName: \"kubernetes.io/projected/fffd01d6-8461-4815-bd7f-52b0fb816346-kube-api-access-zzzm8\") pod \"auto-csr-approver-29536384-d7rj9\" (UID: \"fffd01d6-8461-4815-bd7f-52b0fb816346\") " pod="openshift-infra/auto-csr-approver-29536384-d7rj9" Feb 27 09:04:00 crc kubenswrapper[4612]: I0227 09:04:00.358958 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zzzm8\" (UniqueName: \"kubernetes.io/projected/fffd01d6-8461-4815-bd7f-52b0fb816346-kube-api-access-zzzm8\") pod \"auto-csr-approver-29536384-d7rj9\" (UID: \"fffd01d6-8461-4815-bd7f-52b0fb816346\") " pod="openshift-infra/auto-csr-approver-29536384-d7rj9" Feb 27 09:04:00 crc kubenswrapper[4612]: I0227 09:04:00.381570 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zzzm8\" (UniqueName: \"kubernetes.io/projected/fffd01d6-8461-4815-bd7f-52b0fb816346-kube-api-access-zzzm8\") pod \"auto-csr-approver-29536384-d7rj9\" (UID: \"fffd01d6-8461-4815-bd7f-52b0fb816346\") " pod="openshift-infra/auto-csr-approver-29536384-d7rj9" Feb 27 09:04:00 crc kubenswrapper[4612]: I0227 09:04:00.484059 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536384-d7rj9" Feb 27 09:04:01 crc kubenswrapper[4612]: I0227 09:04:01.020357 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536384-d7rj9"] Feb 27 09:04:01 crc kubenswrapper[4612]: I0227 09:04:01.199707 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536384-d7rj9" event={"ID":"fffd01d6-8461-4815-bd7f-52b0fb816346","Type":"ContainerStarted","Data":"233e4bb6e66f7c8058a97be9a59651c3f3ad8c946a8d89edf3141852d6fa0c45"} Feb 27 09:04:03 crc kubenswrapper[4612]: I0227 09:04:03.244778 4612 generic.go:334] "Generic (PLEG): container finished" podID="fffd01d6-8461-4815-bd7f-52b0fb816346" containerID="e5379e138e2871c6b7feb18a479f446e54a957a15b725efb3f8cb9700829b830" exitCode=0 Feb 27 09:04:03 crc kubenswrapper[4612]: I0227 09:04:03.245093 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536384-d7rj9" event={"ID":"fffd01d6-8461-4815-bd7f-52b0fb816346","Type":"ContainerDied","Data":"e5379e138e2871c6b7feb18a479f446e54a957a15b725efb3f8cb9700829b830"} Feb 27 09:04:04 crc kubenswrapper[4612]: I0227 09:04:04.658407 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536384-d7rj9" Feb 27 09:04:04 crc kubenswrapper[4612]: I0227 09:04:04.853427 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zzzm8\" (UniqueName: \"kubernetes.io/projected/fffd01d6-8461-4815-bd7f-52b0fb816346-kube-api-access-zzzm8\") pod \"fffd01d6-8461-4815-bd7f-52b0fb816346\" (UID: \"fffd01d6-8461-4815-bd7f-52b0fb816346\") " Feb 27 09:04:04 crc kubenswrapper[4612]: I0227 09:04:04.861906 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fffd01d6-8461-4815-bd7f-52b0fb816346-kube-api-access-zzzm8" (OuterVolumeSpecName: "kube-api-access-zzzm8") pod "fffd01d6-8461-4815-bd7f-52b0fb816346" (UID: "fffd01d6-8461-4815-bd7f-52b0fb816346"). InnerVolumeSpecName "kube-api-access-zzzm8". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 09:04:04 crc kubenswrapper[4612]: I0227 09:04:04.956817 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zzzm8\" (UniqueName: \"kubernetes.io/projected/fffd01d6-8461-4815-bd7f-52b0fb816346-kube-api-access-zzzm8\") on node \"crc\" DevicePath \"\"" Feb 27 09:04:05 crc kubenswrapper[4612]: I0227 09:04:05.270303 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536384-d7rj9" event={"ID":"fffd01d6-8461-4815-bd7f-52b0fb816346","Type":"ContainerDied","Data":"233e4bb6e66f7c8058a97be9a59651c3f3ad8c946a8d89edf3141852d6fa0c45"} Feb 27 09:04:05 crc kubenswrapper[4612]: I0227 09:04:05.270360 4612 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="233e4bb6e66f7c8058a97be9a59651c3f3ad8c946a8d89edf3141852d6fa0c45" Feb 27 09:04:05 crc kubenswrapper[4612]: I0227 09:04:05.270378 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536384-d7rj9" Feb 27 09:04:05 crc kubenswrapper[4612]: I0227 09:04:05.737168 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29536378-k6phd"] Feb 27 09:04:05 crc kubenswrapper[4612]: I0227 09:04:05.748142 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29536378-k6phd"] Feb 27 09:04:06 crc kubenswrapper[4612]: I0227 09:04:06.853681 4612 scope.go:117] "RemoveContainer" containerID="aa498fa4b5f3843852a1ebe0b165b06254525eeb49f5224c7bd2a0f8c2ae2480" Feb 27 09:04:06 crc kubenswrapper[4612]: E0227 09:04:06.853921 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-924vb_openshift-machine-config-operator(24599373-7adc-4a1b-8bb4-797bf726f43d)\"" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" Feb 27 09:04:06 crc kubenswrapper[4612]: I0227 09:04:06.869189 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c50f2ffa-23aa-482f-87ee-d0d700753b07" path="/var/lib/kubelet/pods/c50f2ffa-23aa-482f-87ee-d0d700753b07/volumes" Feb 27 09:04:17 crc kubenswrapper[4612]: I0227 09:04:17.853102 4612 scope.go:117] "RemoveContainer" containerID="aa498fa4b5f3843852a1ebe0b165b06254525eeb49f5224c7bd2a0f8c2ae2480" Feb 27 09:04:17 crc kubenswrapper[4612]: E0227 09:04:17.853917 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-924vb_openshift-machine-config-operator(24599373-7adc-4a1b-8bb4-797bf726f43d)\"" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" Feb 27 09:04:20 crc kubenswrapper[4612]: I0227 09:04:20.344504 4612 scope.go:117] "RemoveContainer" containerID="9c008fe33903e854c28f4859f9be27c7651a865e82c16ed2fe42d44c29243c67" Feb 27 09:04:20 crc kubenswrapper[4612]: I0227 09:04:20.504113 4612 scope.go:117] "RemoveContainer" containerID="65beaa35fb57fae1bd4f6c00722d4c06b4b0d5c2ede32f9b100d41e9a0fb3621" Feb 27 09:04:29 crc kubenswrapper[4612]: I0227 09:04:29.853491 4612 scope.go:117] "RemoveContainer" containerID="aa498fa4b5f3843852a1ebe0b165b06254525eeb49f5224c7bd2a0f8c2ae2480" Feb 27 09:04:29 crc kubenswrapper[4612]: E0227 09:04:29.854146 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-924vb_openshift-machine-config-operator(24599373-7adc-4a1b-8bb4-797bf726f43d)\"" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" Feb 27 09:04:41 crc kubenswrapper[4612]: I0227 09:04:41.853590 4612 scope.go:117] "RemoveContainer" containerID="aa498fa4b5f3843852a1ebe0b165b06254525eeb49f5224c7bd2a0f8c2ae2480" Feb 27 09:04:41 crc kubenswrapper[4612]: E0227 09:04:41.854345 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-924vb_openshift-machine-config-operator(24599373-7adc-4a1b-8bb4-797bf726f43d)\"" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" Feb 27 09:04:52 crc kubenswrapper[4612]: I0227 09:04:52.862426 4612 scope.go:117] "RemoveContainer" containerID="aa498fa4b5f3843852a1ebe0b165b06254525eeb49f5224c7bd2a0f8c2ae2480" Feb 27 09:04:52 crc kubenswrapper[4612]: E0227 09:04:52.863270 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-924vb_openshift-machine-config-operator(24599373-7adc-4a1b-8bb4-797bf726f43d)\"" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" Feb 27 09:05:05 crc kubenswrapper[4612]: I0227 09:05:05.909525 4612 generic.go:334] "Generic (PLEG): container finished" podID="3f06074f-6e22-4526-a04b-e842577c6f2e" containerID="e6328b790dfbc8f52b3131fc81f90e6ddef8bfc38dc9e78acb8c6353c4afc402" exitCode=0 Feb 27 09:05:05 crc kubenswrapper[4612]: I0227 09:05:05.909596 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-mchzb/must-gather-t4rbr" event={"ID":"3f06074f-6e22-4526-a04b-e842577c6f2e","Type":"ContainerDied","Data":"e6328b790dfbc8f52b3131fc81f90e6ddef8bfc38dc9e78acb8c6353c4afc402"} Feb 27 09:05:05 crc kubenswrapper[4612]: I0227 09:05:05.911112 4612 scope.go:117] "RemoveContainer" containerID="e6328b790dfbc8f52b3131fc81f90e6ddef8bfc38dc9e78acb8c6353c4afc402" Feb 27 09:05:06 crc kubenswrapper[4612]: I0227 09:05:06.237954 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-mchzb_must-gather-t4rbr_3f06074f-6e22-4526-a04b-e842577c6f2e/gather/0.log" Feb 27 09:05:07 crc kubenswrapper[4612]: I0227 09:05:07.853155 4612 scope.go:117] "RemoveContainer" containerID="aa498fa4b5f3843852a1ebe0b165b06254525eeb49f5224c7bd2a0f8c2ae2480" Feb 27 09:05:07 crc kubenswrapper[4612]: E0227 09:05:07.853725 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-924vb_openshift-machine-config-operator(24599373-7adc-4a1b-8bb4-797bf726f43d)\"" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" Feb 27 09:05:15 crc kubenswrapper[4612]: I0227 09:05:15.353124 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-mchzb/must-gather-t4rbr"] Feb 27 09:05:15 crc kubenswrapper[4612]: I0227 09:05:15.353847 4612 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-mchzb/must-gather-t4rbr" podUID="3f06074f-6e22-4526-a04b-e842577c6f2e" containerName="copy" containerID="cri-o://b9508d72b81df0d55578ad38e42b9f06bdb5f38cb474a26239d791fdd4d9848e" gracePeriod=2 Feb 27 09:05:15 crc kubenswrapper[4612]: I0227 09:05:15.361767 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-mchzb/must-gather-t4rbr"] Feb 27 09:05:15 crc kubenswrapper[4612]: I0227 09:05:15.786927 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-mchzb_must-gather-t4rbr_3f06074f-6e22-4526-a04b-e842577c6f2e/copy/0.log" Feb 27 09:05:15 crc kubenswrapper[4612]: I0227 09:05:15.787883 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mchzb/must-gather-t4rbr" Feb 27 09:05:15 crc kubenswrapper[4612]: I0227 09:05:15.852789 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/3f06074f-6e22-4526-a04b-e842577c6f2e-must-gather-output\") pod \"3f06074f-6e22-4526-a04b-e842577c6f2e\" (UID: \"3f06074f-6e22-4526-a04b-e842577c6f2e\") " Feb 27 09:05:15 crc kubenswrapper[4612]: I0227 09:05:15.852928 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nvrfq\" (UniqueName: \"kubernetes.io/projected/3f06074f-6e22-4526-a04b-e842577c6f2e-kube-api-access-nvrfq\") pod \"3f06074f-6e22-4526-a04b-e842577c6f2e\" (UID: \"3f06074f-6e22-4526-a04b-e842577c6f2e\") " Feb 27 09:05:15 crc kubenswrapper[4612]: I0227 09:05:15.874045 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3f06074f-6e22-4526-a04b-e842577c6f2e-kube-api-access-nvrfq" (OuterVolumeSpecName: "kube-api-access-nvrfq") pod "3f06074f-6e22-4526-a04b-e842577c6f2e" (UID: "3f06074f-6e22-4526-a04b-e842577c6f2e"). InnerVolumeSpecName "kube-api-access-nvrfq". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 09:05:15 crc kubenswrapper[4612]: I0227 09:05:15.954979 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nvrfq\" (UniqueName: \"kubernetes.io/projected/3f06074f-6e22-4526-a04b-e842577c6f2e-kube-api-access-nvrfq\") on node \"crc\" DevicePath \"\"" Feb 27 09:05:16 crc kubenswrapper[4612]: I0227 09:05:16.038830 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-mchzb_must-gather-t4rbr_3f06074f-6e22-4526-a04b-e842577c6f2e/copy/0.log" Feb 27 09:05:16 crc kubenswrapper[4612]: I0227 09:05:16.040503 4612 generic.go:334] "Generic (PLEG): container finished" podID="3f06074f-6e22-4526-a04b-e842577c6f2e" containerID="b9508d72b81df0d55578ad38e42b9f06bdb5f38cb474a26239d791fdd4d9848e" exitCode=143 Feb 27 09:05:16 crc kubenswrapper[4612]: I0227 09:05:16.040716 4612 scope.go:117] "RemoveContainer" containerID="b9508d72b81df0d55578ad38e42b9f06bdb5f38cb474a26239d791fdd4d9848e" Feb 27 09:05:16 crc kubenswrapper[4612]: I0227 09:05:16.040952 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-mchzb/must-gather-t4rbr" Feb 27 09:05:16 crc kubenswrapper[4612]: I0227 09:05:16.071247 4612 scope.go:117] "RemoveContainer" containerID="e6328b790dfbc8f52b3131fc81f90e6ddef8bfc38dc9e78acb8c6353c4afc402" Feb 27 09:05:16 crc kubenswrapper[4612]: I0227 09:05:16.098842 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3f06074f-6e22-4526-a04b-e842577c6f2e-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "3f06074f-6e22-4526-a04b-e842577c6f2e" (UID: "3f06074f-6e22-4526-a04b-e842577c6f2e"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 09:05:16 crc kubenswrapper[4612]: I0227 09:05:16.130293 4612 scope.go:117] "RemoveContainer" containerID="b9508d72b81df0d55578ad38e42b9f06bdb5f38cb474a26239d791fdd4d9848e" Feb 27 09:05:16 crc kubenswrapper[4612]: E0227 09:05:16.130743 4612 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b9508d72b81df0d55578ad38e42b9f06bdb5f38cb474a26239d791fdd4d9848e\": container with ID starting with b9508d72b81df0d55578ad38e42b9f06bdb5f38cb474a26239d791fdd4d9848e not found: ID does not exist" containerID="b9508d72b81df0d55578ad38e42b9f06bdb5f38cb474a26239d791fdd4d9848e" Feb 27 09:05:16 crc kubenswrapper[4612]: I0227 09:05:16.130771 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b9508d72b81df0d55578ad38e42b9f06bdb5f38cb474a26239d791fdd4d9848e"} err="failed to get container status \"b9508d72b81df0d55578ad38e42b9f06bdb5f38cb474a26239d791fdd4d9848e\": rpc error: code = NotFound desc = could not find container \"b9508d72b81df0d55578ad38e42b9f06bdb5f38cb474a26239d791fdd4d9848e\": container with ID starting with b9508d72b81df0d55578ad38e42b9f06bdb5f38cb474a26239d791fdd4d9848e not found: ID does not exist" Feb 27 09:05:16 crc kubenswrapper[4612]: I0227 09:05:16.130791 4612 scope.go:117] "RemoveContainer" containerID="e6328b790dfbc8f52b3131fc81f90e6ddef8bfc38dc9e78acb8c6353c4afc402" Feb 27 09:05:16 crc kubenswrapper[4612]: E0227 09:05:16.130996 4612 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e6328b790dfbc8f52b3131fc81f90e6ddef8bfc38dc9e78acb8c6353c4afc402\": container with ID starting with e6328b790dfbc8f52b3131fc81f90e6ddef8bfc38dc9e78acb8c6353c4afc402 not found: ID does not exist" containerID="e6328b790dfbc8f52b3131fc81f90e6ddef8bfc38dc9e78acb8c6353c4afc402" Feb 27 09:05:16 crc kubenswrapper[4612]: I0227 09:05:16.131011 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e6328b790dfbc8f52b3131fc81f90e6ddef8bfc38dc9e78acb8c6353c4afc402"} err="failed to get container status \"e6328b790dfbc8f52b3131fc81f90e6ddef8bfc38dc9e78acb8c6353c4afc402\": rpc error: code = NotFound desc = could not find container \"e6328b790dfbc8f52b3131fc81f90e6ddef8bfc38dc9e78acb8c6353c4afc402\": container with ID starting with e6328b790dfbc8f52b3131fc81f90e6ddef8bfc38dc9e78acb8c6353c4afc402 not found: ID does not exist" Feb 27 09:05:16 crc kubenswrapper[4612]: I0227 09:05:16.162387 4612 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/3f06074f-6e22-4526-a04b-e842577c6f2e-must-gather-output\") on node \"crc\" DevicePath \"\"" Feb 27 09:05:16 crc kubenswrapper[4612]: I0227 09:05:16.864104 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3f06074f-6e22-4526-a04b-e842577c6f2e" path="/var/lib/kubelet/pods/3f06074f-6e22-4526-a04b-e842577c6f2e/volumes" Feb 27 09:05:20 crc kubenswrapper[4612]: I0227 09:05:20.680528 4612 scope.go:117] "RemoveContainer" containerID="48ebf8c43d201cdb1c751905cabb0358a44ccf8fa725e65bf0cd190df29e5d51" Feb 27 09:05:21 crc kubenswrapper[4612]: I0227 09:05:21.853254 4612 scope.go:117] "RemoveContainer" containerID="aa498fa4b5f3843852a1ebe0b165b06254525eeb49f5224c7bd2a0f8c2ae2480" Feb 27 09:05:21 crc kubenswrapper[4612]: E0227 09:05:21.854631 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-924vb_openshift-machine-config-operator(24599373-7adc-4a1b-8bb4-797bf726f43d)\"" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" Feb 27 09:05:33 crc kubenswrapper[4612]: I0227 09:05:33.853001 4612 scope.go:117] "RemoveContainer" containerID="aa498fa4b5f3843852a1ebe0b165b06254525eeb49f5224c7bd2a0f8c2ae2480" Feb 27 09:05:33 crc kubenswrapper[4612]: E0227 09:05:33.853823 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-924vb_openshift-machine-config-operator(24599373-7adc-4a1b-8bb4-797bf726f43d)\"" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" Feb 27 09:05:48 crc kubenswrapper[4612]: I0227 09:05:48.854249 4612 scope.go:117] "RemoveContainer" containerID="aa498fa4b5f3843852a1ebe0b165b06254525eeb49f5224c7bd2a0f8c2ae2480" Feb 27 09:05:48 crc kubenswrapper[4612]: E0227 09:05:48.854930 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-924vb_openshift-machine-config-operator(24599373-7adc-4a1b-8bb4-797bf726f43d)\"" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" Feb 27 09:05:59 crc kubenswrapper[4612]: I0227 09:05:59.852878 4612 scope.go:117] "RemoveContainer" containerID="aa498fa4b5f3843852a1ebe0b165b06254525eeb49f5224c7bd2a0f8c2ae2480" Feb 27 09:05:59 crc kubenswrapper[4612]: E0227 09:05:59.853870 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-924vb_openshift-machine-config-operator(24599373-7adc-4a1b-8bb4-797bf726f43d)\"" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" Feb 27 09:06:00 crc kubenswrapper[4612]: I0227 09:06:00.162573 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29536386-j6cvj"] Feb 27 09:06:00 crc kubenswrapper[4612]: E0227 09:06:00.163137 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fffd01d6-8461-4815-bd7f-52b0fb816346" containerName="oc" Feb 27 09:06:00 crc kubenswrapper[4612]: I0227 09:06:00.163164 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="fffd01d6-8461-4815-bd7f-52b0fb816346" containerName="oc" Feb 27 09:06:00 crc kubenswrapper[4612]: E0227 09:06:00.163208 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3f06074f-6e22-4526-a04b-e842577c6f2e" containerName="gather" Feb 27 09:06:00 crc kubenswrapper[4612]: I0227 09:06:00.163221 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="3f06074f-6e22-4526-a04b-e842577c6f2e" containerName="gather" Feb 27 09:06:00 crc kubenswrapper[4612]: E0227 09:06:00.163253 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3f06074f-6e22-4526-a04b-e842577c6f2e" containerName="copy" Feb 27 09:06:00 crc kubenswrapper[4612]: I0227 09:06:00.163265 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="3f06074f-6e22-4526-a04b-e842577c6f2e" containerName="copy" Feb 27 09:06:00 crc kubenswrapper[4612]: I0227 09:06:00.163572 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="fffd01d6-8461-4815-bd7f-52b0fb816346" containerName="oc" Feb 27 09:06:00 crc kubenswrapper[4612]: I0227 09:06:00.163624 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="3f06074f-6e22-4526-a04b-e842577c6f2e" containerName="gather" Feb 27 09:06:00 crc kubenswrapper[4612]: I0227 09:06:00.163646 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="3f06074f-6e22-4526-a04b-e842577c6f2e" containerName="copy" Feb 27 09:06:00 crc kubenswrapper[4612]: I0227 09:06:00.164522 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536386-j6cvj" Feb 27 09:06:00 crc kubenswrapper[4612]: I0227 09:06:00.170018 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 27 09:06:00 crc kubenswrapper[4612]: I0227 09:06:00.170387 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 27 09:06:00 crc kubenswrapper[4612]: I0227 09:06:00.172402 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536386-j6cvj"] Feb 27 09:06:00 crc kubenswrapper[4612]: I0227 09:06:00.184758 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-5zwfx" Feb 27 09:06:00 crc kubenswrapper[4612]: I0227 09:06:00.279314 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s8h6n\" (UniqueName: \"kubernetes.io/projected/018188bb-5ac2-4e9a-8bc0-7f7c1ead9115-kube-api-access-s8h6n\") pod \"auto-csr-approver-29536386-j6cvj\" (UID: \"018188bb-5ac2-4e9a-8bc0-7f7c1ead9115\") " pod="openshift-infra/auto-csr-approver-29536386-j6cvj" Feb 27 09:06:00 crc kubenswrapper[4612]: I0227 09:06:00.381092 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s8h6n\" (UniqueName: \"kubernetes.io/projected/018188bb-5ac2-4e9a-8bc0-7f7c1ead9115-kube-api-access-s8h6n\") pod \"auto-csr-approver-29536386-j6cvj\" (UID: \"018188bb-5ac2-4e9a-8bc0-7f7c1ead9115\") " pod="openshift-infra/auto-csr-approver-29536386-j6cvj" Feb 27 09:06:00 crc kubenswrapper[4612]: I0227 09:06:00.402424 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s8h6n\" (UniqueName: \"kubernetes.io/projected/018188bb-5ac2-4e9a-8bc0-7f7c1ead9115-kube-api-access-s8h6n\") pod \"auto-csr-approver-29536386-j6cvj\" (UID: \"018188bb-5ac2-4e9a-8bc0-7f7c1ead9115\") " pod="openshift-infra/auto-csr-approver-29536386-j6cvj" Feb 27 09:06:00 crc kubenswrapper[4612]: I0227 09:06:00.495004 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536386-j6cvj" Feb 27 09:06:01 crc kubenswrapper[4612]: I0227 09:06:01.135477 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536386-j6cvj"] Feb 27 09:06:01 crc kubenswrapper[4612]: I0227 09:06:01.143184 4612 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 27 09:06:01 crc kubenswrapper[4612]: I0227 09:06:01.501455 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536386-j6cvj" event={"ID":"018188bb-5ac2-4e9a-8bc0-7f7c1ead9115","Type":"ContainerStarted","Data":"be00d3322a3c49f8ebd36a110c13745d11b096fcb1e7cb63902517b622b86f1d"} Feb 27 09:06:02 crc kubenswrapper[4612]: I0227 09:06:02.511662 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536386-j6cvj" event={"ID":"018188bb-5ac2-4e9a-8bc0-7f7c1ead9115","Type":"ContainerStarted","Data":"81f9b9bc738c2d1a0d26637aac53e51282a564cb0cc257319f7a8be3d4c7f5e5"} Feb 27 09:06:02 crc kubenswrapper[4612]: I0227 09:06:02.530756 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29536386-j6cvj" podStartSLOduration=1.649808069 podStartE2EDuration="2.530737175s" podCreationTimestamp="2026-02-27 09:06:00 +0000 UTC" firstStartedPulling="2026-02-27 09:06:01.142955257 +0000 UTC m=+4618.996885255" lastFinishedPulling="2026-02-27 09:06:02.023884363 +0000 UTC m=+4619.877814361" observedRunningTime="2026-02-27 09:06:02.526988598 +0000 UTC m=+4620.380918596" watchObservedRunningTime="2026-02-27 09:06:02.530737175 +0000 UTC m=+4620.384667173" Feb 27 09:06:03 crc kubenswrapper[4612]: I0227 09:06:03.523857 4612 generic.go:334] "Generic (PLEG): container finished" podID="018188bb-5ac2-4e9a-8bc0-7f7c1ead9115" containerID="81f9b9bc738c2d1a0d26637aac53e51282a564cb0cc257319f7a8be3d4c7f5e5" exitCode=0 Feb 27 09:06:03 crc kubenswrapper[4612]: I0227 09:06:03.524158 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536386-j6cvj" event={"ID":"018188bb-5ac2-4e9a-8bc0-7f7c1ead9115","Type":"ContainerDied","Data":"81f9b9bc738c2d1a0d26637aac53e51282a564cb0cc257319f7a8be3d4c7f5e5"} Feb 27 09:06:04 crc kubenswrapper[4612]: I0227 09:06:04.928219 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536386-j6cvj" Feb 27 09:06:04 crc kubenswrapper[4612]: I0227 09:06:04.976275 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s8h6n\" (UniqueName: \"kubernetes.io/projected/018188bb-5ac2-4e9a-8bc0-7f7c1ead9115-kube-api-access-s8h6n\") pod \"018188bb-5ac2-4e9a-8bc0-7f7c1ead9115\" (UID: \"018188bb-5ac2-4e9a-8bc0-7f7c1ead9115\") " Feb 27 09:06:04 crc kubenswrapper[4612]: I0227 09:06:04.989655 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/018188bb-5ac2-4e9a-8bc0-7f7c1ead9115-kube-api-access-s8h6n" (OuterVolumeSpecName: "kube-api-access-s8h6n") pod "018188bb-5ac2-4e9a-8bc0-7f7c1ead9115" (UID: "018188bb-5ac2-4e9a-8bc0-7f7c1ead9115"). InnerVolumeSpecName "kube-api-access-s8h6n". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 09:06:05 crc kubenswrapper[4612]: I0227 09:06:05.078937 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s8h6n\" (UniqueName: \"kubernetes.io/projected/018188bb-5ac2-4e9a-8bc0-7f7c1ead9115-kube-api-access-s8h6n\") on node \"crc\" DevicePath \"\"" Feb 27 09:06:05 crc kubenswrapper[4612]: I0227 09:06:05.540919 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536386-j6cvj" event={"ID":"018188bb-5ac2-4e9a-8bc0-7f7c1ead9115","Type":"ContainerDied","Data":"be00d3322a3c49f8ebd36a110c13745d11b096fcb1e7cb63902517b622b86f1d"} Feb 27 09:06:05 crc kubenswrapper[4612]: I0227 09:06:05.541242 4612 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="be00d3322a3c49f8ebd36a110c13745d11b096fcb1e7cb63902517b622b86f1d" Feb 27 09:06:05 crc kubenswrapper[4612]: I0227 09:06:05.540989 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536386-j6cvj" Feb 27 09:06:05 crc kubenswrapper[4612]: I0227 09:06:05.605492 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29536380-wkqc6"] Feb 27 09:06:05 crc kubenswrapper[4612]: I0227 09:06:05.613934 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29536380-wkqc6"] Feb 27 09:06:06 crc kubenswrapper[4612]: I0227 09:06:06.865626 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="45882240-d427-4e23-9415-100cbdcedb42" path="/var/lib/kubelet/pods/45882240-d427-4e23-9415-100cbdcedb42/volumes" Feb 27 09:06:14 crc kubenswrapper[4612]: I0227 09:06:14.853415 4612 scope.go:117] "RemoveContainer" containerID="aa498fa4b5f3843852a1ebe0b165b06254525eeb49f5224c7bd2a0f8c2ae2480" Feb 27 09:06:14 crc kubenswrapper[4612]: E0227 09:06:14.856470 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-924vb_openshift-machine-config-operator(24599373-7adc-4a1b-8bb4-797bf726f43d)\"" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" Feb 27 09:06:20 crc kubenswrapper[4612]: I0227 09:06:20.760257 4612 scope.go:117] "RemoveContainer" containerID="265cc3d5c79c7905f996507b86bb705ce59119077ceb548e02cc90ec3327c871" Feb 27 09:06:25 crc kubenswrapper[4612]: I0227 09:06:25.853640 4612 scope.go:117] "RemoveContainer" containerID="aa498fa4b5f3843852a1ebe0b165b06254525eeb49f5224c7bd2a0f8c2ae2480" Feb 27 09:06:26 crc kubenswrapper[4612]: I0227 09:06:26.761439 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-924vb" event={"ID":"24599373-7adc-4a1b-8bb4-797bf726f43d","Type":"ContainerStarted","Data":"c90278d5d8864d43287d88f2fcf48602a1065bde9d8b6d625ab9498edb9b58db"} Feb 27 09:06:43 crc kubenswrapper[4612]: I0227 09:06:43.417033 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-d5dgw"] Feb 27 09:06:43 crc kubenswrapper[4612]: E0227 09:06:43.418240 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="018188bb-5ac2-4e9a-8bc0-7f7c1ead9115" containerName="oc" Feb 27 09:06:43 crc kubenswrapper[4612]: I0227 09:06:43.418264 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="018188bb-5ac2-4e9a-8bc0-7f7c1ead9115" containerName="oc" Feb 27 09:06:43 crc kubenswrapper[4612]: I0227 09:06:43.418594 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="018188bb-5ac2-4e9a-8bc0-7f7c1ead9115" containerName="oc" Feb 27 09:06:43 crc kubenswrapper[4612]: I0227 09:06:43.421006 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-d5dgw" Feb 27 09:06:43 crc kubenswrapper[4612]: I0227 09:06:43.439101 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-d5dgw"] Feb 27 09:06:43 crc kubenswrapper[4612]: I0227 09:06:43.589777 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-62jqc\" (UniqueName: \"kubernetes.io/projected/46a23f1f-31c1-4fec-b794-2c6bb04cacc5-kube-api-access-62jqc\") pod \"redhat-marketplace-d5dgw\" (UID: \"46a23f1f-31c1-4fec-b794-2c6bb04cacc5\") " pod="openshift-marketplace/redhat-marketplace-d5dgw" Feb 27 09:06:43 crc kubenswrapper[4612]: I0227 09:06:43.590129 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/46a23f1f-31c1-4fec-b794-2c6bb04cacc5-utilities\") pod \"redhat-marketplace-d5dgw\" (UID: \"46a23f1f-31c1-4fec-b794-2c6bb04cacc5\") " pod="openshift-marketplace/redhat-marketplace-d5dgw" Feb 27 09:06:43 crc kubenswrapper[4612]: I0227 09:06:43.590217 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/46a23f1f-31c1-4fec-b794-2c6bb04cacc5-catalog-content\") pod \"redhat-marketplace-d5dgw\" (UID: \"46a23f1f-31c1-4fec-b794-2c6bb04cacc5\") " pod="openshift-marketplace/redhat-marketplace-d5dgw" Feb 27 09:06:43 crc kubenswrapper[4612]: I0227 09:06:43.692596 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/46a23f1f-31c1-4fec-b794-2c6bb04cacc5-catalog-content\") pod \"redhat-marketplace-d5dgw\" (UID: \"46a23f1f-31c1-4fec-b794-2c6bb04cacc5\") " pod="openshift-marketplace/redhat-marketplace-d5dgw" Feb 27 09:06:43 crc kubenswrapper[4612]: I0227 09:06:43.692794 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-62jqc\" (UniqueName: \"kubernetes.io/projected/46a23f1f-31c1-4fec-b794-2c6bb04cacc5-kube-api-access-62jqc\") pod \"redhat-marketplace-d5dgw\" (UID: \"46a23f1f-31c1-4fec-b794-2c6bb04cacc5\") " pod="openshift-marketplace/redhat-marketplace-d5dgw" Feb 27 09:06:43 crc kubenswrapper[4612]: I0227 09:06:43.692910 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/46a23f1f-31c1-4fec-b794-2c6bb04cacc5-utilities\") pod \"redhat-marketplace-d5dgw\" (UID: \"46a23f1f-31c1-4fec-b794-2c6bb04cacc5\") " pod="openshift-marketplace/redhat-marketplace-d5dgw" Feb 27 09:06:43 crc kubenswrapper[4612]: I0227 09:06:43.693256 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/46a23f1f-31c1-4fec-b794-2c6bb04cacc5-catalog-content\") pod \"redhat-marketplace-d5dgw\" (UID: \"46a23f1f-31c1-4fec-b794-2c6bb04cacc5\") " pod="openshift-marketplace/redhat-marketplace-d5dgw" Feb 27 09:06:43 crc kubenswrapper[4612]: I0227 09:06:43.693422 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/46a23f1f-31c1-4fec-b794-2c6bb04cacc5-utilities\") pod \"redhat-marketplace-d5dgw\" (UID: \"46a23f1f-31c1-4fec-b794-2c6bb04cacc5\") " pod="openshift-marketplace/redhat-marketplace-d5dgw" Feb 27 09:06:43 crc kubenswrapper[4612]: I0227 09:06:43.736360 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-62jqc\" (UniqueName: \"kubernetes.io/projected/46a23f1f-31c1-4fec-b794-2c6bb04cacc5-kube-api-access-62jqc\") pod \"redhat-marketplace-d5dgw\" (UID: \"46a23f1f-31c1-4fec-b794-2c6bb04cacc5\") " pod="openshift-marketplace/redhat-marketplace-d5dgw" Feb 27 09:06:43 crc kubenswrapper[4612]: I0227 09:06:43.759516 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-d5dgw" Feb 27 09:06:44 crc kubenswrapper[4612]: I0227 09:06:44.218030 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-d5dgw"] Feb 27 09:06:44 crc kubenswrapper[4612]: W0227 09:06:44.228899 4612 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod46a23f1f_31c1_4fec_b794_2c6bb04cacc5.slice/crio-03cc5ed1dacbde504934fe9da7d3e0f1d905f30e3f77af1fe3f8c683d2935076 WatchSource:0}: Error finding container 03cc5ed1dacbde504934fe9da7d3e0f1d905f30e3f77af1fe3f8c683d2935076: Status 404 returned error can't find the container with id 03cc5ed1dacbde504934fe9da7d3e0f1d905f30e3f77af1fe3f8c683d2935076 Feb 27 09:06:44 crc kubenswrapper[4612]: I0227 09:06:44.961228 4612 generic.go:334] "Generic (PLEG): container finished" podID="46a23f1f-31c1-4fec-b794-2c6bb04cacc5" containerID="ddbf1a9d7b91f756b4f40cb58b0b37b2bca374b593580f06771c4baedbe8721a" exitCode=0 Feb 27 09:06:44 crc kubenswrapper[4612]: I0227 09:06:44.961418 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-d5dgw" event={"ID":"46a23f1f-31c1-4fec-b794-2c6bb04cacc5","Type":"ContainerDied","Data":"ddbf1a9d7b91f756b4f40cb58b0b37b2bca374b593580f06771c4baedbe8721a"} Feb 27 09:06:44 crc kubenswrapper[4612]: I0227 09:06:44.961941 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-d5dgw" event={"ID":"46a23f1f-31c1-4fec-b794-2c6bb04cacc5","Type":"ContainerStarted","Data":"03cc5ed1dacbde504934fe9da7d3e0f1d905f30e3f77af1fe3f8c683d2935076"} Feb 27 09:06:45 crc kubenswrapper[4612]: I0227 09:06:45.978341 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-d5dgw" event={"ID":"46a23f1f-31c1-4fec-b794-2c6bb04cacc5","Type":"ContainerStarted","Data":"4c76d331d16ca43b25fd3b67d6def265bc1d510eb16e245708b87c526a80a05c"} Feb 27 09:06:46 crc kubenswrapper[4612]: I0227 09:06:46.988505 4612 generic.go:334] "Generic (PLEG): container finished" podID="46a23f1f-31c1-4fec-b794-2c6bb04cacc5" containerID="4c76d331d16ca43b25fd3b67d6def265bc1d510eb16e245708b87c526a80a05c" exitCode=0 Feb 27 09:06:46 crc kubenswrapper[4612]: I0227 09:06:46.988556 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-d5dgw" event={"ID":"46a23f1f-31c1-4fec-b794-2c6bb04cacc5","Type":"ContainerDied","Data":"4c76d331d16ca43b25fd3b67d6def265bc1d510eb16e245708b87c526a80a05c"} Feb 27 09:06:47 crc kubenswrapper[4612]: I0227 09:06:47.999344 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-d5dgw" event={"ID":"46a23f1f-31c1-4fec-b794-2c6bb04cacc5","Type":"ContainerStarted","Data":"43bae0cf04b7abf30b4db0c24780c0f7fa7f0adf91c3ca0ffa5e802511e6c89f"} Feb 27 09:06:48 crc kubenswrapper[4612]: I0227 09:06:48.031219 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-d5dgw" podStartSLOduration=2.5612646100000003 podStartE2EDuration="5.031196748s" podCreationTimestamp="2026-02-27 09:06:43 +0000 UTC" firstStartedPulling="2026-02-27 09:06:44.970293807 +0000 UTC m=+4662.824223825" lastFinishedPulling="2026-02-27 09:06:47.440225955 +0000 UTC m=+4665.294155963" observedRunningTime="2026-02-27 09:06:48.025584269 +0000 UTC m=+4665.879514297" watchObservedRunningTime="2026-02-27 09:06:48.031196748 +0000 UTC m=+4665.885126756" Feb 27 09:06:53 crc kubenswrapper[4612]: I0227 09:06:53.760205 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-d5dgw" Feb 27 09:06:53 crc kubenswrapper[4612]: I0227 09:06:53.761618 4612 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-d5dgw" Feb 27 09:06:53 crc kubenswrapper[4612]: I0227 09:06:53.811777 4612 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-d5dgw" Feb 27 09:06:54 crc kubenswrapper[4612]: I0227 09:06:54.103740 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-d5dgw" Feb 27 09:06:54 crc kubenswrapper[4612]: I0227 09:06:54.149985 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-d5dgw"] Feb 27 09:06:56 crc kubenswrapper[4612]: I0227 09:06:56.076399 4612 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-d5dgw" podUID="46a23f1f-31c1-4fec-b794-2c6bb04cacc5" containerName="registry-server" containerID="cri-o://43bae0cf04b7abf30b4db0c24780c0f7fa7f0adf91c3ca0ffa5e802511e6c89f" gracePeriod=2 Feb 27 09:06:56 crc kubenswrapper[4612]: I0227 09:06:56.494465 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-d5dgw" Feb 27 09:06:56 crc kubenswrapper[4612]: I0227 09:06:56.662798 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/46a23f1f-31c1-4fec-b794-2c6bb04cacc5-catalog-content\") pod \"46a23f1f-31c1-4fec-b794-2c6bb04cacc5\" (UID: \"46a23f1f-31c1-4fec-b794-2c6bb04cacc5\") " Feb 27 09:06:56 crc kubenswrapper[4612]: I0227 09:06:56.662853 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/46a23f1f-31c1-4fec-b794-2c6bb04cacc5-utilities\") pod \"46a23f1f-31c1-4fec-b794-2c6bb04cacc5\" (UID: \"46a23f1f-31c1-4fec-b794-2c6bb04cacc5\") " Feb 27 09:06:56 crc kubenswrapper[4612]: I0227 09:06:56.662970 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-62jqc\" (UniqueName: \"kubernetes.io/projected/46a23f1f-31c1-4fec-b794-2c6bb04cacc5-kube-api-access-62jqc\") pod \"46a23f1f-31c1-4fec-b794-2c6bb04cacc5\" (UID: \"46a23f1f-31c1-4fec-b794-2c6bb04cacc5\") " Feb 27 09:06:56 crc kubenswrapper[4612]: I0227 09:06:56.668458 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/46a23f1f-31c1-4fec-b794-2c6bb04cacc5-utilities" (OuterVolumeSpecName: "utilities") pod "46a23f1f-31c1-4fec-b794-2c6bb04cacc5" (UID: "46a23f1f-31c1-4fec-b794-2c6bb04cacc5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 09:06:56 crc kubenswrapper[4612]: I0227 09:06:56.673277 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/46a23f1f-31c1-4fec-b794-2c6bb04cacc5-kube-api-access-62jqc" (OuterVolumeSpecName: "kube-api-access-62jqc") pod "46a23f1f-31c1-4fec-b794-2c6bb04cacc5" (UID: "46a23f1f-31c1-4fec-b794-2c6bb04cacc5"). InnerVolumeSpecName "kube-api-access-62jqc". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 09:06:56 crc kubenswrapper[4612]: I0227 09:06:56.765132 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-62jqc\" (UniqueName: \"kubernetes.io/projected/46a23f1f-31c1-4fec-b794-2c6bb04cacc5-kube-api-access-62jqc\") on node \"crc\" DevicePath \"\"" Feb 27 09:06:56 crc kubenswrapper[4612]: I0227 09:06:56.765163 4612 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/46a23f1f-31c1-4fec-b794-2c6bb04cacc5-utilities\") on node \"crc\" DevicePath \"\"" Feb 27 09:06:56 crc kubenswrapper[4612]: I0227 09:06:56.891561 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/46a23f1f-31c1-4fec-b794-2c6bb04cacc5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "46a23f1f-31c1-4fec-b794-2c6bb04cacc5" (UID: "46a23f1f-31c1-4fec-b794-2c6bb04cacc5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 09:06:56 crc kubenswrapper[4612]: I0227 09:06:56.969670 4612 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/46a23f1f-31c1-4fec-b794-2c6bb04cacc5-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 27 09:06:57 crc kubenswrapper[4612]: I0227 09:06:57.092487 4612 generic.go:334] "Generic (PLEG): container finished" podID="46a23f1f-31c1-4fec-b794-2c6bb04cacc5" containerID="43bae0cf04b7abf30b4db0c24780c0f7fa7f0adf91c3ca0ffa5e802511e6c89f" exitCode=0 Feb 27 09:06:57 crc kubenswrapper[4612]: I0227 09:06:57.092534 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-d5dgw" event={"ID":"46a23f1f-31c1-4fec-b794-2c6bb04cacc5","Type":"ContainerDied","Data":"43bae0cf04b7abf30b4db0c24780c0f7fa7f0adf91c3ca0ffa5e802511e6c89f"} Feb 27 09:06:57 crc kubenswrapper[4612]: I0227 09:06:57.092567 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-d5dgw" Feb 27 09:06:57 crc kubenswrapper[4612]: I0227 09:06:57.092615 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-d5dgw" event={"ID":"46a23f1f-31c1-4fec-b794-2c6bb04cacc5","Type":"ContainerDied","Data":"03cc5ed1dacbde504934fe9da7d3e0f1d905f30e3f77af1fe3f8c683d2935076"} Feb 27 09:06:57 crc kubenswrapper[4612]: I0227 09:06:57.092656 4612 scope.go:117] "RemoveContainer" containerID="43bae0cf04b7abf30b4db0c24780c0f7fa7f0adf91c3ca0ffa5e802511e6c89f" Feb 27 09:06:57 crc kubenswrapper[4612]: I0227 09:06:57.126516 4612 scope.go:117] "RemoveContainer" containerID="4c76d331d16ca43b25fd3b67d6def265bc1d510eb16e245708b87c526a80a05c" Feb 27 09:06:57 crc kubenswrapper[4612]: I0227 09:06:57.162019 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-d5dgw"] Feb 27 09:06:57 crc kubenswrapper[4612]: I0227 09:06:57.179025 4612 scope.go:117] "RemoveContainer" containerID="ddbf1a9d7b91f756b4f40cb58b0b37b2bca374b593580f06771c4baedbe8721a" Feb 27 09:06:57 crc kubenswrapper[4612]: I0227 09:06:57.180114 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-d5dgw"] Feb 27 09:06:57 crc kubenswrapper[4612]: I0227 09:06:57.209951 4612 scope.go:117] "RemoveContainer" containerID="43bae0cf04b7abf30b4db0c24780c0f7fa7f0adf91c3ca0ffa5e802511e6c89f" Feb 27 09:06:57 crc kubenswrapper[4612]: E0227 09:06:57.210303 4612 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"43bae0cf04b7abf30b4db0c24780c0f7fa7f0adf91c3ca0ffa5e802511e6c89f\": container with ID starting with 43bae0cf04b7abf30b4db0c24780c0f7fa7f0adf91c3ca0ffa5e802511e6c89f not found: ID does not exist" containerID="43bae0cf04b7abf30b4db0c24780c0f7fa7f0adf91c3ca0ffa5e802511e6c89f" Feb 27 09:06:57 crc kubenswrapper[4612]: I0227 09:06:57.210347 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"43bae0cf04b7abf30b4db0c24780c0f7fa7f0adf91c3ca0ffa5e802511e6c89f"} err="failed to get container status \"43bae0cf04b7abf30b4db0c24780c0f7fa7f0adf91c3ca0ffa5e802511e6c89f\": rpc error: code = NotFound desc = could not find container \"43bae0cf04b7abf30b4db0c24780c0f7fa7f0adf91c3ca0ffa5e802511e6c89f\": container with ID starting with 43bae0cf04b7abf30b4db0c24780c0f7fa7f0adf91c3ca0ffa5e802511e6c89f not found: ID does not exist" Feb 27 09:06:57 crc kubenswrapper[4612]: I0227 09:06:57.210377 4612 scope.go:117] "RemoveContainer" containerID="4c76d331d16ca43b25fd3b67d6def265bc1d510eb16e245708b87c526a80a05c" Feb 27 09:06:57 crc kubenswrapper[4612]: E0227 09:06:57.210646 4612 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4c76d331d16ca43b25fd3b67d6def265bc1d510eb16e245708b87c526a80a05c\": container with ID starting with 4c76d331d16ca43b25fd3b67d6def265bc1d510eb16e245708b87c526a80a05c not found: ID does not exist" containerID="4c76d331d16ca43b25fd3b67d6def265bc1d510eb16e245708b87c526a80a05c" Feb 27 09:06:57 crc kubenswrapper[4612]: I0227 09:06:57.210678 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4c76d331d16ca43b25fd3b67d6def265bc1d510eb16e245708b87c526a80a05c"} err="failed to get container status \"4c76d331d16ca43b25fd3b67d6def265bc1d510eb16e245708b87c526a80a05c\": rpc error: code = NotFound desc = could not find container \"4c76d331d16ca43b25fd3b67d6def265bc1d510eb16e245708b87c526a80a05c\": container with ID starting with 4c76d331d16ca43b25fd3b67d6def265bc1d510eb16e245708b87c526a80a05c not found: ID does not exist" Feb 27 09:06:57 crc kubenswrapper[4612]: I0227 09:06:57.210803 4612 scope.go:117] "RemoveContainer" containerID="ddbf1a9d7b91f756b4f40cb58b0b37b2bca374b593580f06771c4baedbe8721a" Feb 27 09:06:57 crc kubenswrapper[4612]: E0227 09:06:57.211240 4612 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ddbf1a9d7b91f756b4f40cb58b0b37b2bca374b593580f06771c4baedbe8721a\": container with ID starting with ddbf1a9d7b91f756b4f40cb58b0b37b2bca374b593580f06771c4baedbe8721a not found: ID does not exist" containerID="ddbf1a9d7b91f756b4f40cb58b0b37b2bca374b593580f06771c4baedbe8721a" Feb 27 09:06:57 crc kubenswrapper[4612]: I0227 09:06:57.211266 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ddbf1a9d7b91f756b4f40cb58b0b37b2bca374b593580f06771c4baedbe8721a"} err="failed to get container status \"ddbf1a9d7b91f756b4f40cb58b0b37b2bca374b593580f06771c4baedbe8721a\": rpc error: code = NotFound desc = could not find container \"ddbf1a9d7b91f756b4f40cb58b0b37b2bca374b593580f06771c4baedbe8721a\": container with ID starting with ddbf1a9d7b91f756b4f40cb58b0b37b2bca374b593580f06771c4baedbe8721a not found: ID does not exist" Feb 27 09:06:58 crc kubenswrapper[4612]: I0227 09:06:58.886776 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="46a23f1f-31c1-4fec-b794-2c6bb04cacc5" path="/var/lib/kubelet/pods/46a23f1f-31c1-4fec-b794-2c6bb04cacc5/volumes" Feb 27 09:07:13 crc kubenswrapper[4612]: I0227 09:07:13.377546 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-4pkt7"] Feb 27 09:07:13 crc kubenswrapper[4612]: E0227 09:07:13.378825 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="46a23f1f-31c1-4fec-b794-2c6bb04cacc5" containerName="extract-content" Feb 27 09:07:13 crc kubenswrapper[4612]: I0227 09:07:13.378845 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="46a23f1f-31c1-4fec-b794-2c6bb04cacc5" containerName="extract-content" Feb 27 09:07:13 crc kubenswrapper[4612]: E0227 09:07:13.378866 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="46a23f1f-31c1-4fec-b794-2c6bb04cacc5" containerName="extract-utilities" Feb 27 09:07:13 crc kubenswrapper[4612]: I0227 09:07:13.378878 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="46a23f1f-31c1-4fec-b794-2c6bb04cacc5" containerName="extract-utilities" Feb 27 09:07:13 crc kubenswrapper[4612]: E0227 09:07:13.378912 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="46a23f1f-31c1-4fec-b794-2c6bb04cacc5" containerName="registry-server" Feb 27 09:07:13 crc kubenswrapper[4612]: I0227 09:07:13.378921 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="46a23f1f-31c1-4fec-b794-2c6bb04cacc5" containerName="registry-server" Feb 27 09:07:13 crc kubenswrapper[4612]: I0227 09:07:13.379163 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="46a23f1f-31c1-4fec-b794-2c6bb04cacc5" containerName="registry-server" Feb 27 09:07:13 crc kubenswrapper[4612]: I0227 09:07:13.380927 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4pkt7" Feb 27 09:07:13 crc kubenswrapper[4612]: I0227 09:07:13.391257 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-4pkt7"] Feb 27 09:07:13 crc kubenswrapper[4612]: I0227 09:07:13.458682 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wc2gd\" (UniqueName: \"kubernetes.io/projected/f4cdffae-731e-41f7-96d8-b015a5576a18-kube-api-access-wc2gd\") pod \"certified-operators-4pkt7\" (UID: \"f4cdffae-731e-41f7-96d8-b015a5576a18\") " pod="openshift-marketplace/certified-operators-4pkt7" Feb 27 09:07:13 crc kubenswrapper[4612]: I0227 09:07:13.458804 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f4cdffae-731e-41f7-96d8-b015a5576a18-utilities\") pod \"certified-operators-4pkt7\" (UID: \"f4cdffae-731e-41f7-96d8-b015a5576a18\") " pod="openshift-marketplace/certified-operators-4pkt7" Feb 27 09:07:13 crc kubenswrapper[4612]: I0227 09:07:13.458867 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f4cdffae-731e-41f7-96d8-b015a5576a18-catalog-content\") pod \"certified-operators-4pkt7\" (UID: \"f4cdffae-731e-41f7-96d8-b015a5576a18\") " pod="openshift-marketplace/certified-operators-4pkt7" Feb 27 09:07:13 crc kubenswrapper[4612]: I0227 09:07:13.560828 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wc2gd\" (UniqueName: \"kubernetes.io/projected/f4cdffae-731e-41f7-96d8-b015a5576a18-kube-api-access-wc2gd\") pod \"certified-operators-4pkt7\" (UID: \"f4cdffae-731e-41f7-96d8-b015a5576a18\") " pod="openshift-marketplace/certified-operators-4pkt7" Feb 27 09:07:13 crc kubenswrapper[4612]: I0227 09:07:13.560974 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f4cdffae-731e-41f7-96d8-b015a5576a18-utilities\") pod \"certified-operators-4pkt7\" (UID: \"f4cdffae-731e-41f7-96d8-b015a5576a18\") " pod="openshift-marketplace/certified-operators-4pkt7" Feb 27 09:07:13 crc kubenswrapper[4612]: I0227 09:07:13.561016 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f4cdffae-731e-41f7-96d8-b015a5576a18-catalog-content\") pod \"certified-operators-4pkt7\" (UID: \"f4cdffae-731e-41f7-96d8-b015a5576a18\") " pod="openshift-marketplace/certified-operators-4pkt7" Feb 27 09:07:13 crc kubenswrapper[4612]: I0227 09:07:13.561463 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f4cdffae-731e-41f7-96d8-b015a5576a18-utilities\") pod \"certified-operators-4pkt7\" (UID: \"f4cdffae-731e-41f7-96d8-b015a5576a18\") " pod="openshift-marketplace/certified-operators-4pkt7" Feb 27 09:07:13 crc kubenswrapper[4612]: I0227 09:07:13.561518 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f4cdffae-731e-41f7-96d8-b015a5576a18-catalog-content\") pod \"certified-operators-4pkt7\" (UID: \"f4cdffae-731e-41f7-96d8-b015a5576a18\") " pod="openshift-marketplace/certified-operators-4pkt7" Feb 27 09:07:13 crc kubenswrapper[4612]: I0227 09:07:13.588653 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wc2gd\" (UniqueName: \"kubernetes.io/projected/f4cdffae-731e-41f7-96d8-b015a5576a18-kube-api-access-wc2gd\") pod \"certified-operators-4pkt7\" (UID: \"f4cdffae-731e-41f7-96d8-b015a5576a18\") " pod="openshift-marketplace/certified-operators-4pkt7" Feb 27 09:07:13 crc kubenswrapper[4612]: I0227 09:07:13.732422 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4pkt7" Feb 27 09:07:14 crc kubenswrapper[4612]: I0227 09:07:14.287741 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-4pkt7"] Feb 27 09:07:15 crc kubenswrapper[4612]: I0227 09:07:15.278006 4612 generic.go:334] "Generic (PLEG): container finished" podID="f4cdffae-731e-41f7-96d8-b015a5576a18" containerID="c241e7636871d1005fafdd06e27ff08de0c4af1b9e1e3a1458b4a5eba6a71efe" exitCode=0 Feb 27 09:07:15 crc kubenswrapper[4612]: I0227 09:07:15.278102 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4pkt7" event={"ID":"f4cdffae-731e-41f7-96d8-b015a5576a18","Type":"ContainerDied","Data":"c241e7636871d1005fafdd06e27ff08de0c4af1b9e1e3a1458b4a5eba6a71efe"} Feb 27 09:07:15 crc kubenswrapper[4612]: I0227 09:07:15.278462 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4pkt7" event={"ID":"f4cdffae-731e-41f7-96d8-b015a5576a18","Type":"ContainerStarted","Data":"41c06c0d633bdb0dbe4f0a4819bded50b4490c3fc667ab1f9d7cbc1614e8278b"} Feb 27 09:07:17 crc kubenswrapper[4612]: I0227 09:07:17.303901 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4pkt7" event={"ID":"f4cdffae-731e-41f7-96d8-b015a5576a18","Type":"ContainerStarted","Data":"ae33edc49f11f560fbc693980125c1daba41b91ca98395cd733eea20d1bb3c2d"} Feb 27 09:07:22 crc kubenswrapper[4612]: I0227 09:07:22.354399 4612 generic.go:334] "Generic (PLEG): container finished" podID="f4cdffae-731e-41f7-96d8-b015a5576a18" containerID="ae33edc49f11f560fbc693980125c1daba41b91ca98395cd733eea20d1bb3c2d" exitCode=0 Feb 27 09:07:22 crc kubenswrapper[4612]: I0227 09:07:22.354468 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4pkt7" event={"ID":"f4cdffae-731e-41f7-96d8-b015a5576a18","Type":"ContainerDied","Data":"ae33edc49f11f560fbc693980125c1daba41b91ca98395cd733eea20d1bb3c2d"} Feb 27 09:07:25 crc kubenswrapper[4612]: I0227 09:07:25.389037 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4pkt7" event={"ID":"f4cdffae-731e-41f7-96d8-b015a5576a18","Type":"ContainerStarted","Data":"5275e9480ee7e516b8cbd8c34fb3826a6828480dd93ee62d42beec4042fd8271"} Feb 27 09:07:25 crc kubenswrapper[4612]: I0227 09:07:25.410307 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-4pkt7" podStartSLOduration=3.768711859 podStartE2EDuration="12.410288526s" podCreationTimestamp="2026-02-27 09:07:13 +0000 UTC" firstStartedPulling="2026-02-27 09:07:15.282000568 +0000 UTC m=+4693.135930606" lastFinishedPulling="2026-02-27 09:07:23.923577265 +0000 UTC m=+4701.777507273" observedRunningTime="2026-02-27 09:07:25.408415832 +0000 UTC m=+4703.262345850" watchObservedRunningTime="2026-02-27 09:07:25.410288526 +0000 UTC m=+4703.264218534" Feb 27 09:07:33 crc kubenswrapper[4612]: I0227 09:07:33.733058 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-4pkt7" Feb 27 09:07:33 crc kubenswrapper[4612]: I0227 09:07:33.733684 4612 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-4pkt7" Feb 27 09:07:33 crc kubenswrapper[4612]: I0227 09:07:33.954782 4612 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-4pkt7" Feb 27 09:07:34 crc kubenswrapper[4612]: I0227 09:07:34.518893 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-4pkt7" Feb 27 09:07:34 crc kubenswrapper[4612]: I0227 09:07:34.569119 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-4pkt7"] Feb 27 09:07:36 crc kubenswrapper[4612]: I0227 09:07:36.494157 4612 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-4pkt7" podUID="f4cdffae-731e-41f7-96d8-b015a5576a18" containerName="registry-server" containerID="cri-o://5275e9480ee7e516b8cbd8c34fb3826a6828480dd93ee62d42beec4042fd8271" gracePeriod=2 Feb 27 09:07:36 crc kubenswrapper[4612]: I0227 09:07:36.946439 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4pkt7" Feb 27 09:07:37 crc kubenswrapper[4612]: I0227 09:07:37.084413 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f4cdffae-731e-41f7-96d8-b015a5576a18-utilities\") pod \"f4cdffae-731e-41f7-96d8-b015a5576a18\" (UID: \"f4cdffae-731e-41f7-96d8-b015a5576a18\") " Feb 27 09:07:37 crc kubenswrapper[4612]: I0227 09:07:37.084884 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wc2gd\" (UniqueName: \"kubernetes.io/projected/f4cdffae-731e-41f7-96d8-b015a5576a18-kube-api-access-wc2gd\") pod \"f4cdffae-731e-41f7-96d8-b015a5576a18\" (UID: \"f4cdffae-731e-41f7-96d8-b015a5576a18\") " Feb 27 09:07:37 crc kubenswrapper[4612]: I0227 09:07:37.084918 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f4cdffae-731e-41f7-96d8-b015a5576a18-catalog-content\") pod \"f4cdffae-731e-41f7-96d8-b015a5576a18\" (UID: \"f4cdffae-731e-41f7-96d8-b015a5576a18\") " Feb 27 09:07:37 crc kubenswrapper[4612]: I0227 09:07:37.085492 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f4cdffae-731e-41f7-96d8-b015a5576a18-utilities" (OuterVolumeSpecName: "utilities") pod "f4cdffae-731e-41f7-96d8-b015a5576a18" (UID: "f4cdffae-731e-41f7-96d8-b015a5576a18"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 09:07:37 crc kubenswrapper[4612]: I0227 09:07:37.102284 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f4cdffae-731e-41f7-96d8-b015a5576a18-kube-api-access-wc2gd" (OuterVolumeSpecName: "kube-api-access-wc2gd") pod "f4cdffae-731e-41f7-96d8-b015a5576a18" (UID: "f4cdffae-731e-41f7-96d8-b015a5576a18"). InnerVolumeSpecName "kube-api-access-wc2gd". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 09:07:37 crc kubenswrapper[4612]: I0227 09:07:37.156382 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f4cdffae-731e-41f7-96d8-b015a5576a18-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f4cdffae-731e-41f7-96d8-b015a5576a18" (UID: "f4cdffae-731e-41f7-96d8-b015a5576a18"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 09:07:37 crc kubenswrapper[4612]: I0227 09:07:37.186593 4612 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f4cdffae-731e-41f7-96d8-b015a5576a18-utilities\") on node \"crc\" DevicePath \"\"" Feb 27 09:07:37 crc kubenswrapper[4612]: I0227 09:07:37.186625 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wc2gd\" (UniqueName: \"kubernetes.io/projected/f4cdffae-731e-41f7-96d8-b015a5576a18-kube-api-access-wc2gd\") on node \"crc\" DevicePath \"\"" Feb 27 09:07:37 crc kubenswrapper[4612]: I0227 09:07:37.186639 4612 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f4cdffae-731e-41f7-96d8-b015a5576a18-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 27 09:07:37 crc kubenswrapper[4612]: I0227 09:07:37.510654 4612 generic.go:334] "Generic (PLEG): container finished" podID="f4cdffae-731e-41f7-96d8-b015a5576a18" containerID="5275e9480ee7e516b8cbd8c34fb3826a6828480dd93ee62d42beec4042fd8271" exitCode=0 Feb 27 09:07:37 crc kubenswrapper[4612]: I0227 09:07:37.510878 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4pkt7" event={"ID":"f4cdffae-731e-41f7-96d8-b015a5576a18","Type":"ContainerDied","Data":"5275e9480ee7e516b8cbd8c34fb3826a6828480dd93ee62d42beec4042fd8271"} Feb 27 09:07:37 crc kubenswrapper[4612]: I0227 09:07:37.511909 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4pkt7" event={"ID":"f4cdffae-731e-41f7-96d8-b015a5576a18","Type":"ContainerDied","Data":"41c06c0d633bdb0dbe4f0a4819bded50b4490c3fc667ab1f9d7cbc1614e8278b"} Feb 27 09:07:37 crc kubenswrapper[4612]: I0227 09:07:37.511953 4612 scope.go:117] "RemoveContainer" containerID="5275e9480ee7e516b8cbd8c34fb3826a6828480dd93ee62d42beec4042fd8271" Feb 27 09:07:37 crc kubenswrapper[4612]: I0227 09:07:37.510992 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4pkt7" Feb 27 09:07:37 crc kubenswrapper[4612]: I0227 09:07:37.529649 4612 scope.go:117] "RemoveContainer" containerID="ae33edc49f11f560fbc693980125c1daba41b91ca98395cd733eea20d1bb3c2d" Feb 27 09:07:37 crc kubenswrapper[4612]: I0227 09:07:37.560493 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-4pkt7"] Feb 27 09:07:37 crc kubenswrapper[4612]: I0227 09:07:37.571090 4612 scope.go:117] "RemoveContainer" containerID="c241e7636871d1005fafdd06e27ff08de0c4af1b9e1e3a1458b4a5eba6a71efe" Feb 27 09:07:37 crc kubenswrapper[4612]: I0227 09:07:37.575771 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-4pkt7"] Feb 27 09:07:37 crc kubenswrapper[4612]: I0227 09:07:37.608899 4612 scope.go:117] "RemoveContainer" containerID="5275e9480ee7e516b8cbd8c34fb3826a6828480dd93ee62d42beec4042fd8271" Feb 27 09:07:37 crc kubenswrapper[4612]: E0227 09:07:37.609353 4612 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5275e9480ee7e516b8cbd8c34fb3826a6828480dd93ee62d42beec4042fd8271\": container with ID starting with 5275e9480ee7e516b8cbd8c34fb3826a6828480dd93ee62d42beec4042fd8271 not found: ID does not exist" containerID="5275e9480ee7e516b8cbd8c34fb3826a6828480dd93ee62d42beec4042fd8271" Feb 27 09:07:37 crc kubenswrapper[4612]: I0227 09:07:37.609452 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5275e9480ee7e516b8cbd8c34fb3826a6828480dd93ee62d42beec4042fd8271"} err="failed to get container status \"5275e9480ee7e516b8cbd8c34fb3826a6828480dd93ee62d42beec4042fd8271\": rpc error: code = NotFound desc = could not find container \"5275e9480ee7e516b8cbd8c34fb3826a6828480dd93ee62d42beec4042fd8271\": container with ID starting with 5275e9480ee7e516b8cbd8c34fb3826a6828480dd93ee62d42beec4042fd8271 not found: ID does not exist" Feb 27 09:07:37 crc kubenswrapper[4612]: I0227 09:07:37.609480 4612 scope.go:117] "RemoveContainer" containerID="ae33edc49f11f560fbc693980125c1daba41b91ca98395cd733eea20d1bb3c2d" Feb 27 09:07:37 crc kubenswrapper[4612]: E0227 09:07:37.609978 4612 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ae33edc49f11f560fbc693980125c1daba41b91ca98395cd733eea20d1bb3c2d\": container with ID starting with ae33edc49f11f560fbc693980125c1daba41b91ca98395cd733eea20d1bb3c2d not found: ID does not exist" containerID="ae33edc49f11f560fbc693980125c1daba41b91ca98395cd733eea20d1bb3c2d" Feb 27 09:07:37 crc kubenswrapper[4612]: I0227 09:07:37.610004 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ae33edc49f11f560fbc693980125c1daba41b91ca98395cd733eea20d1bb3c2d"} err="failed to get container status \"ae33edc49f11f560fbc693980125c1daba41b91ca98395cd733eea20d1bb3c2d\": rpc error: code = NotFound desc = could not find container \"ae33edc49f11f560fbc693980125c1daba41b91ca98395cd733eea20d1bb3c2d\": container with ID starting with ae33edc49f11f560fbc693980125c1daba41b91ca98395cd733eea20d1bb3c2d not found: ID does not exist" Feb 27 09:07:37 crc kubenswrapper[4612]: I0227 09:07:37.610019 4612 scope.go:117] "RemoveContainer" containerID="c241e7636871d1005fafdd06e27ff08de0c4af1b9e1e3a1458b4a5eba6a71efe" Feb 27 09:07:37 crc kubenswrapper[4612]: E0227 09:07:37.610324 4612 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c241e7636871d1005fafdd06e27ff08de0c4af1b9e1e3a1458b4a5eba6a71efe\": container with ID starting with c241e7636871d1005fafdd06e27ff08de0c4af1b9e1e3a1458b4a5eba6a71efe not found: ID does not exist" containerID="c241e7636871d1005fafdd06e27ff08de0c4af1b9e1e3a1458b4a5eba6a71efe" Feb 27 09:07:37 crc kubenswrapper[4612]: I0227 09:07:37.610539 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c241e7636871d1005fafdd06e27ff08de0c4af1b9e1e3a1458b4a5eba6a71efe"} err="failed to get container status \"c241e7636871d1005fafdd06e27ff08de0c4af1b9e1e3a1458b4a5eba6a71efe\": rpc error: code = NotFound desc = could not find container \"c241e7636871d1005fafdd06e27ff08de0c4af1b9e1e3a1458b4a5eba6a71efe\": container with ID starting with c241e7636871d1005fafdd06e27ff08de0c4af1b9e1e3a1458b4a5eba6a71efe not found: ID does not exist" Feb 27 09:07:38 crc kubenswrapper[4612]: I0227 09:07:38.864733 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4cdffae-731e-41f7-96d8-b015a5576a18" path="/var/lib/kubelet/pods/f4cdffae-731e-41f7-96d8-b015a5576a18/volumes" Feb 27 09:08:00 crc kubenswrapper[4612]: I0227 09:08:00.155429 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29536388-7rftd"] Feb 27 09:08:00 crc kubenswrapper[4612]: E0227 09:08:00.156916 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4cdffae-731e-41f7-96d8-b015a5576a18" containerName="extract-content" Feb 27 09:08:00 crc kubenswrapper[4612]: I0227 09:08:00.156949 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4cdffae-731e-41f7-96d8-b015a5576a18" containerName="extract-content" Feb 27 09:08:00 crc kubenswrapper[4612]: E0227 09:08:00.157011 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4cdffae-731e-41f7-96d8-b015a5576a18" containerName="registry-server" Feb 27 09:08:00 crc kubenswrapper[4612]: I0227 09:08:00.157024 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4cdffae-731e-41f7-96d8-b015a5576a18" containerName="registry-server" Feb 27 09:08:00 crc kubenswrapper[4612]: E0227 09:08:00.157056 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4cdffae-731e-41f7-96d8-b015a5576a18" containerName="extract-utilities" Feb 27 09:08:00 crc kubenswrapper[4612]: I0227 09:08:00.157070 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4cdffae-731e-41f7-96d8-b015a5576a18" containerName="extract-utilities" Feb 27 09:08:00 crc kubenswrapper[4612]: I0227 09:08:00.157442 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4cdffae-731e-41f7-96d8-b015a5576a18" containerName="registry-server" Feb 27 09:08:00 crc kubenswrapper[4612]: I0227 09:08:00.158490 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536388-7rftd" Feb 27 09:08:00 crc kubenswrapper[4612]: I0227 09:08:00.161072 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 27 09:08:00 crc kubenswrapper[4612]: I0227 09:08:00.162263 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-5zwfx" Feb 27 09:08:00 crc kubenswrapper[4612]: I0227 09:08:00.162782 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 27 09:08:00 crc kubenswrapper[4612]: I0227 09:08:00.168709 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536388-7rftd"] Feb 27 09:08:00 crc kubenswrapper[4612]: I0227 09:08:00.208226 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cgl24\" (UniqueName: \"kubernetes.io/projected/ca8320dc-a538-4330-936e-74e7fe62dd30-kube-api-access-cgl24\") pod \"auto-csr-approver-29536388-7rftd\" (UID: \"ca8320dc-a538-4330-936e-74e7fe62dd30\") " pod="openshift-infra/auto-csr-approver-29536388-7rftd" Feb 27 09:08:00 crc kubenswrapper[4612]: I0227 09:08:00.310003 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cgl24\" (UniqueName: \"kubernetes.io/projected/ca8320dc-a538-4330-936e-74e7fe62dd30-kube-api-access-cgl24\") pod \"auto-csr-approver-29536388-7rftd\" (UID: \"ca8320dc-a538-4330-936e-74e7fe62dd30\") " pod="openshift-infra/auto-csr-approver-29536388-7rftd" Feb 27 09:08:00 crc kubenswrapper[4612]: I0227 09:08:00.336998 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cgl24\" (UniqueName: \"kubernetes.io/projected/ca8320dc-a538-4330-936e-74e7fe62dd30-kube-api-access-cgl24\") pod \"auto-csr-approver-29536388-7rftd\" (UID: \"ca8320dc-a538-4330-936e-74e7fe62dd30\") " pod="openshift-infra/auto-csr-approver-29536388-7rftd" Feb 27 09:08:00 crc kubenswrapper[4612]: I0227 09:08:00.485925 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536388-7rftd" Feb 27 09:08:00 crc kubenswrapper[4612]: I0227 09:08:00.794815 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536388-7rftd"] Feb 27 09:08:01 crc kubenswrapper[4612]: I0227 09:08:01.787082 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536388-7rftd" event={"ID":"ca8320dc-a538-4330-936e-74e7fe62dd30","Type":"ContainerStarted","Data":"cad7364e804b4afd0367b36b6f389185a3a0144dd0aa1bbb4db645da0a0daddf"} Feb 27 09:08:02 crc kubenswrapper[4612]: I0227 09:08:02.803165 4612 generic.go:334] "Generic (PLEG): container finished" podID="ca8320dc-a538-4330-936e-74e7fe62dd30" containerID="7615e9a05300b174510034ccbceefb060369b203b78547262124d37e85cc2a88" exitCode=0 Feb 27 09:08:02 crc kubenswrapper[4612]: I0227 09:08:02.803250 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536388-7rftd" event={"ID":"ca8320dc-a538-4330-936e-74e7fe62dd30","Type":"ContainerDied","Data":"7615e9a05300b174510034ccbceefb060369b203b78547262124d37e85cc2a88"} Feb 27 09:08:04 crc kubenswrapper[4612]: I0227 09:08:04.237726 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536388-7rftd" Feb 27 09:08:04 crc kubenswrapper[4612]: I0227 09:08:04.292661 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cgl24\" (UniqueName: \"kubernetes.io/projected/ca8320dc-a538-4330-936e-74e7fe62dd30-kube-api-access-cgl24\") pod \"ca8320dc-a538-4330-936e-74e7fe62dd30\" (UID: \"ca8320dc-a538-4330-936e-74e7fe62dd30\") " Feb 27 09:08:04 crc kubenswrapper[4612]: I0227 09:08:04.307132 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ca8320dc-a538-4330-936e-74e7fe62dd30-kube-api-access-cgl24" (OuterVolumeSpecName: "kube-api-access-cgl24") pod "ca8320dc-a538-4330-936e-74e7fe62dd30" (UID: "ca8320dc-a538-4330-936e-74e7fe62dd30"). InnerVolumeSpecName "kube-api-access-cgl24". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 09:08:04 crc kubenswrapper[4612]: I0227 09:08:04.396056 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cgl24\" (UniqueName: \"kubernetes.io/projected/ca8320dc-a538-4330-936e-74e7fe62dd30-kube-api-access-cgl24\") on node \"crc\" DevicePath \"\"" Feb 27 09:08:04 crc kubenswrapper[4612]: I0227 09:08:04.841546 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536388-7rftd" event={"ID":"ca8320dc-a538-4330-936e-74e7fe62dd30","Type":"ContainerDied","Data":"cad7364e804b4afd0367b36b6f389185a3a0144dd0aa1bbb4db645da0a0daddf"} Feb 27 09:08:04 crc kubenswrapper[4612]: I0227 09:08:04.841945 4612 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cad7364e804b4afd0367b36b6f389185a3a0144dd0aa1bbb4db645da0a0daddf" Feb 27 09:08:04 crc kubenswrapper[4612]: I0227 09:08:04.841592 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536388-7rftd" Feb 27 09:08:05 crc kubenswrapper[4612]: I0227 09:08:05.313760 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29536382-2jd9v"] Feb 27 09:08:05 crc kubenswrapper[4612]: I0227 09:08:05.320743 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29536382-2jd9v"] Feb 27 09:08:06 crc kubenswrapper[4612]: I0227 09:08:06.874810 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="80a082d2-95dd-4d63-965f-8d79c871ad0e" path="/var/lib/kubelet/pods/80a082d2-95dd-4d63-965f-8d79c871ad0e/volumes" Feb 27 09:08:19 crc kubenswrapper[4612]: I0227 09:08:19.069559 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-gxfb5/must-gather-ksdms"] Feb 27 09:08:19 crc kubenswrapper[4612]: E0227 09:08:19.071546 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ca8320dc-a538-4330-936e-74e7fe62dd30" containerName="oc" Feb 27 09:08:19 crc kubenswrapper[4612]: I0227 09:08:19.071676 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="ca8320dc-a538-4330-936e-74e7fe62dd30" containerName="oc" Feb 27 09:08:19 crc kubenswrapper[4612]: I0227 09:08:19.072099 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="ca8320dc-a538-4330-936e-74e7fe62dd30" containerName="oc" Feb 27 09:08:19 crc kubenswrapper[4612]: I0227 09:08:19.073467 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-gxfb5/must-gather-ksdms" Feb 27 09:08:19 crc kubenswrapper[4612]: I0227 09:08:19.076580 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-gxfb5"/"openshift-service-ca.crt" Feb 27 09:08:19 crc kubenswrapper[4612]: I0227 09:08:19.098977 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-gxfb5"/"kube-root-ca.crt" Feb 27 09:08:19 crc kubenswrapper[4612]: I0227 09:08:19.192204 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-gxfb5/must-gather-ksdms"] Feb 27 09:08:19 crc kubenswrapper[4612]: I0227 09:08:19.201830 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/445b752f-7d64-4300-a5fe-bebc7d94f962-must-gather-output\") pod \"must-gather-ksdms\" (UID: \"445b752f-7d64-4300-a5fe-bebc7d94f962\") " pod="openshift-must-gather-gxfb5/must-gather-ksdms" Feb 27 09:08:19 crc kubenswrapper[4612]: I0227 09:08:19.202010 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-twk7b\" (UniqueName: \"kubernetes.io/projected/445b752f-7d64-4300-a5fe-bebc7d94f962-kube-api-access-twk7b\") pod \"must-gather-ksdms\" (UID: \"445b752f-7d64-4300-a5fe-bebc7d94f962\") " pod="openshift-must-gather-gxfb5/must-gather-ksdms" Feb 27 09:08:19 crc kubenswrapper[4612]: I0227 09:08:19.303581 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-twk7b\" (UniqueName: \"kubernetes.io/projected/445b752f-7d64-4300-a5fe-bebc7d94f962-kube-api-access-twk7b\") pod \"must-gather-ksdms\" (UID: \"445b752f-7d64-4300-a5fe-bebc7d94f962\") " pod="openshift-must-gather-gxfb5/must-gather-ksdms" Feb 27 09:08:19 crc kubenswrapper[4612]: I0227 09:08:19.303656 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/445b752f-7d64-4300-a5fe-bebc7d94f962-must-gather-output\") pod \"must-gather-ksdms\" (UID: \"445b752f-7d64-4300-a5fe-bebc7d94f962\") " pod="openshift-must-gather-gxfb5/must-gather-ksdms" Feb 27 09:08:19 crc kubenswrapper[4612]: I0227 09:08:19.304295 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/445b752f-7d64-4300-a5fe-bebc7d94f962-must-gather-output\") pod \"must-gather-ksdms\" (UID: \"445b752f-7d64-4300-a5fe-bebc7d94f962\") " pod="openshift-must-gather-gxfb5/must-gather-ksdms" Feb 27 09:08:19 crc kubenswrapper[4612]: I0227 09:08:19.326568 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-twk7b\" (UniqueName: \"kubernetes.io/projected/445b752f-7d64-4300-a5fe-bebc7d94f962-kube-api-access-twk7b\") pod \"must-gather-ksdms\" (UID: \"445b752f-7d64-4300-a5fe-bebc7d94f962\") " pod="openshift-must-gather-gxfb5/must-gather-ksdms" Feb 27 09:08:19 crc kubenswrapper[4612]: I0227 09:08:19.417952 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-gxfb5/must-gather-ksdms" Feb 27 09:08:20 crc kubenswrapper[4612]: I0227 09:08:20.542433 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-gxfb5/must-gather-ksdms"] Feb 27 09:08:20 crc kubenswrapper[4612]: I0227 09:08:20.918262 4612 scope.go:117] "RemoveContainer" containerID="89ca1cead2b1d034bd87ad1cb0b15782f88020d83a8b54fbedef5cfbb6857046" Feb 27 09:08:21 crc kubenswrapper[4612]: I0227 09:08:21.050225 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-gxfb5/must-gather-ksdms" event={"ID":"445b752f-7d64-4300-a5fe-bebc7d94f962","Type":"ContainerStarted","Data":"5758c84ae5ba6363334c142523e1a977fdcc41a6086265bba40ba4e563d7b5f4"} Feb 27 09:08:21 crc kubenswrapper[4612]: I0227 09:08:21.050548 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-gxfb5/must-gather-ksdms" event={"ID":"445b752f-7d64-4300-a5fe-bebc7d94f962","Type":"ContainerStarted","Data":"b165a22365e30cd326f84767f0cb5e9a8124f50353ae9aa08fa2888bee319652"} Feb 27 09:08:22 crc kubenswrapper[4612]: I0227 09:08:22.062234 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-gxfb5/must-gather-ksdms" event={"ID":"445b752f-7d64-4300-a5fe-bebc7d94f962","Type":"ContainerStarted","Data":"6f641cde04e671db752583be64a094e25103923c86300f24a4e12540dec044b3"} Feb 27 09:08:22 crc kubenswrapper[4612]: I0227 09:08:22.087778 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-gxfb5/must-gather-ksdms" podStartSLOduration=3.087757745 podStartE2EDuration="3.087757745s" podCreationTimestamp="2026-02-27 09:08:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 09:08:22.082430853 +0000 UTC m=+4759.936360861" watchObservedRunningTime="2026-02-27 09:08:22.087757745 +0000 UTC m=+4759.941687743" Feb 27 09:08:24 crc kubenswrapper[4612]: I0227 09:08:24.656198 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-gxfb5/crc-debug-wgczc"] Feb 27 09:08:24 crc kubenswrapper[4612]: I0227 09:08:24.658554 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-gxfb5/crc-debug-wgczc" Feb 27 09:08:24 crc kubenswrapper[4612]: I0227 09:08:24.660506 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-gxfb5"/"default-dockercfg-szm2q" Feb 27 09:08:24 crc kubenswrapper[4612]: I0227 09:08:24.711580 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/86e7ebf2-4396-4985-b1b7-faf3b72eaf22-host\") pod \"crc-debug-wgczc\" (UID: \"86e7ebf2-4396-4985-b1b7-faf3b72eaf22\") " pod="openshift-must-gather-gxfb5/crc-debug-wgczc" Feb 27 09:08:24 crc kubenswrapper[4612]: I0227 09:08:24.711729 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zz7mr\" (UniqueName: \"kubernetes.io/projected/86e7ebf2-4396-4985-b1b7-faf3b72eaf22-kube-api-access-zz7mr\") pod \"crc-debug-wgczc\" (UID: \"86e7ebf2-4396-4985-b1b7-faf3b72eaf22\") " pod="openshift-must-gather-gxfb5/crc-debug-wgczc" Feb 27 09:08:24 crc kubenswrapper[4612]: I0227 09:08:24.816136 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zz7mr\" (UniqueName: \"kubernetes.io/projected/86e7ebf2-4396-4985-b1b7-faf3b72eaf22-kube-api-access-zz7mr\") pod \"crc-debug-wgczc\" (UID: \"86e7ebf2-4396-4985-b1b7-faf3b72eaf22\") " pod="openshift-must-gather-gxfb5/crc-debug-wgczc" Feb 27 09:08:24 crc kubenswrapper[4612]: I0227 09:08:24.816268 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/86e7ebf2-4396-4985-b1b7-faf3b72eaf22-host\") pod \"crc-debug-wgczc\" (UID: \"86e7ebf2-4396-4985-b1b7-faf3b72eaf22\") " pod="openshift-must-gather-gxfb5/crc-debug-wgczc" Feb 27 09:08:24 crc kubenswrapper[4612]: I0227 09:08:24.816387 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/86e7ebf2-4396-4985-b1b7-faf3b72eaf22-host\") pod \"crc-debug-wgczc\" (UID: \"86e7ebf2-4396-4985-b1b7-faf3b72eaf22\") " pod="openshift-must-gather-gxfb5/crc-debug-wgczc" Feb 27 09:08:24 crc kubenswrapper[4612]: I0227 09:08:24.843555 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zz7mr\" (UniqueName: \"kubernetes.io/projected/86e7ebf2-4396-4985-b1b7-faf3b72eaf22-kube-api-access-zz7mr\") pod \"crc-debug-wgczc\" (UID: \"86e7ebf2-4396-4985-b1b7-faf3b72eaf22\") " pod="openshift-must-gather-gxfb5/crc-debug-wgczc" Feb 27 09:08:24 crc kubenswrapper[4612]: I0227 09:08:24.975482 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-gxfb5/crc-debug-wgczc" Feb 27 09:08:25 crc kubenswrapper[4612]: I0227 09:08:25.094137 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-gxfb5/crc-debug-wgczc" event={"ID":"86e7ebf2-4396-4985-b1b7-faf3b72eaf22","Type":"ContainerStarted","Data":"cb03cba398d90b62a8408b9b19c1bf67c2d7d5801b798231b2b88b714e64d74c"} Feb 27 09:08:26 crc kubenswrapper[4612]: I0227 09:08:26.104741 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-gxfb5/crc-debug-wgczc" event={"ID":"86e7ebf2-4396-4985-b1b7-faf3b72eaf22","Type":"ContainerStarted","Data":"774af867d843dca9879978e4163a00adbc674a17692277e6f8632c4ce3c205c9"} Feb 27 09:08:26 crc kubenswrapper[4612]: I0227 09:08:26.130651 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-gxfb5/crc-debug-wgczc" podStartSLOduration=2.130628805 podStartE2EDuration="2.130628805s" podCreationTimestamp="2026-02-27 09:08:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-27 09:08:26.121419354 +0000 UTC m=+4763.975349362" watchObservedRunningTime="2026-02-27 09:08:26.130628805 +0000 UTC m=+4763.984558803" Feb 27 09:08:46 crc kubenswrapper[4612]: I0227 09:08:46.027131 4612 patch_prober.go:28] interesting pod/machine-config-daemon-924vb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 27 09:08:46 crc kubenswrapper[4612]: I0227 09:08:46.027749 4612 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 27 09:09:06 crc kubenswrapper[4612]: I0227 09:09:06.482083 4612 generic.go:334] "Generic (PLEG): container finished" podID="86e7ebf2-4396-4985-b1b7-faf3b72eaf22" containerID="774af867d843dca9879978e4163a00adbc674a17692277e6f8632c4ce3c205c9" exitCode=0 Feb 27 09:09:06 crc kubenswrapper[4612]: I0227 09:09:06.482132 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-gxfb5/crc-debug-wgczc" event={"ID":"86e7ebf2-4396-4985-b1b7-faf3b72eaf22","Type":"ContainerDied","Data":"774af867d843dca9879978e4163a00adbc674a17692277e6f8632c4ce3c205c9"} Feb 27 09:09:07 crc kubenswrapper[4612]: I0227 09:09:07.601293 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-gxfb5/crc-debug-wgczc" Feb 27 09:09:07 crc kubenswrapper[4612]: I0227 09:09:07.634311 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-gxfb5/crc-debug-wgczc"] Feb 27 09:09:07 crc kubenswrapper[4612]: I0227 09:09:07.640972 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/86e7ebf2-4396-4985-b1b7-faf3b72eaf22-host\") pod \"86e7ebf2-4396-4985-b1b7-faf3b72eaf22\" (UID: \"86e7ebf2-4396-4985-b1b7-faf3b72eaf22\") " Feb 27 09:09:07 crc kubenswrapper[4612]: I0227 09:09:07.641073 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zz7mr\" (UniqueName: \"kubernetes.io/projected/86e7ebf2-4396-4985-b1b7-faf3b72eaf22-kube-api-access-zz7mr\") pod \"86e7ebf2-4396-4985-b1b7-faf3b72eaf22\" (UID: \"86e7ebf2-4396-4985-b1b7-faf3b72eaf22\") " Feb 27 09:09:07 crc kubenswrapper[4612]: I0227 09:09:07.642610 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/86e7ebf2-4396-4985-b1b7-faf3b72eaf22-host" (OuterVolumeSpecName: "host") pod "86e7ebf2-4396-4985-b1b7-faf3b72eaf22" (UID: "86e7ebf2-4396-4985-b1b7-faf3b72eaf22"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 27 09:09:07 crc kubenswrapper[4612]: I0227 09:09:07.642679 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-gxfb5/crc-debug-wgczc"] Feb 27 09:09:07 crc kubenswrapper[4612]: I0227 09:09:07.677392 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/86e7ebf2-4396-4985-b1b7-faf3b72eaf22-kube-api-access-zz7mr" (OuterVolumeSpecName: "kube-api-access-zz7mr") pod "86e7ebf2-4396-4985-b1b7-faf3b72eaf22" (UID: "86e7ebf2-4396-4985-b1b7-faf3b72eaf22"). InnerVolumeSpecName "kube-api-access-zz7mr". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 09:09:07 crc kubenswrapper[4612]: I0227 09:09:07.742364 4612 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/86e7ebf2-4396-4985-b1b7-faf3b72eaf22-host\") on node \"crc\" DevicePath \"\"" Feb 27 09:09:07 crc kubenswrapper[4612]: I0227 09:09:07.742394 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zz7mr\" (UniqueName: \"kubernetes.io/projected/86e7ebf2-4396-4985-b1b7-faf3b72eaf22-kube-api-access-zz7mr\") on node \"crc\" DevicePath \"\"" Feb 27 09:09:08 crc kubenswrapper[4612]: I0227 09:09:08.499292 4612 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cb03cba398d90b62a8408b9b19c1bf67c2d7d5801b798231b2b88b714e64d74c" Feb 27 09:09:08 crc kubenswrapper[4612]: I0227 09:09:08.499333 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-gxfb5/crc-debug-wgczc" Feb 27 09:09:08 crc kubenswrapper[4612]: I0227 09:09:08.861647 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="86e7ebf2-4396-4985-b1b7-faf3b72eaf22" path="/var/lib/kubelet/pods/86e7ebf2-4396-4985-b1b7-faf3b72eaf22/volumes" Feb 27 09:09:09 crc kubenswrapper[4612]: I0227 09:09:09.065720 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-gxfb5/crc-debug-hgstp"] Feb 27 09:09:09 crc kubenswrapper[4612]: E0227 09:09:09.066093 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="86e7ebf2-4396-4985-b1b7-faf3b72eaf22" containerName="container-00" Feb 27 09:09:09 crc kubenswrapper[4612]: I0227 09:09:09.066108 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="86e7ebf2-4396-4985-b1b7-faf3b72eaf22" containerName="container-00" Feb 27 09:09:09 crc kubenswrapper[4612]: I0227 09:09:09.066280 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="86e7ebf2-4396-4985-b1b7-faf3b72eaf22" containerName="container-00" Feb 27 09:09:09 crc kubenswrapper[4612]: I0227 09:09:09.066928 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-gxfb5/crc-debug-hgstp" Feb 27 09:09:09 crc kubenswrapper[4612]: I0227 09:09:09.069200 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-gxfb5"/"default-dockercfg-szm2q" Feb 27 09:09:09 crc kubenswrapper[4612]: I0227 09:09:09.181572 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hxb7k\" (UniqueName: \"kubernetes.io/projected/1347613e-f879-4dc8-b726-7f87100ebf04-kube-api-access-hxb7k\") pod \"crc-debug-hgstp\" (UID: \"1347613e-f879-4dc8-b726-7f87100ebf04\") " pod="openshift-must-gather-gxfb5/crc-debug-hgstp" Feb 27 09:09:09 crc kubenswrapper[4612]: I0227 09:09:09.182044 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/1347613e-f879-4dc8-b726-7f87100ebf04-host\") pod \"crc-debug-hgstp\" (UID: \"1347613e-f879-4dc8-b726-7f87100ebf04\") " pod="openshift-must-gather-gxfb5/crc-debug-hgstp" Feb 27 09:09:09 crc kubenswrapper[4612]: I0227 09:09:09.283382 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hxb7k\" (UniqueName: \"kubernetes.io/projected/1347613e-f879-4dc8-b726-7f87100ebf04-kube-api-access-hxb7k\") pod \"crc-debug-hgstp\" (UID: \"1347613e-f879-4dc8-b726-7f87100ebf04\") " pod="openshift-must-gather-gxfb5/crc-debug-hgstp" Feb 27 09:09:09 crc kubenswrapper[4612]: I0227 09:09:09.283467 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/1347613e-f879-4dc8-b726-7f87100ebf04-host\") pod \"crc-debug-hgstp\" (UID: \"1347613e-f879-4dc8-b726-7f87100ebf04\") " pod="openshift-must-gather-gxfb5/crc-debug-hgstp" Feb 27 09:09:09 crc kubenswrapper[4612]: I0227 09:09:09.283618 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/1347613e-f879-4dc8-b726-7f87100ebf04-host\") pod \"crc-debug-hgstp\" (UID: \"1347613e-f879-4dc8-b726-7f87100ebf04\") " pod="openshift-must-gather-gxfb5/crc-debug-hgstp" Feb 27 09:09:09 crc kubenswrapper[4612]: I0227 09:09:09.310078 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hxb7k\" (UniqueName: \"kubernetes.io/projected/1347613e-f879-4dc8-b726-7f87100ebf04-kube-api-access-hxb7k\") pod \"crc-debug-hgstp\" (UID: \"1347613e-f879-4dc8-b726-7f87100ebf04\") " pod="openshift-must-gather-gxfb5/crc-debug-hgstp" Feb 27 09:09:09 crc kubenswrapper[4612]: I0227 09:09:09.383878 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-gxfb5/crc-debug-hgstp" Feb 27 09:09:09 crc kubenswrapper[4612]: I0227 09:09:09.518265 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-gxfb5/crc-debug-hgstp" event={"ID":"1347613e-f879-4dc8-b726-7f87100ebf04","Type":"ContainerStarted","Data":"181ff5d5d5ae425fc3b8c2eed9d1dcfc9d9ea858e77df31a9f8b4ca6e23dafcd"} Feb 27 09:09:10 crc kubenswrapper[4612]: I0227 09:09:10.527607 4612 generic.go:334] "Generic (PLEG): container finished" podID="1347613e-f879-4dc8-b726-7f87100ebf04" containerID="0a303417d47b93f56fe9e76626b77d3e695b9328364bdf2dfdd9ec678b311d79" exitCode=0 Feb 27 09:09:10 crc kubenswrapper[4612]: I0227 09:09:10.527703 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-gxfb5/crc-debug-hgstp" event={"ID":"1347613e-f879-4dc8-b726-7f87100ebf04","Type":"ContainerDied","Data":"0a303417d47b93f56fe9e76626b77d3e695b9328364bdf2dfdd9ec678b311d79"} Feb 27 09:09:11 crc kubenswrapper[4612]: I0227 09:09:11.633412 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-gxfb5/crc-debug-hgstp" Feb 27 09:09:11 crc kubenswrapper[4612]: I0227 09:09:11.825501 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/1347613e-f879-4dc8-b726-7f87100ebf04-host\") pod \"1347613e-f879-4dc8-b726-7f87100ebf04\" (UID: \"1347613e-f879-4dc8-b726-7f87100ebf04\") " Feb 27 09:09:11 crc kubenswrapper[4612]: I0227 09:09:11.825555 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hxb7k\" (UniqueName: \"kubernetes.io/projected/1347613e-f879-4dc8-b726-7f87100ebf04-kube-api-access-hxb7k\") pod \"1347613e-f879-4dc8-b726-7f87100ebf04\" (UID: \"1347613e-f879-4dc8-b726-7f87100ebf04\") " Feb 27 09:09:11 crc kubenswrapper[4612]: I0227 09:09:11.825768 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1347613e-f879-4dc8-b726-7f87100ebf04-host" (OuterVolumeSpecName: "host") pod "1347613e-f879-4dc8-b726-7f87100ebf04" (UID: "1347613e-f879-4dc8-b726-7f87100ebf04"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 27 09:09:11 crc kubenswrapper[4612]: I0227 09:09:11.825994 4612 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/1347613e-f879-4dc8-b726-7f87100ebf04-host\") on node \"crc\" DevicePath \"\"" Feb 27 09:09:11 crc kubenswrapper[4612]: I0227 09:09:11.873957 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1347613e-f879-4dc8-b726-7f87100ebf04-kube-api-access-hxb7k" (OuterVolumeSpecName: "kube-api-access-hxb7k") pod "1347613e-f879-4dc8-b726-7f87100ebf04" (UID: "1347613e-f879-4dc8-b726-7f87100ebf04"). InnerVolumeSpecName "kube-api-access-hxb7k". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 09:09:11 crc kubenswrapper[4612]: I0227 09:09:11.927726 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hxb7k\" (UniqueName: \"kubernetes.io/projected/1347613e-f879-4dc8-b726-7f87100ebf04-kube-api-access-hxb7k\") on node \"crc\" DevicePath \"\"" Feb 27 09:09:12 crc kubenswrapper[4612]: I0227 09:09:12.266368 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-gxfb5/crc-debug-hgstp"] Feb 27 09:09:12 crc kubenswrapper[4612]: I0227 09:09:12.272410 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-gxfb5/crc-debug-hgstp"] Feb 27 09:09:12 crc kubenswrapper[4612]: I0227 09:09:12.545752 4612 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="181ff5d5d5ae425fc3b8c2eed9d1dcfc9d9ea858e77df31a9f8b4ca6e23dafcd" Feb 27 09:09:12 crc kubenswrapper[4612]: I0227 09:09:12.546077 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-gxfb5/crc-debug-hgstp" Feb 27 09:09:12 crc kubenswrapper[4612]: I0227 09:09:12.863656 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1347613e-f879-4dc8-b726-7f87100ebf04" path="/var/lib/kubelet/pods/1347613e-f879-4dc8-b726-7f87100ebf04/volumes" Feb 27 09:09:13 crc kubenswrapper[4612]: I0227 09:09:13.521993 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-gxfb5/crc-debug-szvn4"] Feb 27 09:09:13 crc kubenswrapper[4612]: E0227 09:09:13.522582 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1347613e-f879-4dc8-b726-7f87100ebf04" containerName="container-00" Feb 27 09:09:13 crc kubenswrapper[4612]: I0227 09:09:13.522597 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="1347613e-f879-4dc8-b726-7f87100ebf04" containerName="container-00" Feb 27 09:09:13 crc kubenswrapper[4612]: I0227 09:09:13.522818 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="1347613e-f879-4dc8-b726-7f87100ebf04" containerName="container-00" Feb 27 09:09:13 crc kubenswrapper[4612]: I0227 09:09:13.523356 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-gxfb5/crc-debug-szvn4" Feb 27 09:09:13 crc kubenswrapper[4612]: I0227 09:09:13.530914 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-gxfb5"/"default-dockercfg-szm2q" Feb 27 09:09:13 crc kubenswrapper[4612]: I0227 09:09:13.553860 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/540cc9bf-009f-4586-ad66-c742b9b39ace-host\") pod \"crc-debug-szvn4\" (UID: \"540cc9bf-009f-4586-ad66-c742b9b39ace\") " pod="openshift-must-gather-gxfb5/crc-debug-szvn4" Feb 27 09:09:13 crc kubenswrapper[4612]: I0227 09:09:13.553901 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v7h5x\" (UniqueName: \"kubernetes.io/projected/540cc9bf-009f-4586-ad66-c742b9b39ace-kube-api-access-v7h5x\") pod \"crc-debug-szvn4\" (UID: \"540cc9bf-009f-4586-ad66-c742b9b39ace\") " pod="openshift-must-gather-gxfb5/crc-debug-szvn4" Feb 27 09:09:13 crc kubenswrapper[4612]: I0227 09:09:13.654802 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/540cc9bf-009f-4586-ad66-c742b9b39ace-host\") pod \"crc-debug-szvn4\" (UID: \"540cc9bf-009f-4586-ad66-c742b9b39ace\") " pod="openshift-must-gather-gxfb5/crc-debug-szvn4" Feb 27 09:09:13 crc kubenswrapper[4612]: I0227 09:09:13.654849 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v7h5x\" (UniqueName: \"kubernetes.io/projected/540cc9bf-009f-4586-ad66-c742b9b39ace-kube-api-access-v7h5x\") pod \"crc-debug-szvn4\" (UID: \"540cc9bf-009f-4586-ad66-c742b9b39ace\") " pod="openshift-must-gather-gxfb5/crc-debug-szvn4" Feb 27 09:09:13 crc kubenswrapper[4612]: I0227 09:09:13.654904 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/540cc9bf-009f-4586-ad66-c742b9b39ace-host\") pod \"crc-debug-szvn4\" (UID: \"540cc9bf-009f-4586-ad66-c742b9b39ace\") " pod="openshift-must-gather-gxfb5/crc-debug-szvn4" Feb 27 09:09:13 crc kubenswrapper[4612]: I0227 09:09:13.679669 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v7h5x\" (UniqueName: \"kubernetes.io/projected/540cc9bf-009f-4586-ad66-c742b9b39ace-kube-api-access-v7h5x\") pod \"crc-debug-szvn4\" (UID: \"540cc9bf-009f-4586-ad66-c742b9b39ace\") " pod="openshift-must-gather-gxfb5/crc-debug-szvn4" Feb 27 09:09:13 crc kubenswrapper[4612]: I0227 09:09:13.838805 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-gxfb5/crc-debug-szvn4" Feb 27 09:09:13 crc kubenswrapper[4612]: W0227 09:09:13.885913 4612 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod540cc9bf_009f_4586_ad66_c742b9b39ace.slice/crio-816eb4c5442cb867aaf4acc2af549d174e7cd416062e79049835d2602d4575e4 WatchSource:0}: Error finding container 816eb4c5442cb867aaf4acc2af549d174e7cd416062e79049835d2602d4575e4: Status 404 returned error can't find the container with id 816eb4c5442cb867aaf4acc2af549d174e7cd416062e79049835d2602d4575e4 Feb 27 09:09:14 crc kubenswrapper[4612]: I0227 09:09:14.561595 4612 generic.go:334] "Generic (PLEG): container finished" podID="540cc9bf-009f-4586-ad66-c742b9b39ace" containerID="7d0a063387a46052c838d3ee53023c025db059279a9c86348297c86a7891a7d5" exitCode=0 Feb 27 09:09:14 crc kubenswrapper[4612]: I0227 09:09:14.561758 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-gxfb5/crc-debug-szvn4" event={"ID":"540cc9bf-009f-4586-ad66-c742b9b39ace","Type":"ContainerDied","Data":"7d0a063387a46052c838d3ee53023c025db059279a9c86348297c86a7891a7d5"} Feb 27 09:09:14 crc kubenswrapper[4612]: I0227 09:09:14.561956 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-gxfb5/crc-debug-szvn4" event={"ID":"540cc9bf-009f-4586-ad66-c742b9b39ace","Type":"ContainerStarted","Data":"816eb4c5442cb867aaf4acc2af549d174e7cd416062e79049835d2602d4575e4"} Feb 27 09:09:14 crc kubenswrapper[4612]: I0227 09:09:14.594511 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-gxfb5/crc-debug-szvn4"] Feb 27 09:09:14 crc kubenswrapper[4612]: I0227 09:09:14.601993 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-gxfb5/crc-debug-szvn4"] Feb 27 09:09:15 crc kubenswrapper[4612]: I0227 09:09:15.657966 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-gxfb5/crc-debug-szvn4" Feb 27 09:09:15 crc kubenswrapper[4612]: I0227 09:09:15.790300 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v7h5x\" (UniqueName: \"kubernetes.io/projected/540cc9bf-009f-4586-ad66-c742b9b39ace-kube-api-access-v7h5x\") pod \"540cc9bf-009f-4586-ad66-c742b9b39ace\" (UID: \"540cc9bf-009f-4586-ad66-c742b9b39ace\") " Feb 27 09:09:15 crc kubenswrapper[4612]: I0227 09:09:15.790656 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/540cc9bf-009f-4586-ad66-c742b9b39ace-host\") pod \"540cc9bf-009f-4586-ad66-c742b9b39ace\" (UID: \"540cc9bf-009f-4586-ad66-c742b9b39ace\") " Feb 27 09:09:15 crc kubenswrapper[4612]: I0227 09:09:15.790756 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/540cc9bf-009f-4586-ad66-c742b9b39ace-host" (OuterVolumeSpecName: "host") pod "540cc9bf-009f-4586-ad66-c742b9b39ace" (UID: "540cc9bf-009f-4586-ad66-c742b9b39ace"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 27 09:09:15 crc kubenswrapper[4612]: I0227 09:09:15.791093 4612 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/540cc9bf-009f-4586-ad66-c742b9b39ace-host\") on node \"crc\" DevicePath \"\"" Feb 27 09:09:15 crc kubenswrapper[4612]: I0227 09:09:15.798483 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/540cc9bf-009f-4586-ad66-c742b9b39ace-kube-api-access-v7h5x" (OuterVolumeSpecName: "kube-api-access-v7h5x") pod "540cc9bf-009f-4586-ad66-c742b9b39ace" (UID: "540cc9bf-009f-4586-ad66-c742b9b39ace"). InnerVolumeSpecName "kube-api-access-v7h5x". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 09:09:15 crc kubenswrapper[4612]: I0227 09:09:15.892951 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v7h5x\" (UniqueName: \"kubernetes.io/projected/540cc9bf-009f-4586-ad66-c742b9b39ace-kube-api-access-v7h5x\") on node \"crc\" DevicePath \"\"" Feb 27 09:09:16 crc kubenswrapper[4612]: I0227 09:09:16.027678 4612 patch_prober.go:28] interesting pod/machine-config-daemon-924vb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 27 09:09:16 crc kubenswrapper[4612]: I0227 09:09:16.027766 4612 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 27 09:09:16 crc kubenswrapper[4612]: I0227 09:09:16.598554 4612 scope.go:117] "RemoveContainer" containerID="7d0a063387a46052c838d3ee53023c025db059279a9c86348297c86a7891a7d5" Feb 27 09:09:16 crc kubenswrapper[4612]: I0227 09:09:16.598597 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-gxfb5/crc-debug-szvn4" Feb 27 09:09:16 crc kubenswrapper[4612]: I0227 09:09:16.863801 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="540cc9bf-009f-4586-ad66-c742b9b39ace" path="/var/lib/kubelet/pods/540cc9bf-009f-4586-ad66-c742b9b39ace/volumes" Feb 27 09:09:46 crc kubenswrapper[4612]: I0227 09:09:46.027542 4612 patch_prober.go:28] interesting pod/machine-config-daemon-924vb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 27 09:09:46 crc kubenswrapper[4612]: I0227 09:09:46.028029 4612 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 27 09:09:46 crc kubenswrapper[4612]: I0227 09:09:46.028073 4612 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-924vb" Feb 27 09:09:46 crc kubenswrapper[4612]: I0227 09:09:46.028816 4612 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"c90278d5d8864d43287d88f2fcf48602a1065bde9d8b6d625ab9498edb9b58db"} pod="openshift-machine-config-operator/machine-config-daemon-924vb" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 27 09:09:46 crc kubenswrapper[4612]: I0227 09:09:46.028873 4612 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" containerName="machine-config-daemon" containerID="cri-o://c90278d5d8864d43287d88f2fcf48602a1065bde9d8b6d625ab9498edb9b58db" gracePeriod=600 Feb 27 09:09:46 crc kubenswrapper[4612]: I0227 09:09:46.864636 4612 generic.go:334] "Generic (PLEG): container finished" podID="24599373-7adc-4a1b-8bb4-797bf726f43d" containerID="c90278d5d8864d43287d88f2fcf48602a1065bde9d8b6d625ab9498edb9b58db" exitCode=0 Feb 27 09:09:46 crc kubenswrapper[4612]: I0227 09:09:46.864723 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-924vb" event={"ID":"24599373-7adc-4a1b-8bb4-797bf726f43d","Type":"ContainerDied","Data":"c90278d5d8864d43287d88f2fcf48602a1065bde9d8b6d625ab9498edb9b58db"} Feb 27 09:09:46 crc kubenswrapper[4612]: I0227 09:09:46.865215 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-924vb" event={"ID":"24599373-7adc-4a1b-8bb4-797bf726f43d","Type":"ContainerStarted","Data":"83e903dc65dc37650f4243bfdb54ee1e09257b50cf32c05bb8e521882531bc38"} Feb 27 09:09:46 crc kubenswrapper[4612]: I0227 09:09:46.865239 4612 scope.go:117] "RemoveContainer" containerID="aa498fa4b5f3843852a1ebe0b165b06254525eeb49f5224c7bd2a0f8c2ae2480" Feb 27 09:09:51 crc kubenswrapper[4612]: I0227 09:09:51.989072 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-vxhvr"] Feb 27 09:09:51 crc kubenswrapper[4612]: E0227 09:09:51.990445 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="540cc9bf-009f-4586-ad66-c742b9b39ace" containerName="container-00" Feb 27 09:09:51 crc kubenswrapper[4612]: I0227 09:09:51.990466 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="540cc9bf-009f-4586-ad66-c742b9b39ace" containerName="container-00" Feb 27 09:09:51 crc kubenswrapper[4612]: I0227 09:09:51.995271 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="540cc9bf-009f-4586-ad66-c742b9b39ace" containerName="container-00" Feb 27 09:09:51 crc kubenswrapper[4612]: I0227 09:09:51.998350 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vxhvr" Feb 27 09:09:52 crc kubenswrapper[4612]: I0227 09:09:52.012409 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-vxhvr"] Feb 27 09:09:52 crc kubenswrapper[4612]: I0227 09:09:52.024612 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/247cf880-b83d-4493-b12f-e25afaa861fd-utilities\") pod \"redhat-operators-vxhvr\" (UID: \"247cf880-b83d-4493-b12f-e25afaa861fd\") " pod="openshift-marketplace/redhat-operators-vxhvr" Feb 27 09:09:52 crc kubenswrapper[4612]: I0227 09:09:52.024686 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6fxfj\" (UniqueName: \"kubernetes.io/projected/247cf880-b83d-4493-b12f-e25afaa861fd-kube-api-access-6fxfj\") pod \"redhat-operators-vxhvr\" (UID: \"247cf880-b83d-4493-b12f-e25afaa861fd\") " pod="openshift-marketplace/redhat-operators-vxhvr" Feb 27 09:09:52 crc kubenswrapper[4612]: I0227 09:09:52.024810 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/247cf880-b83d-4493-b12f-e25afaa861fd-catalog-content\") pod \"redhat-operators-vxhvr\" (UID: \"247cf880-b83d-4493-b12f-e25afaa861fd\") " pod="openshift-marketplace/redhat-operators-vxhvr" Feb 27 09:09:52 crc kubenswrapper[4612]: I0227 09:09:52.126124 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/247cf880-b83d-4493-b12f-e25afaa861fd-utilities\") pod \"redhat-operators-vxhvr\" (UID: \"247cf880-b83d-4493-b12f-e25afaa861fd\") " pod="openshift-marketplace/redhat-operators-vxhvr" Feb 27 09:09:52 crc kubenswrapper[4612]: I0227 09:09:52.126199 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6fxfj\" (UniqueName: \"kubernetes.io/projected/247cf880-b83d-4493-b12f-e25afaa861fd-kube-api-access-6fxfj\") pod \"redhat-operators-vxhvr\" (UID: \"247cf880-b83d-4493-b12f-e25afaa861fd\") " pod="openshift-marketplace/redhat-operators-vxhvr" Feb 27 09:09:52 crc kubenswrapper[4612]: I0227 09:09:52.126229 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/247cf880-b83d-4493-b12f-e25afaa861fd-catalog-content\") pod \"redhat-operators-vxhvr\" (UID: \"247cf880-b83d-4493-b12f-e25afaa861fd\") " pod="openshift-marketplace/redhat-operators-vxhvr" Feb 27 09:09:52 crc kubenswrapper[4612]: I0227 09:09:52.126611 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/247cf880-b83d-4493-b12f-e25afaa861fd-utilities\") pod \"redhat-operators-vxhvr\" (UID: \"247cf880-b83d-4493-b12f-e25afaa861fd\") " pod="openshift-marketplace/redhat-operators-vxhvr" Feb 27 09:09:52 crc kubenswrapper[4612]: I0227 09:09:52.126772 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/247cf880-b83d-4493-b12f-e25afaa861fd-catalog-content\") pod \"redhat-operators-vxhvr\" (UID: \"247cf880-b83d-4493-b12f-e25afaa861fd\") " pod="openshift-marketplace/redhat-operators-vxhvr" Feb 27 09:09:52 crc kubenswrapper[4612]: I0227 09:09:52.153058 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6fxfj\" (UniqueName: \"kubernetes.io/projected/247cf880-b83d-4493-b12f-e25afaa861fd-kube-api-access-6fxfj\") pod \"redhat-operators-vxhvr\" (UID: \"247cf880-b83d-4493-b12f-e25afaa861fd\") " pod="openshift-marketplace/redhat-operators-vxhvr" Feb 27 09:09:52 crc kubenswrapper[4612]: I0227 09:09:52.339893 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vxhvr" Feb 27 09:09:52 crc kubenswrapper[4612]: I0227 09:09:52.817571 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-vxhvr"] Feb 27 09:09:53 crc kubenswrapper[4612]: I0227 09:09:53.926629 4612 generic.go:334] "Generic (PLEG): container finished" podID="247cf880-b83d-4493-b12f-e25afaa861fd" containerID="137cd75031a64cc229591f7f3dacce26a9b19e760575d5be1dfec51a63436cb4" exitCode=0 Feb 27 09:09:53 crc kubenswrapper[4612]: I0227 09:09:53.926824 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vxhvr" event={"ID":"247cf880-b83d-4493-b12f-e25afaa861fd","Type":"ContainerDied","Data":"137cd75031a64cc229591f7f3dacce26a9b19e760575d5be1dfec51a63436cb4"} Feb 27 09:09:53 crc kubenswrapper[4612]: I0227 09:09:53.927011 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vxhvr" event={"ID":"247cf880-b83d-4493-b12f-e25afaa861fd","Type":"ContainerStarted","Data":"8f99a17ced44e454690efa852ec537d1c00e7e927e254f2da8e0a2ac9e3b6f9a"} Feb 27 09:09:55 crc kubenswrapper[4612]: I0227 09:09:55.945999 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vxhvr" event={"ID":"247cf880-b83d-4493-b12f-e25afaa861fd","Type":"ContainerStarted","Data":"83a2bbeaae26da1239f997503f6fa70146491baaab31b6cbaea96c7b5f953cb6"} Feb 27 09:10:00 crc kubenswrapper[4612]: I0227 09:10:00.141638 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29536390-lz95w"] Feb 27 09:10:00 crc kubenswrapper[4612]: I0227 09:10:00.143312 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536390-lz95w" Feb 27 09:10:00 crc kubenswrapper[4612]: I0227 09:10:00.145169 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 27 09:10:00 crc kubenswrapper[4612]: I0227 09:10:00.147059 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 27 09:10:00 crc kubenswrapper[4612]: I0227 09:10:00.147556 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-5zwfx" Feb 27 09:10:00 crc kubenswrapper[4612]: I0227 09:10:00.151382 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536390-lz95w"] Feb 27 09:10:00 crc kubenswrapper[4612]: I0227 09:10:00.192876 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bf2gp\" (UniqueName: \"kubernetes.io/projected/70bd3725-f383-45ff-a62f-45af7fd4f542-kube-api-access-bf2gp\") pod \"auto-csr-approver-29536390-lz95w\" (UID: \"70bd3725-f383-45ff-a62f-45af7fd4f542\") " pod="openshift-infra/auto-csr-approver-29536390-lz95w" Feb 27 09:10:00 crc kubenswrapper[4612]: I0227 09:10:00.294203 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bf2gp\" (UniqueName: \"kubernetes.io/projected/70bd3725-f383-45ff-a62f-45af7fd4f542-kube-api-access-bf2gp\") pod \"auto-csr-approver-29536390-lz95w\" (UID: \"70bd3725-f383-45ff-a62f-45af7fd4f542\") " pod="openshift-infra/auto-csr-approver-29536390-lz95w" Feb 27 09:10:00 crc kubenswrapper[4612]: I0227 09:10:00.335664 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bf2gp\" (UniqueName: \"kubernetes.io/projected/70bd3725-f383-45ff-a62f-45af7fd4f542-kube-api-access-bf2gp\") pod \"auto-csr-approver-29536390-lz95w\" (UID: \"70bd3725-f383-45ff-a62f-45af7fd4f542\") " pod="openshift-infra/auto-csr-approver-29536390-lz95w" Feb 27 09:10:00 crc kubenswrapper[4612]: I0227 09:10:00.497354 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536390-lz95w" Feb 27 09:10:00 crc kubenswrapper[4612]: I0227 09:10:00.970396 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536390-lz95w"] Feb 27 09:10:00 crc kubenswrapper[4612]: I0227 09:10:00.987785 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536390-lz95w" event={"ID":"70bd3725-f383-45ff-a62f-45af7fd4f542","Type":"ContainerStarted","Data":"2a5ecb2bf25b84b817180176e08b18b3970e9216a412a6d75c9edc56b121e962"} Feb 27 09:10:00 crc kubenswrapper[4612]: I0227 09:10:00.989758 4612 generic.go:334] "Generic (PLEG): container finished" podID="247cf880-b83d-4493-b12f-e25afaa861fd" containerID="83a2bbeaae26da1239f997503f6fa70146491baaab31b6cbaea96c7b5f953cb6" exitCode=0 Feb 27 09:10:00 crc kubenswrapper[4612]: I0227 09:10:00.989792 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vxhvr" event={"ID":"247cf880-b83d-4493-b12f-e25afaa861fd","Type":"ContainerDied","Data":"83a2bbeaae26da1239f997503f6fa70146491baaab31b6cbaea96c7b5f953cb6"} Feb 27 09:10:02 crc kubenswrapper[4612]: I0227 09:10:02.000356 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vxhvr" event={"ID":"247cf880-b83d-4493-b12f-e25afaa861fd","Type":"ContainerStarted","Data":"8dca1d714a17638cab306ee7f33ed3d80d59826130f47ba5fc6d2a63ab292164"} Feb 27 09:10:02 crc kubenswrapper[4612]: I0227 09:10:02.026610 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-vxhvr" podStartSLOduration=3.515953102 podStartE2EDuration="11.026594992s" podCreationTimestamp="2026-02-27 09:09:51 +0000 UTC" firstStartedPulling="2026-02-27 09:09:53.92817214 +0000 UTC m=+4851.782102138" lastFinishedPulling="2026-02-27 09:10:01.43881403 +0000 UTC m=+4859.292744028" observedRunningTime="2026-02-27 09:10:02.024296116 +0000 UTC m=+4859.878226114" watchObservedRunningTime="2026-02-27 09:10:02.026594992 +0000 UTC m=+4859.880524990" Feb 27 09:10:02 crc kubenswrapper[4612]: I0227 09:10:02.340564 4612 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-vxhvr" Feb 27 09:10:02 crc kubenswrapper[4612]: I0227 09:10:02.340669 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-vxhvr" Feb 27 09:10:03 crc kubenswrapper[4612]: I0227 09:10:03.008802 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536390-lz95w" event={"ID":"70bd3725-f383-45ff-a62f-45af7fd4f542","Type":"ContainerStarted","Data":"f6d0962857a99b47dbbd06edd001e3ef631ea6576c5f4450c0495eaa9d6ad379"} Feb 27 09:10:03 crc kubenswrapper[4612]: I0227 09:10:03.026986 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29536390-lz95w" podStartSLOduration=1.423365241 podStartE2EDuration="3.026966796s" podCreationTimestamp="2026-02-27 09:10:00 +0000 UTC" firstStartedPulling="2026-02-27 09:10:00.977415491 +0000 UTC m=+4858.831345489" lastFinishedPulling="2026-02-27 09:10:02.581017046 +0000 UTC m=+4860.434947044" observedRunningTime="2026-02-27 09:10:03.021912052 +0000 UTC m=+4860.875842040" watchObservedRunningTime="2026-02-27 09:10:03.026966796 +0000 UTC m=+4860.880896794" Feb 27 09:10:03 crc kubenswrapper[4612]: I0227 09:10:03.491146 4612 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-vxhvr" podUID="247cf880-b83d-4493-b12f-e25afaa861fd" containerName="registry-server" probeResult="failure" output=< Feb 27 09:10:03 crc kubenswrapper[4612]: timeout: failed to connect service ":50051" within 1s Feb 27 09:10:03 crc kubenswrapper[4612]: > Feb 27 09:10:04 crc kubenswrapper[4612]: I0227 09:10:04.043355 4612 generic.go:334] "Generic (PLEG): container finished" podID="70bd3725-f383-45ff-a62f-45af7fd4f542" containerID="f6d0962857a99b47dbbd06edd001e3ef631ea6576c5f4450c0495eaa9d6ad379" exitCode=0 Feb 27 09:10:04 crc kubenswrapper[4612]: I0227 09:10:04.044988 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536390-lz95w" event={"ID":"70bd3725-f383-45ff-a62f-45af7fd4f542","Type":"ContainerDied","Data":"f6d0962857a99b47dbbd06edd001e3ef631ea6576c5f4450c0495eaa9d6ad379"} Feb 27 09:10:05 crc kubenswrapper[4612]: I0227 09:10:05.363148 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536390-lz95w" Feb 27 09:10:05 crc kubenswrapper[4612]: I0227 09:10:05.498264 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2gp\" (UniqueName: \"kubernetes.io/projected/70bd3725-f383-45ff-a62f-45af7fd4f542-kube-api-access-bf2gp\") pod \"70bd3725-f383-45ff-a62f-45af7fd4f542\" (UID: \"70bd3725-f383-45ff-a62f-45af7fd4f542\") " Feb 27 09:10:05 crc kubenswrapper[4612]: I0227 09:10:05.504856 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/70bd3725-f383-45ff-a62f-45af7fd4f542-kube-api-access-bf2gp" (OuterVolumeSpecName: "kube-api-access-bf2gp") pod "70bd3725-f383-45ff-a62f-45af7fd4f542" (UID: "70bd3725-f383-45ff-a62f-45af7fd4f542"). InnerVolumeSpecName "kube-api-access-bf2gp". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 09:10:05 crc kubenswrapper[4612]: I0227 09:10:05.600728 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2gp\" (UniqueName: \"kubernetes.io/projected/70bd3725-f383-45ff-a62f-45af7fd4f542-kube-api-access-bf2gp\") on node \"crc\" DevicePath \"\"" Feb 27 09:10:05 crc kubenswrapper[4612]: I0227 09:10:05.955407 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29536384-d7rj9"] Feb 27 09:10:05 crc kubenswrapper[4612]: I0227 09:10:05.967966 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29536384-d7rj9"] Feb 27 09:10:06 crc kubenswrapper[4612]: I0227 09:10:06.060570 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536390-lz95w" event={"ID":"70bd3725-f383-45ff-a62f-45af7fd4f542","Type":"ContainerDied","Data":"2a5ecb2bf25b84b817180176e08b18b3970e9216a412a6d75c9edc56b121e962"} Feb 27 09:10:06 crc kubenswrapper[4612]: I0227 09:10:06.060610 4612 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2a5ecb2bf25b84b817180176e08b18b3970e9216a412a6d75c9edc56b121e962" Feb 27 09:10:06 crc kubenswrapper[4612]: I0227 09:10:06.060659 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536390-lz95w" Feb 27 09:10:06 crc kubenswrapper[4612]: I0227 09:10:06.867322 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fffd01d6-8461-4815-bd7f-52b0fb816346" path="/var/lib/kubelet/pods/fffd01d6-8461-4815-bd7f-52b0fb816346/volumes" Feb 27 09:10:11 crc kubenswrapper[4612]: I0227 09:10:11.162615 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-7db7f7dcc8-f95qr_16621861-feac-4972-ad1c-1e365692c23e/barbican-api/0.log" Feb 27 09:10:11 crc kubenswrapper[4612]: I0227 09:10:11.307104 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-7db7f7dcc8-f95qr_16621861-feac-4972-ad1c-1e365692c23e/barbican-api-log/0.log" Feb 27 09:10:11 crc kubenswrapper[4612]: I0227 09:10:11.387629 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-745b984886-tzpzt_e2579b13-034e-474a-bbe6-2062a710f380/barbican-keystone-listener/0.log" Feb 27 09:10:11 crc kubenswrapper[4612]: I0227 09:10:11.496220 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-745b984886-tzpzt_e2579b13-034e-474a-bbe6-2062a710f380/barbican-keystone-listener-log/0.log" Feb 27 09:10:11 crc kubenswrapper[4612]: I0227 09:10:11.620568 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-7558bddf6c-j9f2z_74bec848-d5b3-43d0-97d7-eda229642bb5/barbican-worker/0.log" Feb 27 09:10:11 crc kubenswrapper[4612]: I0227 09:10:11.657520 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-7558bddf6c-j9f2z_74bec848-d5b3-43d0-97d7-eda229642bb5/barbican-worker-log/0.log" Feb 27 09:10:12 crc kubenswrapper[4612]: I0227 09:10:12.347434 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_16ab2b55-65e7-4d63-b521-0c6421ba52ea/ceilometer-central-agent/0.log" Feb 27 09:10:12 crc kubenswrapper[4612]: I0227 09:10:12.414474 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-k4jmw_9caaee00-687f-47de-8336-859a249751f7/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Feb 27 09:10:12 crc kubenswrapper[4612]: I0227 09:10:12.442311 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_16ab2b55-65e7-4d63-b521-0c6421ba52ea/ceilometer-notification-agent/0.log" Feb 27 09:10:12 crc kubenswrapper[4612]: I0227 09:10:12.567837 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_16ab2b55-65e7-4d63-b521-0c6421ba52ea/sg-core/0.log" Feb 27 09:10:12 crc kubenswrapper[4612]: I0227 09:10:12.608963 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_16ab2b55-65e7-4d63-b521-0c6421ba52ea/proxy-httpd/0.log" Feb 27 09:10:12 crc kubenswrapper[4612]: I0227 09:10:12.739324 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_9ba31a95-0fce-4eb6-98be-6292b7146e21/cinder-api/0.log" Feb 27 09:10:12 crc kubenswrapper[4612]: I0227 09:10:12.838700 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_9ba31a95-0fce-4eb6-98be-6292b7146e21/cinder-api-log/0.log" Feb 27 09:10:12 crc kubenswrapper[4612]: I0227 09:10:12.970520 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_0cf1b618-59b9-4688-85b9-1a7f6d2af657/cinder-scheduler/0.log" Feb 27 09:10:13 crc kubenswrapper[4612]: I0227 09:10:13.041153 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_0cf1b618-59b9-4688-85b9-1a7f6d2af657/probe/0.log" Feb 27 09:10:13 crc kubenswrapper[4612]: I0227 09:10:13.180277 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-v4fds_c6f08c4d-80c1-49fb-b8ac-6d25b9d7f143/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Feb 27 09:10:13 crc kubenswrapper[4612]: I0227 09:10:13.273410 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-w9sl2_68ded60c-e2ee-48fb-986b-cebf9e447b82/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Feb 27 09:10:13 crc kubenswrapper[4612]: I0227 09:10:13.390737 4612 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-vxhvr" podUID="247cf880-b83d-4493-b12f-e25afaa861fd" containerName="registry-server" probeResult="failure" output=< Feb 27 09:10:13 crc kubenswrapper[4612]: timeout: failed to connect service ":50051" within 1s Feb 27 09:10:13 crc kubenswrapper[4612]: > Feb 27 09:10:14 crc kubenswrapper[4612]: I0227 09:10:14.007064 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-798f84bf5f-g5zjp_3c836155-a5e3-4181-9460-4147825f36ab/init/0.log" Feb 27 09:10:14 crc kubenswrapper[4612]: I0227 09:10:14.323347 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-t4m6d_8439abb9-0c80-44f9-9ec4-ef0c99a50182/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Feb 27 09:10:14 crc kubenswrapper[4612]: I0227 09:10:14.330068 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-798f84bf5f-g5zjp_3c836155-a5e3-4181-9460-4147825f36ab/init/0.log" Feb 27 09:10:14 crc kubenswrapper[4612]: I0227 09:10:14.426636 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-798f84bf5f-g5zjp_3c836155-a5e3-4181-9460-4147825f36ab/dnsmasq-dns/0.log" Feb 27 09:10:14 crc kubenswrapper[4612]: I0227 09:10:14.580749 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_58945c32-1660-410d-9a49-b0a38829f907/glance-httpd/0.log" Feb 27 09:10:14 crc kubenswrapper[4612]: I0227 09:10:14.639268 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_58945c32-1660-410d-9a49-b0a38829f907/glance-log/0.log" Feb 27 09:10:14 crc kubenswrapper[4612]: I0227 09:10:14.873948 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_f4ea14ec-7119-4952-a548-8afa0ad086bb/glance-log/0.log" Feb 27 09:10:14 crc kubenswrapper[4612]: I0227 09:10:14.941456 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_f4ea14ec-7119-4952-a548-8afa0ad086bb/glance-httpd/0.log" Feb 27 09:10:15 crc kubenswrapper[4612]: I0227 09:10:15.360514 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-77948db5bb-97qh6_b72b541d-4661-44a7-a121-c2a8aef6db11/horizon/1.log" Feb 27 09:10:15 crc kubenswrapper[4612]: I0227 09:10:15.368442 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-77948db5bb-97qh6_b72b541d-4661-44a7-a121-c2a8aef6db11/horizon/0.log" Feb 27 09:10:15 crc kubenswrapper[4612]: I0227 09:10:15.452356 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-frrmr_4f248d51-18f9-4d20-b64b-d23765054818/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Feb 27 09:10:15 crc kubenswrapper[4612]: I0227 09:10:15.620807 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-7fwd7_301bc35d-5ee7-4a2e-af9b-47a57e6c5123/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Feb 27 09:10:15 crc kubenswrapper[4612]: I0227 09:10:15.730046 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-77948db5bb-97qh6_b72b541d-4661-44a7-a121-c2a8aef6db11/horizon-log/0.log" Feb 27 09:10:15 crc kubenswrapper[4612]: I0227 09:10:15.874149 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29536381-vc87r_7aa90d53-2716-4f3f-aacf-fe667ce25f5d/keystone-cron/0.log" Feb 27 09:10:16 crc kubenswrapper[4612]: I0227 09:10:16.341046 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-9c6fdd6-lvn8p_71940e05-ecfa-46de-aefe-c213add5362c/keystone-api/0.log" Feb 27 09:10:16 crc kubenswrapper[4612]: I0227 09:10:16.372825 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_7a3af89c-deaa-4d32-b00b-ae3ca1de8ee9/kube-state-metrics/0.log" Feb 27 09:10:16 crc kubenswrapper[4612]: I0227 09:10:16.458235 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-27c56_f4308354-0bda-4fb3-976f-cc2fa471dcb4/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Feb 27 09:10:16 crc kubenswrapper[4612]: I0227 09:10:16.991630 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-x52ht_a213a1c0-9409-487d-834a-a60e11fd8a66/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Feb 27 09:10:17 crc kubenswrapper[4612]: I0227 09:10:17.142270 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-cc6d746d7-b8wbb_41eb38cf-80fa-408e-b16c-686ef58a143f/neutron-httpd/0.log" Feb 27 09:10:17 crc kubenswrapper[4612]: I0227 09:10:17.392887 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-cc6d746d7-b8wbb_41eb38cf-80fa-408e-b16c-686ef58a143f/neutron-api/0.log" Feb 27 09:10:18 crc kubenswrapper[4612]: I0227 09:10:18.143282 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_602ab3aa-60d5-45ff-9890-647a6f8ffb6c/nova-cell0-conductor-conductor/0.log" Feb 27 09:10:18 crc kubenswrapper[4612]: I0227 09:10:18.182680 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_fb906a42-d4ee-4e3e-b4c2-8924890a9e26/nova-cell1-conductor-conductor/0.log" Feb 27 09:10:18 crc kubenswrapper[4612]: I0227 09:10:18.600376 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_e492f61b-94d4-4c4e-ac47-168fdb1cd997/nova-api-log/0.log" Feb 27 09:10:18 crc kubenswrapper[4612]: I0227 09:10:18.629799 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_211046a5-967c-44f0-8320-ddcc223dda97/nova-cell1-novncproxy-novncproxy/0.log" Feb 27 09:10:18 crc kubenswrapper[4612]: I0227 09:10:18.959941 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-bnlnx_24758178-1be1-49f2-a15e-196426ec0dcd/nova-edpm-deployment-openstack-edpm-ipam/0.log" Feb 27 09:10:19 crc kubenswrapper[4612]: I0227 09:10:19.029432 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_e492f61b-94d4-4c4e-ac47-168fdb1cd997/nova-api-api/0.log" Feb 27 09:10:19 crc kubenswrapper[4612]: I0227 09:10:19.038860 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_24c836ae-ba59-4432-acee-fd17b60c8236/nova-metadata-log/0.log" Feb 27 09:10:19 crc kubenswrapper[4612]: I0227 09:10:19.582748 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_e02a40f2-3de1-4f01-84b0-2fe56acdd396/mysql-bootstrap/0.log" Feb 27 09:10:19 crc kubenswrapper[4612]: I0227 09:10:19.815205 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_e02a40f2-3de1-4f01-84b0-2fe56acdd396/mysql-bootstrap/0.log" Feb 27 09:10:19 crc kubenswrapper[4612]: I0227 09:10:19.855307 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_e02a40f2-3de1-4f01-84b0-2fe56acdd396/galera/0.log" Feb 27 09:10:19 crc kubenswrapper[4612]: I0227 09:10:19.911729 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_43270ba1-f597-4ca7-bc1e-741d7b915940/nova-scheduler-scheduler/0.log" Feb 27 09:10:20 crc kubenswrapper[4612]: I0227 09:10:20.137590 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_f189e394-0c97-48a4-9c55-4bd6710923f1/mysql-bootstrap/0.log" Feb 27 09:10:20 crc kubenswrapper[4612]: I0227 09:10:20.291451 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_f189e394-0c97-48a4-9c55-4bd6710923f1/mysql-bootstrap/0.log" Feb 27 09:10:20 crc kubenswrapper[4612]: I0227 09:10:20.442512 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_f189e394-0c97-48a4-9c55-4bd6710923f1/galera/0.log" Feb 27 09:10:20 crc kubenswrapper[4612]: I0227 09:10:20.555168 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_9fd44447-aa5f-436f-8aaa-bf3b53d99e64/openstackclient/0.log" Feb 27 09:10:20 crc kubenswrapper[4612]: I0227 09:10:20.619326 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-m6w89_99cc2aa0-2437-49fe-a7e0-1df45cecd195/openstack-network-exporter/0.log" Feb 27 09:10:20 crc kubenswrapper[4612]: I0227 09:10:20.923186 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-x255x_20c3f793-4286-4067-a372-37b2de08fa8e/ovsdb-server-init/0.log" Feb 27 09:10:21 crc kubenswrapper[4612]: I0227 09:10:21.065530 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_24c836ae-ba59-4432-acee-fd17b60c8236/nova-metadata-metadata/0.log" Feb 27 09:10:21 crc kubenswrapper[4612]: I0227 09:10:21.068403 4612 scope.go:117] "RemoveContainer" containerID="e5379e138e2871c6b7feb18a479f446e54a957a15b725efb3f8cb9700829b830" Feb 27 09:10:21 crc kubenswrapper[4612]: I0227 09:10:21.117114 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-x255x_20c3f793-4286-4067-a372-37b2de08fa8e/ovsdb-server-init/0.log" Feb 27 09:10:21 crc kubenswrapper[4612]: I0227 09:10:21.138841 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-x255x_20c3f793-4286-4067-a372-37b2de08fa8e/ovs-vswitchd/0.log" Feb 27 09:10:21 crc kubenswrapper[4612]: I0227 09:10:21.180555 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_d19bfe90-c5ca-4089-b4f0-bf3eccff30c8/memcached/0.log" Feb 27 09:10:21 crc kubenswrapper[4612]: I0227 09:10:21.221525 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-x255x_20c3f793-4286-4067-a372-37b2de08fa8e/ovsdb-server/0.log" Feb 27 09:10:21 crc kubenswrapper[4612]: I0227 09:10:21.347816 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-z7f9r_1ef66a51-46e7-4b76-a70b-b475d47822fd/ovn-controller/0.log" Feb 27 09:10:21 crc kubenswrapper[4612]: I0227 09:10:21.428994 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-xhcrn_dfdf6a2b-0ca5-4ea9-b87c-fd7a61f4981c/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Feb 27 09:10:21 crc kubenswrapper[4612]: I0227 09:10:21.720905 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_44e7bca8-b600-4ae0-8205-28f110f82dae/openstack-network-exporter/0.log" Feb 27 09:10:21 crc kubenswrapper[4612]: I0227 09:10:21.806195 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_44e7bca8-b600-4ae0-8205-28f110f82dae/ovn-northd/0.log" Feb 27 09:10:21 crc kubenswrapper[4612]: I0227 09:10:21.908376 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_1336636f-bc7a-45d4-a2ee-886a6f64ecd3/ovsdbserver-nb/0.log" Feb 27 09:10:21 crc kubenswrapper[4612]: I0227 09:10:21.921361 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_1336636f-bc7a-45d4-a2ee-886a6f64ecd3/openstack-network-exporter/0.log" Feb 27 09:10:22 crc kubenswrapper[4612]: I0227 09:10:22.013773 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_ddac95c9-8d81-420b-836e-d1902ce8beb5/openstack-network-exporter/0.log" Feb 27 09:10:22 crc kubenswrapper[4612]: I0227 09:10:22.126241 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_ddac95c9-8d81-420b-836e-d1902ce8beb5/ovsdbserver-sb/0.log" Feb 27 09:10:22 crc kubenswrapper[4612]: I0227 09:10:22.331967 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-7d477cc4b6-5qtkx_5fb13c78-0dab-4785-90bc-21d5dabe81f6/placement-api/0.log" Feb 27 09:10:22 crc kubenswrapper[4612]: I0227 09:10:22.384828 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_40399d37-c675-448e-8b7e-eb50b5ed5629/setup-container/0.log" Feb 27 09:10:22 crc kubenswrapper[4612]: I0227 09:10:22.408142 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-7d477cc4b6-5qtkx_5fb13c78-0dab-4785-90bc-21d5dabe81f6/placement-log/0.log" Feb 27 09:10:22 crc kubenswrapper[4612]: I0227 09:10:22.613201 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_40399d37-c675-448e-8b7e-eb50b5ed5629/rabbitmq/0.log" Feb 27 09:10:22 crc kubenswrapper[4612]: I0227 09:10:22.662962 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_ba8e827f-294b-4532-8636-c81eeb511ba4/setup-container/0.log" Feb 27 09:10:22 crc kubenswrapper[4612]: I0227 09:10:22.682483 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_40399d37-c675-448e-8b7e-eb50b5ed5629/setup-container/0.log" Feb 27 09:10:22 crc kubenswrapper[4612]: I0227 09:10:22.834654 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_ba8e827f-294b-4532-8636-c81eeb511ba4/rabbitmq/0.log" Feb 27 09:10:22 crc kubenswrapper[4612]: I0227 09:10:22.847552 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_ba8e827f-294b-4532-8636-c81eeb511ba4/setup-container/0.log" Feb 27 09:10:22 crc kubenswrapper[4612]: I0227 09:10:22.913707 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-cl6sd_ac2d7b40-cd87-4726-8225-6c9ce8afb4d7/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Feb 27 09:10:23 crc kubenswrapper[4612]: I0227 09:10:23.030600 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-vz4qv_9995bc24-fa2b-4916-8e71-627b11a39de7/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Feb 27 09:10:23 crc kubenswrapper[4612]: I0227 09:10:23.113636 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-g99jn_ab3a1ccc-dae6-40e3-859e-e4c0997b9071/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Feb 27 09:10:23 crc kubenswrapper[4612]: I0227 09:10:23.236867 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-psmxx_340ef584-802e-4696-b1b2-c7c8be815482/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Feb 27 09:10:23 crc kubenswrapper[4612]: I0227 09:10:23.387208 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-2g86c_d99b6ff5-a143-40ce-bee0-d4300122aff2/ssh-known-hosts-edpm-deployment/0.log" Feb 27 09:10:23 crc kubenswrapper[4612]: I0227 09:10:23.391356 4612 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-vxhvr" podUID="247cf880-b83d-4493-b12f-e25afaa861fd" containerName="registry-server" probeResult="failure" output=< Feb 27 09:10:23 crc kubenswrapper[4612]: timeout: failed to connect service ":50051" within 1s Feb 27 09:10:23 crc kubenswrapper[4612]: > Feb 27 09:10:23 crc kubenswrapper[4612]: I0227 09:10:23.504766 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-59ddd57cfc-qvv4h_2e631b0d-d44b-472b-90d9-6794cf71d3fd/proxy-server/0.log" Feb 27 09:10:23 crc kubenswrapper[4612]: I0227 09:10:23.530381 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-59ddd57cfc-qvv4h_2e631b0d-d44b-472b-90d9-6794cf71d3fd/proxy-httpd/0.log" Feb 27 09:10:23 crc kubenswrapper[4612]: I0227 09:10:23.609018 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-gp8mt_76db835b-ff4b-4d34-9578-38c5e3a605ef/swift-ring-rebalance/0.log" Feb 27 09:10:23 crc kubenswrapper[4612]: I0227 09:10:23.685008 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_77288860-cf3c-47e6-a752-5b41aa1967fa/account-auditor/0.log" Feb 27 09:10:23 crc kubenswrapper[4612]: I0227 09:10:23.761441 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_77288860-cf3c-47e6-a752-5b41aa1967fa/account-reaper/0.log" Feb 27 09:10:23 crc kubenswrapper[4612]: I0227 09:10:23.821258 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_77288860-cf3c-47e6-a752-5b41aa1967fa/account-replicator/0.log" Feb 27 09:10:23 crc kubenswrapper[4612]: I0227 09:10:23.880312 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_77288860-cf3c-47e6-a752-5b41aa1967fa/container-auditor/0.log" Feb 27 09:10:23 crc kubenswrapper[4612]: I0227 09:10:23.923917 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_77288860-cf3c-47e6-a752-5b41aa1967fa/account-server/0.log" Feb 27 09:10:23 crc kubenswrapper[4612]: I0227 09:10:23.990803 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_77288860-cf3c-47e6-a752-5b41aa1967fa/container-replicator/0.log" Feb 27 09:10:23 crc kubenswrapper[4612]: I0227 09:10:23.998237 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_77288860-cf3c-47e6-a752-5b41aa1967fa/container-server/0.log" Feb 27 09:10:24 crc kubenswrapper[4612]: I0227 09:10:24.040799 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_77288860-cf3c-47e6-a752-5b41aa1967fa/container-updater/0.log" Feb 27 09:10:24 crc kubenswrapper[4612]: I0227 09:10:24.139285 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_77288860-cf3c-47e6-a752-5b41aa1967fa/object-auditor/0.log" Feb 27 09:10:24 crc kubenswrapper[4612]: I0227 09:10:24.225751 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_77288860-cf3c-47e6-a752-5b41aa1967fa/object-expirer/0.log" Feb 27 09:10:24 crc kubenswrapper[4612]: I0227 09:10:24.226470 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_77288860-cf3c-47e6-a752-5b41aa1967fa/object-replicator/0.log" Feb 27 09:10:24 crc kubenswrapper[4612]: I0227 09:10:24.260406 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_77288860-cf3c-47e6-a752-5b41aa1967fa/object-server/0.log" Feb 27 09:10:24 crc kubenswrapper[4612]: I0227 09:10:24.260807 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_77288860-cf3c-47e6-a752-5b41aa1967fa/object-updater/0.log" Feb 27 09:10:24 crc kubenswrapper[4612]: I0227 09:10:24.353235 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_77288860-cf3c-47e6-a752-5b41aa1967fa/rsync/0.log" Feb 27 09:10:25 crc kubenswrapper[4612]: I0227 09:10:25.010183 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_77288860-cf3c-47e6-a752-5b41aa1967fa/swift-recon-cron/0.log" Feb 27 09:10:25 crc kubenswrapper[4612]: I0227 09:10:25.116045 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_4700642e-da3a-4e53-84de-0bf4c1870ccd/tempest-tests-tempest-tests-runner/0.log" Feb 27 09:10:25 crc kubenswrapper[4612]: I0227 09:10:25.216566 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-qthr6_f69e6377-1d38-4128-b6ad-af91cf8f7793/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Feb 27 09:10:25 crc kubenswrapper[4612]: I0227 09:10:25.382727 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_8aad3c36-76d6-4991-9494-0725b3e67aea/test-operator-logs-container/0.log" Feb 27 09:10:25 crc kubenswrapper[4612]: I0227 09:10:25.416294 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-vfpr6_ff510046-7399-4fdd-9880-67f6a51b61db/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Feb 27 09:10:32 crc kubenswrapper[4612]: I0227 09:10:32.381771 4612 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-vxhvr" Feb 27 09:10:32 crc kubenswrapper[4612]: I0227 09:10:32.430246 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-vxhvr" Feb 27 09:10:32 crc kubenswrapper[4612]: I0227 09:10:32.618078 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-vxhvr"] Feb 27 09:10:34 crc kubenswrapper[4612]: I0227 09:10:34.291173 4612 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-vxhvr" podUID="247cf880-b83d-4493-b12f-e25afaa861fd" containerName="registry-server" containerID="cri-o://8dca1d714a17638cab306ee7f33ed3d80d59826130f47ba5fc6d2a63ab292164" gracePeriod=2 Feb 27 09:10:34 crc kubenswrapper[4612]: I0227 09:10:34.771744 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vxhvr" Feb 27 09:10:34 crc kubenswrapper[4612]: I0227 09:10:34.836900 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6fxfj\" (UniqueName: \"kubernetes.io/projected/247cf880-b83d-4493-b12f-e25afaa861fd-kube-api-access-6fxfj\") pod \"247cf880-b83d-4493-b12f-e25afaa861fd\" (UID: \"247cf880-b83d-4493-b12f-e25afaa861fd\") " Feb 27 09:10:34 crc kubenswrapper[4612]: I0227 09:10:34.836976 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/247cf880-b83d-4493-b12f-e25afaa861fd-utilities\") pod \"247cf880-b83d-4493-b12f-e25afaa861fd\" (UID: \"247cf880-b83d-4493-b12f-e25afaa861fd\") " Feb 27 09:10:34 crc kubenswrapper[4612]: I0227 09:10:34.837060 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/247cf880-b83d-4493-b12f-e25afaa861fd-catalog-content\") pod \"247cf880-b83d-4493-b12f-e25afaa861fd\" (UID: \"247cf880-b83d-4493-b12f-e25afaa861fd\") " Feb 27 09:10:34 crc kubenswrapper[4612]: I0227 09:10:34.839291 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/247cf880-b83d-4493-b12f-e25afaa861fd-utilities" (OuterVolumeSpecName: "utilities") pod "247cf880-b83d-4493-b12f-e25afaa861fd" (UID: "247cf880-b83d-4493-b12f-e25afaa861fd"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 09:10:34 crc kubenswrapper[4612]: I0227 09:10:34.842754 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/247cf880-b83d-4493-b12f-e25afaa861fd-kube-api-access-6fxfj" (OuterVolumeSpecName: "kube-api-access-6fxfj") pod "247cf880-b83d-4493-b12f-e25afaa861fd" (UID: "247cf880-b83d-4493-b12f-e25afaa861fd"). InnerVolumeSpecName "kube-api-access-6fxfj". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 09:10:34 crc kubenswrapper[4612]: I0227 09:10:34.939727 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6fxfj\" (UniqueName: \"kubernetes.io/projected/247cf880-b83d-4493-b12f-e25afaa861fd-kube-api-access-6fxfj\") on node \"crc\" DevicePath \"\"" Feb 27 09:10:34 crc kubenswrapper[4612]: I0227 09:10:34.939949 4612 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/247cf880-b83d-4493-b12f-e25afaa861fd-utilities\") on node \"crc\" DevicePath \"\"" Feb 27 09:10:35 crc kubenswrapper[4612]: I0227 09:10:35.008398 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/247cf880-b83d-4493-b12f-e25afaa861fd-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "247cf880-b83d-4493-b12f-e25afaa861fd" (UID: "247cf880-b83d-4493-b12f-e25afaa861fd"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 09:10:35 crc kubenswrapper[4612]: I0227 09:10:35.041971 4612 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/247cf880-b83d-4493-b12f-e25afaa861fd-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 27 09:10:35 crc kubenswrapper[4612]: I0227 09:10:35.301873 4612 generic.go:334] "Generic (PLEG): container finished" podID="247cf880-b83d-4493-b12f-e25afaa861fd" containerID="8dca1d714a17638cab306ee7f33ed3d80d59826130f47ba5fc6d2a63ab292164" exitCode=0 Feb 27 09:10:35 crc kubenswrapper[4612]: I0227 09:10:35.301918 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vxhvr" event={"ID":"247cf880-b83d-4493-b12f-e25afaa861fd","Type":"ContainerDied","Data":"8dca1d714a17638cab306ee7f33ed3d80d59826130f47ba5fc6d2a63ab292164"} Feb 27 09:10:35 crc kubenswrapper[4612]: I0227 09:10:35.301943 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vxhvr" event={"ID":"247cf880-b83d-4493-b12f-e25afaa861fd","Type":"ContainerDied","Data":"8f99a17ced44e454690efa852ec537d1c00e7e927e254f2da8e0a2ac9e3b6f9a"} Feb 27 09:10:35 crc kubenswrapper[4612]: I0227 09:10:35.301959 4612 scope.go:117] "RemoveContainer" containerID="8dca1d714a17638cab306ee7f33ed3d80d59826130f47ba5fc6d2a63ab292164" Feb 27 09:10:35 crc kubenswrapper[4612]: I0227 09:10:35.302104 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vxhvr" Feb 27 09:10:35 crc kubenswrapper[4612]: I0227 09:10:35.355500 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-vxhvr"] Feb 27 09:10:35 crc kubenswrapper[4612]: I0227 09:10:35.362749 4612 scope.go:117] "RemoveContainer" containerID="83a2bbeaae26da1239f997503f6fa70146491baaab31b6cbaea96c7b5f953cb6" Feb 27 09:10:35 crc kubenswrapper[4612]: I0227 09:10:35.369534 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-vxhvr"] Feb 27 09:10:35 crc kubenswrapper[4612]: I0227 09:10:35.387486 4612 scope.go:117] "RemoveContainer" containerID="137cd75031a64cc229591f7f3dacce26a9b19e760575d5be1dfec51a63436cb4" Feb 27 09:10:35 crc kubenswrapper[4612]: I0227 09:10:35.435345 4612 scope.go:117] "RemoveContainer" containerID="8dca1d714a17638cab306ee7f33ed3d80d59826130f47ba5fc6d2a63ab292164" Feb 27 09:10:35 crc kubenswrapper[4612]: E0227 09:10:35.435748 4612 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8dca1d714a17638cab306ee7f33ed3d80d59826130f47ba5fc6d2a63ab292164\": container with ID starting with 8dca1d714a17638cab306ee7f33ed3d80d59826130f47ba5fc6d2a63ab292164 not found: ID does not exist" containerID="8dca1d714a17638cab306ee7f33ed3d80d59826130f47ba5fc6d2a63ab292164" Feb 27 09:10:35 crc kubenswrapper[4612]: I0227 09:10:35.435817 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8dca1d714a17638cab306ee7f33ed3d80d59826130f47ba5fc6d2a63ab292164"} err="failed to get container status \"8dca1d714a17638cab306ee7f33ed3d80d59826130f47ba5fc6d2a63ab292164\": rpc error: code = NotFound desc = could not find container \"8dca1d714a17638cab306ee7f33ed3d80d59826130f47ba5fc6d2a63ab292164\": container with ID starting with 8dca1d714a17638cab306ee7f33ed3d80d59826130f47ba5fc6d2a63ab292164 not found: ID does not exist" Feb 27 09:10:35 crc kubenswrapper[4612]: I0227 09:10:35.435849 4612 scope.go:117] "RemoveContainer" containerID="83a2bbeaae26da1239f997503f6fa70146491baaab31b6cbaea96c7b5f953cb6" Feb 27 09:10:35 crc kubenswrapper[4612]: E0227 09:10:35.438291 4612 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"83a2bbeaae26da1239f997503f6fa70146491baaab31b6cbaea96c7b5f953cb6\": container with ID starting with 83a2bbeaae26da1239f997503f6fa70146491baaab31b6cbaea96c7b5f953cb6 not found: ID does not exist" containerID="83a2bbeaae26da1239f997503f6fa70146491baaab31b6cbaea96c7b5f953cb6" Feb 27 09:10:35 crc kubenswrapper[4612]: I0227 09:10:35.438354 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"83a2bbeaae26da1239f997503f6fa70146491baaab31b6cbaea96c7b5f953cb6"} err="failed to get container status \"83a2bbeaae26da1239f997503f6fa70146491baaab31b6cbaea96c7b5f953cb6\": rpc error: code = NotFound desc = could not find container \"83a2bbeaae26da1239f997503f6fa70146491baaab31b6cbaea96c7b5f953cb6\": container with ID starting with 83a2bbeaae26da1239f997503f6fa70146491baaab31b6cbaea96c7b5f953cb6 not found: ID does not exist" Feb 27 09:10:35 crc kubenswrapper[4612]: I0227 09:10:35.438384 4612 scope.go:117] "RemoveContainer" containerID="137cd75031a64cc229591f7f3dacce26a9b19e760575d5be1dfec51a63436cb4" Feb 27 09:10:35 crc kubenswrapper[4612]: E0227 09:10:35.439942 4612 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"137cd75031a64cc229591f7f3dacce26a9b19e760575d5be1dfec51a63436cb4\": container with ID starting with 137cd75031a64cc229591f7f3dacce26a9b19e760575d5be1dfec51a63436cb4 not found: ID does not exist" containerID="137cd75031a64cc229591f7f3dacce26a9b19e760575d5be1dfec51a63436cb4" Feb 27 09:10:35 crc kubenswrapper[4612]: I0227 09:10:35.439967 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"137cd75031a64cc229591f7f3dacce26a9b19e760575d5be1dfec51a63436cb4"} err="failed to get container status \"137cd75031a64cc229591f7f3dacce26a9b19e760575d5be1dfec51a63436cb4\": rpc error: code = NotFound desc = could not find container \"137cd75031a64cc229591f7f3dacce26a9b19e760575d5be1dfec51a63436cb4\": container with ID starting with 137cd75031a64cc229591f7f3dacce26a9b19e760575d5be1dfec51a63436cb4 not found: ID does not exist" Feb 27 09:10:35 crc kubenswrapper[4612]: E0227 09:10:35.548338 4612 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod247cf880_b83d_4493_b12f_e25afaa861fd.slice/crio-8f99a17ced44e454690efa852ec537d1c00e7e927e254f2da8e0a2ac9e3b6f9a\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod247cf880_b83d_4493_b12f_e25afaa861fd.slice\": RecentStats: unable to find data in memory cache]" Feb 27 09:10:36 crc kubenswrapper[4612]: I0227 09:10:36.865432 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="247cf880-b83d-4493-b12f-e25afaa861fd" path="/var/lib/kubelet/pods/247cf880-b83d-4493-b12f-e25afaa861fd/volumes" Feb 27 09:10:52 crc kubenswrapper[4612]: I0227 09:10:52.030623 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_1e2105768d85fd59d354c5be43e261a5f6290b57040396b6c3a32dc181djh99_dfad17e6-8a21-44be-8328-2a156da54578/util/0.log" Feb 27 09:10:52 crc kubenswrapper[4612]: I0227 09:10:52.211864 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_1e2105768d85fd59d354c5be43e261a5f6290b57040396b6c3a32dc181djh99_dfad17e6-8a21-44be-8328-2a156da54578/util/0.log" Feb 27 09:10:52 crc kubenswrapper[4612]: I0227 09:10:52.270445 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_1e2105768d85fd59d354c5be43e261a5f6290b57040396b6c3a32dc181djh99_dfad17e6-8a21-44be-8328-2a156da54578/pull/0.log" Feb 27 09:10:52 crc kubenswrapper[4612]: I0227 09:10:52.305634 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_1e2105768d85fd59d354c5be43e261a5f6290b57040396b6c3a32dc181djh99_dfad17e6-8a21-44be-8328-2a156da54578/pull/0.log" Feb 27 09:10:52 crc kubenswrapper[4612]: I0227 09:10:52.453517 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_1e2105768d85fd59d354c5be43e261a5f6290b57040396b6c3a32dc181djh99_dfad17e6-8a21-44be-8328-2a156da54578/extract/0.log" Feb 27 09:10:52 crc kubenswrapper[4612]: I0227 09:10:52.464174 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_1e2105768d85fd59d354c5be43e261a5f6290b57040396b6c3a32dc181djh99_dfad17e6-8a21-44be-8328-2a156da54578/util/0.log" Feb 27 09:10:52 crc kubenswrapper[4612]: I0227 09:10:52.524563 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_1e2105768d85fd59d354c5be43e261a5f6290b57040396b6c3a32dc181djh99_dfad17e6-8a21-44be-8328-2a156da54578/pull/0.log" Feb 27 09:10:53 crc kubenswrapper[4612]: I0227 09:10:53.045622 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-6d8bf5c495-t95hp_63d519c2-d2f9-4349-ad82-0399e472d78a/manager/0.log" Feb 27 09:10:53 crc kubenswrapper[4612]: I0227 09:10:53.402928 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-784b5bb6c5-g4p5v_71c482bc-7818-4193-a80b-323682cdcfb3/manager/0.log" Feb 27 09:10:53 crc kubenswrapper[4612]: I0227 09:10:53.630734 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-69f49c598c-7f9bj_6065186a-5435-4db3-8adc-698bec4358ef/manager/0.log" Feb 27 09:10:54 crc kubenswrapper[4612]: I0227 09:10:54.310850 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-5b9b8895d5-t4rct_1903598c-2428-4e90-ab81-9a47fa5af9ef/manager/0.log" Feb 27 09:10:54 crc kubenswrapper[4612]: I0227 09:10:54.998188 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-79d975b745-plm9k_98296d45-b487-47c3-a44a-fa63ba52426b/manager/0.log" Feb 27 09:10:55 crc kubenswrapper[4612]: I0227 09:10:55.014115 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-554564d7fc-8kpwf_bd687bfa-c843-4d11-80f9-cedcdce41e3d/manager/0.log" Feb 27 09:10:55 crc kubenswrapper[4612]: I0227 09:10:55.248461 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-55d77d7b5c-462m7_36fc734e-4b3e-46bb-b914-896b1ec5c2ee/manager/0.log" Feb 27 09:10:55 crc kubenswrapper[4612]: I0227 09:10:55.316801 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-b4d948c87-2hm5l_259df4bd-af1b-4c61-a5bd-0ca28ac531d4/manager/0.log" Feb 27 09:10:55 crc kubenswrapper[4612]: I0227 09:10:55.544430 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-67d996989d-4dhhw_b60d4209-cd53-4cfe-99f7-7e7351041857/manager/0.log" Feb 27 09:10:55 crc kubenswrapper[4612]: I0227 09:10:55.831026 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-6994f66f48-qs4l7_f62fea1f-71de-4872-9baa-42ad2448d5f5/manager/0.log" Feb 27 09:10:55 crc kubenswrapper[4612]: I0227 09:10:55.848325 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-6bd4687957-2qmzb_f292b2e2-2b32-4e4e-bfca-d5db60f0b358/manager/0.log" Feb 27 09:10:56 crc kubenswrapper[4612]: I0227 09:10:56.927275 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-659dc6bbfc-vlczg_bd923861-d2ad-4a06-b0c7-613605afefbb/manager/0.log" Feb 27 09:10:56 crc kubenswrapper[4612]: I0227 09:10:56.966707 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-567668f5cf-tjqwp_d84adbda-b53c-46c7-9196-682910a0009c/manager/0.log" Feb 27 09:10:56 crc kubenswrapper[4612]: I0227 09:10:56.990246 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-7c6767dc9clp9dx_cf5ef40d-5e5c-4b2c-9570-63e552800b4e/manager/0.log" Feb 27 09:10:57 crc kubenswrapper[4612]: I0227 09:10:57.343259 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-init-8667ddcbf9-7stg7_a8e2b07a-a4cf-4a98-afeb-2a7554b687ee/operator/0.log" Feb 27 09:10:57 crc kubenswrapper[4612]: I0227 09:10:57.592328 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-f9cqc_5ed1acaa-e783-4d8d-bac0-adf0ab4a3b0f/registry-server/0.log" Feb 27 09:10:57 crc kubenswrapper[4612]: I0227 09:10:57.770938 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-5955d8c787-q74ht_e0304d11-114b-4aa3-86ca-1747c60b5bfc/manager/0.log" Feb 27 09:10:57 crc kubenswrapper[4612]: I0227 09:10:57.928646 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-8497b45c89-dvq8j_8414b96c-7e5b-4824-998e-549a0c43d9d5/manager/0.log" Feb 27 09:10:58 crc kubenswrapper[4612]: I0227 09:10:58.098276 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-8s6n2_57ba67ae-3e41-4e2b-9ee6-30b2643805dd/operator/0.log" Feb 27 09:10:58 crc kubenswrapper[4612]: I0227 09:10:58.306594 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-68f46476f-r8lqs_713a994e-490d-47b4-86ab-1203a6b806fc/manager/0.log" Feb 27 09:10:58 crc kubenswrapper[4612]: I0227 09:10:58.555549 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-589c568786-qclqp_ee3f21f6-2f91-45c6-8c80-5fb8f89a41ce/manager/0.log" Feb 27 09:10:58 crc kubenswrapper[4612]: I0227 09:10:58.609827 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5dc6794d5b-s2jnh_ad4fca59-6a87-4ebc-8681-d5a3a40f0b82/manager/0.log" Feb 27 09:10:58 crc kubenswrapper[4612]: I0227 09:10:58.761007 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-bccc79885-zdtzv_a4619f7b-7242-459b-8558-f84c2893ee55/manager/0.log" Feb 27 09:10:59 crc kubenswrapper[4612]: I0227 09:10:59.052685 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-69f9884c8f-v4kkz_41aed211-3a1f-4f7c-86ca-53e445dcd585/manager/0.log" Feb 27 09:11:03 crc kubenswrapper[4612]: I0227 09:11:03.433077 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-868647ff47-mlgz7_2ddc15ea-b7c1-44e7-b795-392890e9c14e/manager/0.log" Feb 27 09:11:22 crc kubenswrapper[4612]: I0227 09:11:22.539384 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-zpshm_d951f066-9e39-4e0a-9fa3-da9c868552d1/control-plane-machine-set-operator/0.log" Feb 27 09:11:22 crc kubenswrapper[4612]: I0227 09:11:22.752190 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-q6h7h_094b001c-e80c-4d07-a161-e00514667462/kube-rbac-proxy/0.log" Feb 27 09:11:22 crc kubenswrapper[4612]: I0227 09:11:22.776922 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-q6h7h_094b001c-e80c-4d07-a161-e00514667462/machine-api-operator/0.log" Feb 27 09:11:36 crc kubenswrapper[4612]: I0227 09:11:36.446288 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-858654f9db-lq868_095685d4-89ee-421e-ba88-a2df75c3a160/cert-manager-controller/0.log" Feb 27 09:11:36 crc kubenswrapper[4612]: I0227 09:11:36.574035 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-cf98fcc89-gk4r5_4df7eb22-d428-4e88-bdc5-0bab2b4806a3/cert-manager-cainjector/0.log" Feb 27 09:11:36 crc kubenswrapper[4612]: I0227 09:11:36.625218 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-687f57d79b-krsbj_5bba164a-78bf-4a9a-9147-729a1a91740c/cert-manager-webhook/0.log" Feb 27 09:11:46 crc kubenswrapper[4612]: I0227 09:11:46.027045 4612 patch_prober.go:28] interesting pod/machine-config-daemon-924vb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 27 09:11:46 crc kubenswrapper[4612]: I0227 09:11:46.027647 4612 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 27 09:11:50 crc kubenswrapper[4612]: I0227 09:11:50.404651 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-5dcbbd79cf-4cfq5_a82cc418-4f77-4788-addb-25341d42d9fa/nmstate-console-plugin/0.log" Feb 27 09:11:50 crc kubenswrapper[4612]: I0227 09:11:50.547988 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-26zsp_5553e6c7-b45f-40c9-99ce-abd46767d31a/nmstate-handler/0.log" Feb 27 09:11:50 crc kubenswrapper[4612]: I0227 09:11:50.586842 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-69594cc75-czssc_d058174c-69c3-49cf-b846-af074b709972/kube-rbac-proxy/0.log" Feb 27 09:11:50 crc kubenswrapper[4612]: I0227 09:11:50.604247 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-69594cc75-czssc_d058174c-69c3-49cf-b846-af074b709972/nmstate-metrics/0.log" Feb 27 09:11:50 crc kubenswrapper[4612]: I0227 09:11:50.822804 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-75c5dccd6c-p6nbk_49fa2d59-6289-4edc-a6d8-2c3f7733a536/nmstate-operator/0.log" Feb 27 09:11:50 crc kubenswrapper[4612]: I0227 09:11:50.869592 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-786f45cff4-m8llg_b141dc9e-912e-4096-97fc-5df6e0ae245c/nmstate-webhook/0.log" Feb 27 09:12:00 crc kubenswrapper[4612]: I0227 09:12:00.143661 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29536392-z9mh7"] Feb 27 09:12:00 crc kubenswrapper[4612]: E0227 09:12:00.145431 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="247cf880-b83d-4493-b12f-e25afaa861fd" containerName="registry-server" Feb 27 09:12:00 crc kubenswrapper[4612]: I0227 09:12:00.145450 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="247cf880-b83d-4493-b12f-e25afaa861fd" containerName="registry-server" Feb 27 09:12:00 crc kubenswrapper[4612]: E0227 09:12:00.145475 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="247cf880-b83d-4493-b12f-e25afaa861fd" containerName="extract-utilities" Feb 27 09:12:00 crc kubenswrapper[4612]: I0227 09:12:00.145501 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="247cf880-b83d-4493-b12f-e25afaa861fd" containerName="extract-utilities" Feb 27 09:12:00 crc kubenswrapper[4612]: E0227 09:12:00.145530 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="70bd3725-f383-45ff-a62f-45af7fd4f542" containerName="oc" Feb 27 09:12:00 crc kubenswrapper[4612]: I0227 09:12:00.145540 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="70bd3725-f383-45ff-a62f-45af7fd4f542" containerName="oc" Feb 27 09:12:00 crc kubenswrapper[4612]: E0227 09:12:00.145569 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="247cf880-b83d-4493-b12f-e25afaa861fd" containerName="extract-content" Feb 27 09:12:00 crc kubenswrapper[4612]: I0227 09:12:00.145577 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="247cf880-b83d-4493-b12f-e25afaa861fd" containerName="extract-content" Feb 27 09:12:00 crc kubenswrapper[4612]: I0227 09:12:00.145911 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="247cf880-b83d-4493-b12f-e25afaa861fd" containerName="registry-server" Feb 27 09:12:00 crc kubenswrapper[4612]: I0227 09:12:00.145939 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="70bd3725-f383-45ff-a62f-45af7fd4f542" containerName="oc" Feb 27 09:12:00 crc kubenswrapper[4612]: I0227 09:12:00.146669 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536392-z9mh7" Feb 27 09:12:00 crc kubenswrapper[4612]: I0227 09:12:00.149609 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 27 09:12:00 crc kubenswrapper[4612]: I0227 09:12:00.149803 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-5zwfx" Feb 27 09:12:00 crc kubenswrapper[4612]: I0227 09:12:00.149890 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 27 09:12:00 crc kubenswrapper[4612]: I0227 09:12:00.158153 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536392-z9mh7"] Feb 27 09:12:00 crc kubenswrapper[4612]: I0227 09:12:00.325943 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jrkgb\" (UniqueName: \"kubernetes.io/projected/087335eb-b232-4434-aca3-2ed95ec4350a-kube-api-access-jrkgb\") pod \"auto-csr-approver-29536392-z9mh7\" (UID: \"087335eb-b232-4434-aca3-2ed95ec4350a\") " pod="openshift-infra/auto-csr-approver-29536392-z9mh7" Feb 27 09:12:00 crc kubenswrapper[4612]: I0227 09:12:00.428318 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jrkgb\" (UniqueName: \"kubernetes.io/projected/087335eb-b232-4434-aca3-2ed95ec4350a-kube-api-access-jrkgb\") pod \"auto-csr-approver-29536392-z9mh7\" (UID: \"087335eb-b232-4434-aca3-2ed95ec4350a\") " pod="openshift-infra/auto-csr-approver-29536392-z9mh7" Feb 27 09:12:00 crc kubenswrapper[4612]: I0227 09:12:00.461807 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jrkgb\" (UniqueName: \"kubernetes.io/projected/087335eb-b232-4434-aca3-2ed95ec4350a-kube-api-access-jrkgb\") pod \"auto-csr-approver-29536392-z9mh7\" (UID: \"087335eb-b232-4434-aca3-2ed95ec4350a\") " pod="openshift-infra/auto-csr-approver-29536392-z9mh7" Feb 27 09:12:00 crc kubenswrapper[4612]: I0227 09:12:00.474092 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536392-z9mh7" Feb 27 09:12:00 crc kubenswrapper[4612]: I0227 09:12:00.975816 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536392-z9mh7"] Feb 27 09:12:00 crc kubenswrapper[4612]: I0227 09:12:00.980984 4612 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 27 09:12:01 crc kubenswrapper[4612]: I0227 09:12:01.042618 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536392-z9mh7" event={"ID":"087335eb-b232-4434-aca3-2ed95ec4350a","Type":"ContainerStarted","Data":"cd8ca4f0773213b8c1ed92cfb3e93654ed19a5395b23e6c0d9d3ee037bcbdc78"} Feb 27 09:12:03 crc kubenswrapper[4612]: I0227 09:12:03.066841 4612 generic.go:334] "Generic (PLEG): container finished" podID="087335eb-b232-4434-aca3-2ed95ec4350a" containerID="9c3bb02736109f1379c55cc74edc9bf0104377cd1266287b430d3a31e68c8c71" exitCode=0 Feb 27 09:12:03 crc kubenswrapper[4612]: I0227 09:12:03.067343 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536392-z9mh7" event={"ID":"087335eb-b232-4434-aca3-2ed95ec4350a","Type":"ContainerDied","Data":"9c3bb02736109f1379c55cc74edc9bf0104377cd1266287b430d3a31e68c8c71"} Feb 27 09:12:04 crc kubenswrapper[4612]: I0227 09:12:04.430303 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536392-z9mh7" Feb 27 09:12:04 crc kubenswrapper[4612]: I0227 09:12:04.611062 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jrkgb\" (UniqueName: \"kubernetes.io/projected/087335eb-b232-4434-aca3-2ed95ec4350a-kube-api-access-jrkgb\") pod \"087335eb-b232-4434-aca3-2ed95ec4350a\" (UID: \"087335eb-b232-4434-aca3-2ed95ec4350a\") " Feb 27 09:12:04 crc kubenswrapper[4612]: I0227 09:12:04.621147 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/087335eb-b232-4434-aca3-2ed95ec4350a-kube-api-access-jrkgb" (OuterVolumeSpecName: "kube-api-access-jrkgb") pod "087335eb-b232-4434-aca3-2ed95ec4350a" (UID: "087335eb-b232-4434-aca3-2ed95ec4350a"). InnerVolumeSpecName "kube-api-access-jrkgb". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 09:12:04 crc kubenswrapper[4612]: I0227 09:12:04.714000 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jrkgb\" (UniqueName: \"kubernetes.io/projected/087335eb-b232-4434-aca3-2ed95ec4350a-kube-api-access-jrkgb\") on node \"crc\" DevicePath \"\"" Feb 27 09:12:05 crc kubenswrapper[4612]: I0227 09:12:05.084403 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536392-z9mh7" event={"ID":"087335eb-b232-4434-aca3-2ed95ec4350a","Type":"ContainerDied","Data":"cd8ca4f0773213b8c1ed92cfb3e93654ed19a5395b23e6c0d9d3ee037bcbdc78"} Feb 27 09:12:05 crc kubenswrapper[4612]: I0227 09:12:05.084440 4612 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cd8ca4f0773213b8c1ed92cfb3e93654ed19a5395b23e6c0d9d3ee037bcbdc78" Feb 27 09:12:05 crc kubenswrapper[4612]: I0227 09:12:05.084460 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536392-z9mh7" Feb 27 09:12:05 crc kubenswrapper[4612]: I0227 09:12:05.497762 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29536386-j6cvj"] Feb 27 09:12:05 crc kubenswrapper[4612]: I0227 09:12:05.506721 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29536386-j6cvj"] Feb 27 09:12:06 crc kubenswrapper[4612]: I0227 09:12:06.868871 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="018188bb-5ac2-4e9a-8bc0-7f7c1ead9115" path="/var/lib/kubelet/pods/018188bb-5ac2-4e9a-8bc0-7f7c1ead9115/volumes" Feb 27 09:12:16 crc kubenswrapper[4612]: I0227 09:12:16.027246 4612 patch_prober.go:28] interesting pod/machine-config-daemon-924vb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 27 09:12:16 crc kubenswrapper[4612]: I0227 09:12:16.028046 4612 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 27 09:12:21 crc kubenswrapper[4612]: I0227 09:12:21.206940 4612 scope.go:117] "RemoveContainer" containerID="81f9b9bc738c2d1a0d26637aac53e51282a564cb0cc257319f7a8be3d4c7f5e5" Feb 27 09:12:21 crc kubenswrapper[4612]: I0227 09:12:21.598378 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-86ddb6bd46-gk8pf_2b49ec1b-e464-4e71-badd-04acfa13ef75/controller/0.log" Feb 27 09:12:21 crc kubenswrapper[4612]: I0227 09:12:21.607887 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-86ddb6bd46-gk8pf_2b49ec1b-e464-4e71-badd-04acfa13ef75/kube-rbac-proxy/0.log" Feb 27 09:12:21 crc kubenswrapper[4612]: I0227 09:12:21.801804 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-8drht_8fe63fdd-8b6f-4ecf-81d2-959ae060d567/cp-frr-files/0.log" Feb 27 09:12:21 crc kubenswrapper[4612]: I0227 09:12:21.996779 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-8drht_8fe63fdd-8b6f-4ecf-81d2-959ae060d567/cp-reloader/0.log" Feb 27 09:12:22 crc kubenswrapper[4612]: I0227 09:12:22.045031 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-8drht_8fe63fdd-8b6f-4ecf-81d2-959ae060d567/cp-metrics/0.log" Feb 27 09:12:22 crc kubenswrapper[4612]: I0227 09:12:22.057438 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-8drht_8fe63fdd-8b6f-4ecf-81d2-959ae060d567/cp-frr-files/0.log" Feb 27 09:12:22 crc kubenswrapper[4612]: I0227 09:12:22.120632 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-8drht_8fe63fdd-8b6f-4ecf-81d2-959ae060d567/cp-reloader/0.log" Feb 27 09:12:22 crc kubenswrapper[4612]: I0227 09:12:22.253099 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-8drht_8fe63fdd-8b6f-4ecf-81d2-959ae060d567/cp-reloader/0.log" Feb 27 09:12:22 crc kubenswrapper[4612]: I0227 09:12:22.283326 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-8drht_8fe63fdd-8b6f-4ecf-81d2-959ae060d567/cp-frr-files/0.log" Feb 27 09:12:22 crc kubenswrapper[4612]: I0227 09:12:22.310498 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-8drht_8fe63fdd-8b6f-4ecf-81d2-959ae060d567/cp-metrics/0.log" Feb 27 09:12:22 crc kubenswrapper[4612]: I0227 09:12:22.334849 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-8drht_8fe63fdd-8b6f-4ecf-81d2-959ae060d567/cp-metrics/0.log" Feb 27 09:12:22 crc kubenswrapper[4612]: I0227 09:12:22.520072 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-8drht_8fe63fdd-8b6f-4ecf-81d2-959ae060d567/cp-metrics/0.log" Feb 27 09:12:22 crc kubenswrapper[4612]: I0227 09:12:22.528238 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-8drht_8fe63fdd-8b6f-4ecf-81d2-959ae060d567/cp-reloader/0.log" Feb 27 09:12:22 crc kubenswrapper[4612]: I0227 09:12:22.535902 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-8drht_8fe63fdd-8b6f-4ecf-81d2-959ae060d567/cp-frr-files/0.log" Feb 27 09:12:22 crc kubenswrapper[4612]: I0227 09:12:22.557032 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-8drht_8fe63fdd-8b6f-4ecf-81d2-959ae060d567/controller/0.log" Feb 27 09:12:22 crc kubenswrapper[4612]: I0227 09:12:22.801785 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-8drht_8fe63fdd-8b6f-4ecf-81d2-959ae060d567/kube-rbac-proxy/0.log" Feb 27 09:12:22 crc kubenswrapper[4612]: I0227 09:12:22.817044 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-8drht_8fe63fdd-8b6f-4ecf-81d2-959ae060d567/frr-metrics/0.log" Feb 27 09:12:22 crc kubenswrapper[4612]: I0227 09:12:22.824674 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-8drht_8fe63fdd-8b6f-4ecf-81d2-959ae060d567/kube-rbac-proxy-frr/0.log" Feb 27 09:12:23 crc kubenswrapper[4612]: I0227 09:12:23.049679 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-8drht_8fe63fdd-8b6f-4ecf-81d2-959ae060d567/reloader/0.log" Feb 27 09:12:23 crc kubenswrapper[4612]: I0227 09:12:23.227226 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-7f989f654f-x2k2n_161b64e1-f414-4519-9fc1-c564d1cc0316/frr-k8s-webhook-server/0.log" Feb 27 09:12:23 crc kubenswrapper[4612]: I0227 09:12:23.330593 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-78d5b849b-c2rb5_29cfdfc3-d4a9-44df-b46e-638a9707a9f1/manager/0.log" Feb 27 09:12:23 crc kubenswrapper[4612]: I0227 09:12:23.581897 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-6f9ff8875f-jphph_62df4a96-aac6-4169-bbc4-d41ea85b8722/webhook-server/0.log" Feb 27 09:12:23 crc kubenswrapper[4612]: I0227 09:12:23.855616 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-zlmtg_9b1f5f78-3c58-42fb-9a8f-0166eab4f500/kube-rbac-proxy/0.log" Feb 27 09:12:24 crc kubenswrapper[4612]: I0227 09:12:24.355668 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-zlmtg_9b1f5f78-3c58-42fb-9a8f-0166eab4f500/speaker/0.log" Feb 27 09:12:24 crc kubenswrapper[4612]: I0227 09:12:24.490131 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-8drht_8fe63fdd-8b6f-4ecf-81d2-959ae060d567/frr/0.log" Feb 27 09:12:39 crc kubenswrapper[4612]: I0227 09:12:39.403331 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82g84wq_d7ee024d-7259-4702-bc99-fb938a71345f/util/0.log" Feb 27 09:12:39 crc kubenswrapper[4612]: I0227 09:12:39.893418 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82g84wq_d7ee024d-7259-4702-bc99-fb938a71345f/util/0.log" Feb 27 09:12:39 crc kubenswrapper[4612]: I0227 09:12:39.942534 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82g84wq_d7ee024d-7259-4702-bc99-fb938a71345f/pull/0.log" Feb 27 09:12:39 crc kubenswrapper[4612]: I0227 09:12:39.988037 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82g84wq_d7ee024d-7259-4702-bc99-fb938a71345f/pull/0.log" Feb 27 09:12:40 crc kubenswrapper[4612]: I0227 09:12:40.114534 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82g84wq_d7ee024d-7259-4702-bc99-fb938a71345f/util/0.log" Feb 27 09:12:40 crc kubenswrapper[4612]: I0227 09:12:40.183052 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82g84wq_d7ee024d-7259-4702-bc99-fb938a71345f/pull/0.log" Feb 27 09:12:40 crc kubenswrapper[4612]: I0227 09:12:40.237125 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_0e94e7566f739476ccec6d16e58de3f1c434cfa3060893f90f3e473a82g84wq_d7ee024d-7259-4702-bc99-fb938a71345f/extract/0.log" Feb 27 09:12:40 crc kubenswrapper[4612]: I0227 09:12:40.366013 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-ln97g_4a605168-76bd-4605-9a60-5ea574a0b2c1/extract-utilities/0.log" Feb 27 09:12:40 crc kubenswrapper[4612]: I0227 09:12:40.623968 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-ln97g_4a605168-76bd-4605-9a60-5ea574a0b2c1/extract-content/0.log" Feb 27 09:12:40 crc kubenswrapper[4612]: I0227 09:12:40.637009 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-ln97g_4a605168-76bd-4605-9a60-5ea574a0b2c1/extract-utilities/0.log" Feb 27 09:12:40 crc kubenswrapper[4612]: I0227 09:12:40.637390 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-ln97g_4a605168-76bd-4605-9a60-5ea574a0b2c1/extract-content/0.log" Feb 27 09:12:40 crc kubenswrapper[4612]: I0227 09:12:40.841360 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-ln97g_4a605168-76bd-4605-9a60-5ea574a0b2c1/extract-content/0.log" Feb 27 09:12:40 crc kubenswrapper[4612]: I0227 09:12:40.844673 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-ln97g_4a605168-76bd-4605-9a60-5ea574a0b2c1/extract-utilities/0.log" Feb 27 09:12:41 crc kubenswrapper[4612]: I0227 09:12:41.087259 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-ck8cc_f2b71124-3e61-48fc-a493-60c759124936/extract-utilities/0.log" Feb 27 09:12:41 crc kubenswrapper[4612]: I0227 09:12:41.470332 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-ln97g_4a605168-76bd-4605-9a60-5ea574a0b2c1/registry-server/0.log" Feb 27 09:12:41 crc kubenswrapper[4612]: I0227 09:12:41.493269 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-ck8cc_f2b71124-3e61-48fc-a493-60c759124936/extract-content/0.log" Feb 27 09:12:41 crc kubenswrapper[4612]: I0227 09:12:41.525314 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-ck8cc_f2b71124-3e61-48fc-a493-60c759124936/extract-content/0.log" Feb 27 09:12:41 crc kubenswrapper[4612]: I0227 09:12:41.525937 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-ck8cc_f2b71124-3e61-48fc-a493-60c759124936/extract-utilities/0.log" Feb 27 09:12:41 crc kubenswrapper[4612]: I0227 09:12:41.700031 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-ck8cc_f2b71124-3e61-48fc-a493-60c759124936/extract-utilities/0.log" Feb 27 09:12:41 crc kubenswrapper[4612]: I0227 09:12:41.743985 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-ck8cc_f2b71124-3e61-48fc-a493-60c759124936/extract-content/0.log" Feb 27 09:12:42 crc kubenswrapper[4612]: I0227 09:12:42.098254 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4nq7w2_f3841018-8eb3-4d1d-b78a-c023bbb050cd/util/0.log" Feb 27 09:12:42 crc kubenswrapper[4612]: I0227 09:12:42.364658 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-ck8cc_f2b71124-3e61-48fc-a493-60c759124936/registry-server/0.log" Feb 27 09:12:42 crc kubenswrapper[4612]: I0227 09:12:42.409199 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4nq7w2_f3841018-8eb3-4d1d-b78a-c023bbb050cd/pull/0.log" Feb 27 09:12:42 crc kubenswrapper[4612]: I0227 09:12:42.496710 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4nq7w2_f3841018-8eb3-4d1d-b78a-c023bbb050cd/util/0.log" Feb 27 09:12:42 crc kubenswrapper[4612]: I0227 09:12:42.499820 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4nq7w2_f3841018-8eb3-4d1d-b78a-c023bbb050cd/pull/0.log" Feb 27 09:12:42 crc kubenswrapper[4612]: I0227 09:12:42.677852 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4nq7w2_f3841018-8eb3-4d1d-b78a-c023bbb050cd/util/0.log" Feb 27 09:12:42 crc kubenswrapper[4612]: I0227 09:12:42.689200 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4nq7w2_f3841018-8eb3-4d1d-b78a-c023bbb050cd/extract/0.log" Feb 27 09:12:42 crc kubenswrapper[4612]: I0227 09:12:42.690367 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_d146760600e43041070ad4572d9c23f31a62e3aefc01a54998863bc5f4nq7w2_f3841018-8eb3-4d1d-b78a-c023bbb050cd/pull/0.log" Feb 27 09:12:42 crc kubenswrapper[4612]: I0227 09:12:42.899726 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-88rfh_6f0a0683-7e38-410b-8c31-137e9d3a841a/marketplace-operator/0.log" Feb 27 09:12:42 crc kubenswrapper[4612]: I0227 09:12:42.954022 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-8dbvm_eb785264-58b2-4d87-89f3-8dc86f784419/extract-utilities/0.log" Feb 27 09:12:43 crc kubenswrapper[4612]: I0227 09:12:43.164383 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-8dbvm_eb785264-58b2-4d87-89f3-8dc86f784419/extract-utilities/0.log" Feb 27 09:12:43 crc kubenswrapper[4612]: I0227 09:12:43.167226 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-8dbvm_eb785264-58b2-4d87-89f3-8dc86f784419/extract-content/0.log" Feb 27 09:12:43 crc kubenswrapper[4612]: I0227 09:12:43.204370 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-8dbvm_eb785264-58b2-4d87-89f3-8dc86f784419/extract-content/0.log" Feb 27 09:12:43 crc kubenswrapper[4612]: I0227 09:12:43.408667 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-8dbvm_eb785264-58b2-4d87-89f3-8dc86f784419/extract-utilities/0.log" Feb 27 09:12:43 crc kubenswrapper[4612]: I0227 09:12:43.435136 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-8dbvm_eb785264-58b2-4d87-89f3-8dc86f784419/extract-content/0.log" Feb 27 09:12:43 crc kubenswrapper[4612]: I0227 09:12:43.585391 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-8dbvm_eb785264-58b2-4d87-89f3-8dc86f784419/registry-server/0.log" Feb 27 09:12:43 crc kubenswrapper[4612]: I0227 09:12:43.639056 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-x56jg_cba13562-dcb0-4896-94eb-5d5246ffa1c2/extract-utilities/0.log" Feb 27 09:12:43 crc kubenswrapper[4612]: I0227 09:12:43.804017 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-x56jg_cba13562-dcb0-4896-94eb-5d5246ffa1c2/extract-utilities/0.log" Feb 27 09:12:43 crc kubenswrapper[4612]: I0227 09:12:43.804407 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-x56jg_cba13562-dcb0-4896-94eb-5d5246ffa1c2/extract-content/0.log" Feb 27 09:12:43 crc kubenswrapper[4612]: I0227 09:12:43.804440 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-x56jg_cba13562-dcb0-4896-94eb-5d5246ffa1c2/extract-content/0.log" Feb 27 09:12:43 crc kubenswrapper[4612]: I0227 09:12:43.971318 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-x56jg_cba13562-dcb0-4896-94eb-5d5246ffa1c2/extract-utilities/0.log" Feb 27 09:12:44 crc kubenswrapper[4612]: I0227 09:12:44.001280 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-x56jg_cba13562-dcb0-4896-94eb-5d5246ffa1c2/extract-content/0.log" Feb 27 09:12:44 crc kubenswrapper[4612]: I0227 09:12:44.540966 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-x56jg_cba13562-dcb0-4896-94eb-5d5246ffa1c2/registry-server/0.log" Feb 27 09:12:46 crc kubenswrapper[4612]: I0227 09:12:46.027091 4612 patch_prober.go:28] interesting pod/machine-config-daemon-924vb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 27 09:12:46 crc kubenswrapper[4612]: I0227 09:12:46.027378 4612 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 27 09:12:46 crc kubenswrapper[4612]: I0227 09:12:46.027445 4612 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-924vb" Feb 27 09:12:46 crc kubenswrapper[4612]: I0227 09:12:46.028285 4612 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"83e903dc65dc37650f4243bfdb54ee1e09257b50cf32c05bb8e521882531bc38"} pod="openshift-machine-config-operator/machine-config-daemon-924vb" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 27 09:12:46 crc kubenswrapper[4612]: I0227 09:12:46.028358 4612 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" containerName="machine-config-daemon" containerID="cri-o://83e903dc65dc37650f4243bfdb54ee1e09257b50cf32c05bb8e521882531bc38" gracePeriod=600 Feb 27 09:12:46 crc kubenswrapper[4612]: I0227 09:12:46.486485 4612 generic.go:334] "Generic (PLEG): container finished" podID="24599373-7adc-4a1b-8bb4-797bf726f43d" containerID="83e903dc65dc37650f4243bfdb54ee1e09257b50cf32c05bb8e521882531bc38" exitCode=0 Feb 27 09:12:46 crc kubenswrapper[4612]: I0227 09:12:46.486566 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-924vb" event={"ID":"24599373-7adc-4a1b-8bb4-797bf726f43d","Type":"ContainerDied","Data":"83e903dc65dc37650f4243bfdb54ee1e09257b50cf32c05bb8e521882531bc38"} Feb 27 09:12:46 crc kubenswrapper[4612]: I0227 09:12:46.486937 4612 scope.go:117] "RemoveContainer" containerID="c90278d5d8864d43287d88f2fcf48602a1065bde9d8b6d625ab9498edb9b58db" Feb 27 09:12:46 crc kubenswrapper[4612]: E0227 09:12:46.661139 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-924vb_openshift-machine-config-operator(24599373-7adc-4a1b-8bb4-797bf726f43d)\"" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" Feb 27 09:12:47 crc kubenswrapper[4612]: I0227 09:12:47.496668 4612 scope.go:117] "RemoveContainer" containerID="83e903dc65dc37650f4243bfdb54ee1e09257b50cf32c05bb8e521882531bc38" Feb 27 09:12:47 crc kubenswrapper[4612]: E0227 09:12:47.497179 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-924vb_openshift-machine-config-operator(24599373-7adc-4a1b-8bb4-797bf726f43d)\"" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" Feb 27 09:12:57 crc kubenswrapper[4612]: I0227 09:12:57.238927 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-bfsf2"] Feb 27 09:12:57 crc kubenswrapper[4612]: E0227 09:12:57.239813 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="087335eb-b232-4434-aca3-2ed95ec4350a" containerName="oc" Feb 27 09:12:57 crc kubenswrapper[4612]: I0227 09:12:57.239827 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="087335eb-b232-4434-aca3-2ed95ec4350a" containerName="oc" Feb 27 09:12:57 crc kubenswrapper[4612]: I0227 09:12:57.240008 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="087335eb-b232-4434-aca3-2ed95ec4350a" containerName="oc" Feb 27 09:12:57 crc kubenswrapper[4612]: I0227 09:12:57.241251 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-bfsf2" Feb 27 09:12:57 crc kubenswrapper[4612]: I0227 09:12:57.264630 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-bfsf2"] Feb 27 09:12:57 crc kubenswrapper[4612]: I0227 09:12:57.328021 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fb733a33-a2ba-462f-b033-ce0f924e51dd-catalog-content\") pod \"community-operators-bfsf2\" (UID: \"fb733a33-a2ba-462f-b033-ce0f924e51dd\") " pod="openshift-marketplace/community-operators-bfsf2" Feb 27 09:12:57 crc kubenswrapper[4612]: I0227 09:12:57.328070 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fb733a33-a2ba-462f-b033-ce0f924e51dd-utilities\") pod \"community-operators-bfsf2\" (UID: \"fb733a33-a2ba-462f-b033-ce0f924e51dd\") " pod="openshift-marketplace/community-operators-bfsf2" Feb 27 09:12:57 crc kubenswrapper[4612]: I0227 09:12:57.328096 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vnj6q\" (UniqueName: \"kubernetes.io/projected/fb733a33-a2ba-462f-b033-ce0f924e51dd-kube-api-access-vnj6q\") pod \"community-operators-bfsf2\" (UID: \"fb733a33-a2ba-462f-b033-ce0f924e51dd\") " pod="openshift-marketplace/community-operators-bfsf2" Feb 27 09:12:57 crc kubenswrapper[4612]: I0227 09:12:57.429990 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fb733a33-a2ba-462f-b033-ce0f924e51dd-catalog-content\") pod \"community-operators-bfsf2\" (UID: \"fb733a33-a2ba-462f-b033-ce0f924e51dd\") " pod="openshift-marketplace/community-operators-bfsf2" Feb 27 09:12:57 crc kubenswrapper[4612]: I0227 09:12:57.430358 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fb733a33-a2ba-462f-b033-ce0f924e51dd-utilities\") pod \"community-operators-bfsf2\" (UID: \"fb733a33-a2ba-462f-b033-ce0f924e51dd\") " pod="openshift-marketplace/community-operators-bfsf2" Feb 27 09:12:57 crc kubenswrapper[4612]: I0227 09:12:57.430384 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vnj6q\" (UniqueName: \"kubernetes.io/projected/fb733a33-a2ba-462f-b033-ce0f924e51dd-kube-api-access-vnj6q\") pod \"community-operators-bfsf2\" (UID: \"fb733a33-a2ba-462f-b033-ce0f924e51dd\") " pod="openshift-marketplace/community-operators-bfsf2" Feb 27 09:12:57 crc kubenswrapper[4612]: I0227 09:12:57.430507 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fb733a33-a2ba-462f-b033-ce0f924e51dd-catalog-content\") pod \"community-operators-bfsf2\" (UID: \"fb733a33-a2ba-462f-b033-ce0f924e51dd\") " pod="openshift-marketplace/community-operators-bfsf2" Feb 27 09:12:57 crc kubenswrapper[4612]: I0227 09:12:57.430763 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fb733a33-a2ba-462f-b033-ce0f924e51dd-utilities\") pod \"community-operators-bfsf2\" (UID: \"fb733a33-a2ba-462f-b033-ce0f924e51dd\") " pod="openshift-marketplace/community-operators-bfsf2" Feb 27 09:12:57 crc kubenswrapper[4612]: I0227 09:12:57.451429 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vnj6q\" (UniqueName: \"kubernetes.io/projected/fb733a33-a2ba-462f-b033-ce0f924e51dd-kube-api-access-vnj6q\") pod \"community-operators-bfsf2\" (UID: \"fb733a33-a2ba-462f-b033-ce0f924e51dd\") " pod="openshift-marketplace/community-operators-bfsf2" Feb 27 09:12:57 crc kubenswrapper[4612]: I0227 09:12:57.564837 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-bfsf2" Feb 27 09:12:58 crc kubenswrapper[4612]: W0227 09:12:58.172793 4612 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfb733a33_a2ba_462f_b033_ce0f924e51dd.slice/crio-3250f0f8088402d46d0e99a536a93f45e015c89a632166ea8f9f40ba26e71984 WatchSource:0}: Error finding container 3250f0f8088402d46d0e99a536a93f45e015c89a632166ea8f9f40ba26e71984: Status 404 returned error can't find the container with id 3250f0f8088402d46d0e99a536a93f45e015c89a632166ea8f9f40ba26e71984 Feb 27 09:12:58 crc kubenswrapper[4612]: I0227 09:12:58.172929 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-bfsf2"] Feb 27 09:12:58 crc kubenswrapper[4612]: I0227 09:12:58.580624 4612 generic.go:334] "Generic (PLEG): container finished" podID="fb733a33-a2ba-462f-b033-ce0f924e51dd" containerID="c8850408b90d3ccd253b6e4e3ab564197dfb20b51fdf672f7727c24c261b3467" exitCode=0 Feb 27 09:12:58 crc kubenswrapper[4612]: I0227 09:12:58.580718 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bfsf2" event={"ID":"fb733a33-a2ba-462f-b033-ce0f924e51dd","Type":"ContainerDied","Data":"c8850408b90d3ccd253b6e4e3ab564197dfb20b51fdf672f7727c24c261b3467"} Feb 27 09:12:58 crc kubenswrapper[4612]: I0227 09:12:58.580960 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bfsf2" event={"ID":"fb733a33-a2ba-462f-b033-ce0f924e51dd","Type":"ContainerStarted","Data":"3250f0f8088402d46d0e99a536a93f45e015c89a632166ea8f9f40ba26e71984"} Feb 27 09:13:00 crc kubenswrapper[4612]: I0227 09:13:00.600859 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bfsf2" event={"ID":"fb733a33-a2ba-462f-b033-ce0f924e51dd","Type":"ContainerStarted","Data":"c00d2b470830f6a47a2fbd5bd7698ad57e098776be5aa60cbb64779247a3695d"} Feb 27 09:13:00 crc kubenswrapper[4612]: I0227 09:13:00.853513 4612 scope.go:117] "RemoveContainer" containerID="83e903dc65dc37650f4243bfdb54ee1e09257b50cf32c05bb8e521882531bc38" Feb 27 09:13:00 crc kubenswrapper[4612]: E0227 09:13:00.853799 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-924vb_openshift-machine-config-operator(24599373-7adc-4a1b-8bb4-797bf726f43d)\"" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" Feb 27 09:13:01 crc kubenswrapper[4612]: I0227 09:13:01.615396 4612 generic.go:334] "Generic (PLEG): container finished" podID="fb733a33-a2ba-462f-b033-ce0f924e51dd" containerID="c00d2b470830f6a47a2fbd5bd7698ad57e098776be5aa60cbb64779247a3695d" exitCode=0 Feb 27 09:13:01 crc kubenswrapper[4612]: I0227 09:13:01.615462 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bfsf2" event={"ID":"fb733a33-a2ba-462f-b033-ce0f924e51dd","Type":"ContainerDied","Data":"c00d2b470830f6a47a2fbd5bd7698ad57e098776be5aa60cbb64779247a3695d"} Feb 27 09:13:02 crc kubenswrapper[4612]: I0227 09:13:02.636558 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bfsf2" event={"ID":"fb733a33-a2ba-462f-b033-ce0f924e51dd","Type":"ContainerStarted","Data":"ab021344cb3073be29c173f0be436cc821e67d35b49f25d61d53db646201d5f1"} Feb 27 09:13:07 crc kubenswrapper[4612]: I0227 09:13:07.565301 4612 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-bfsf2" Feb 27 09:13:07 crc kubenswrapper[4612]: I0227 09:13:07.565886 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-bfsf2" Feb 27 09:13:08 crc kubenswrapper[4612]: I0227 09:13:08.634609 4612 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-bfsf2" podUID="fb733a33-a2ba-462f-b033-ce0f924e51dd" containerName="registry-server" probeResult="failure" output=< Feb 27 09:13:08 crc kubenswrapper[4612]: timeout: failed to connect service ":50051" within 1s Feb 27 09:13:08 crc kubenswrapper[4612]: > Feb 27 09:13:15 crc kubenswrapper[4612]: I0227 09:13:15.853865 4612 scope.go:117] "RemoveContainer" containerID="83e903dc65dc37650f4243bfdb54ee1e09257b50cf32c05bb8e521882531bc38" Feb 27 09:13:15 crc kubenswrapper[4612]: E0227 09:13:15.854624 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-924vb_openshift-machine-config-operator(24599373-7adc-4a1b-8bb4-797bf726f43d)\"" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" Feb 27 09:13:17 crc kubenswrapper[4612]: I0227 09:13:17.817675 4612 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-bfsf2" Feb 27 09:13:17 crc kubenswrapper[4612]: I0227 09:13:17.845055 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-bfsf2" podStartSLOduration=17.355756382 podStartE2EDuration="20.845041423s" podCreationTimestamp="2026-02-27 09:12:57 +0000 UTC" firstStartedPulling="2026-02-27 09:12:58.5838961 +0000 UTC m=+5036.437826098" lastFinishedPulling="2026-02-27 09:13:02.073181151 +0000 UTC m=+5039.927111139" observedRunningTime="2026-02-27 09:13:02.678663477 +0000 UTC m=+5040.532593475" watchObservedRunningTime="2026-02-27 09:13:17.845041423 +0000 UTC m=+5055.698971421" Feb 27 09:13:17 crc kubenswrapper[4612]: I0227 09:13:17.926277 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-bfsf2" Feb 27 09:13:18 crc kubenswrapper[4612]: I0227 09:13:18.061682 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-bfsf2"] Feb 27 09:13:19 crc kubenswrapper[4612]: I0227 09:13:19.834622 4612 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-bfsf2" podUID="fb733a33-a2ba-462f-b033-ce0f924e51dd" containerName="registry-server" containerID="cri-o://ab021344cb3073be29c173f0be436cc821e67d35b49f25d61d53db646201d5f1" gracePeriod=2 Feb 27 09:13:20 crc kubenswrapper[4612]: I0227 09:13:20.408635 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-bfsf2" Feb 27 09:13:20 crc kubenswrapper[4612]: I0227 09:13:20.598628 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vnj6q\" (UniqueName: \"kubernetes.io/projected/fb733a33-a2ba-462f-b033-ce0f924e51dd-kube-api-access-vnj6q\") pod \"fb733a33-a2ba-462f-b033-ce0f924e51dd\" (UID: \"fb733a33-a2ba-462f-b033-ce0f924e51dd\") " Feb 27 09:13:20 crc kubenswrapper[4612]: I0227 09:13:20.598846 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fb733a33-a2ba-462f-b033-ce0f924e51dd-catalog-content\") pod \"fb733a33-a2ba-462f-b033-ce0f924e51dd\" (UID: \"fb733a33-a2ba-462f-b033-ce0f924e51dd\") " Feb 27 09:13:20 crc kubenswrapper[4612]: I0227 09:13:20.598915 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fb733a33-a2ba-462f-b033-ce0f924e51dd-utilities\") pod \"fb733a33-a2ba-462f-b033-ce0f924e51dd\" (UID: \"fb733a33-a2ba-462f-b033-ce0f924e51dd\") " Feb 27 09:13:20 crc kubenswrapper[4612]: I0227 09:13:20.600071 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fb733a33-a2ba-462f-b033-ce0f924e51dd-utilities" (OuterVolumeSpecName: "utilities") pod "fb733a33-a2ba-462f-b033-ce0f924e51dd" (UID: "fb733a33-a2ba-462f-b033-ce0f924e51dd"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 09:13:20 crc kubenswrapper[4612]: I0227 09:13:20.607069 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fb733a33-a2ba-462f-b033-ce0f924e51dd-kube-api-access-vnj6q" (OuterVolumeSpecName: "kube-api-access-vnj6q") pod "fb733a33-a2ba-462f-b033-ce0f924e51dd" (UID: "fb733a33-a2ba-462f-b033-ce0f924e51dd"). InnerVolumeSpecName "kube-api-access-vnj6q". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 09:13:20 crc kubenswrapper[4612]: I0227 09:13:20.652530 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fb733a33-a2ba-462f-b033-ce0f924e51dd-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "fb733a33-a2ba-462f-b033-ce0f924e51dd" (UID: "fb733a33-a2ba-462f-b033-ce0f924e51dd"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 09:13:20 crc kubenswrapper[4612]: I0227 09:13:20.701364 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vnj6q\" (UniqueName: \"kubernetes.io/projected/fb733a33-a2ba-462f-b033-ce0f924e51dd-kube-api-access-vnj6q\") on node \"crc\" DevicePath \"\"" Feb 27 09:13:20 crc kubenswrapper[4612]: I0227 09:13:20.701405 4612 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fb733a33-a2ba-462f-b033-ce0f924e51dd-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 27 09:13:20 crc kubenswrapper[4612]: I0227 09:13:20.701417 4612 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fb733a33-a2ba-462f-b033-ce0f924e51dd-utilities\") on node \"crc\" DevicePath \"\"" Feb 27 09:13:20 crc kubenswrapper[4612]: I0227 09:13:20.843500 4612 generic.go:334] "Generic (PLEG): container finished" podID="fb733a33-a2ba-462f-b033-ce0f924e51dd" containerID="ab021344cb3073be29c173f0be436cc821e67d35b49f25d61d53db646201d5f1" exitCode=0 Feb 27 09:13:20 crc kubenswrapper[4612]: I0227 09:13:20.843543 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bfsf2" event={"ID":"fb733a33-a2ba-462f-b033-ce0f924e51dd","Type":"ContainerDied","Data":"ab021344cb3073be29c173f0be436cc821e67d35b49f25d61d53db646201d5f1"} Feb 27 09:13:20 crc kubenswrapper[4612]: I0227 09:13:20.843604 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-bfsf2" Feb 27 09:13:20 crc kubenswrapper[4612]: I0227 09:13:20.843804 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bfsf2" event={"ID":"fb733a33-a2ba-462f-b033-ce0f924e51dd","Type":"ContainerDied","Data":"3250f0f8088402d46d0e99a536a93f45e015c89a632166ea8f9f40ba26e71984"} Feb 27 09:13:20 crc kubenswrapper[4612]: I0227 09:13:20.843827 4612 scope.go:117] "RemoveContainer" containerID="ab021344cb3073be29c173f0be436cc821e67d35b49f25d61d53db646201d5f1" Feb 27 09:13:20 crc kubenswrapper[4612]: I0227 09:13:20.864276 4612 scope.go:117] "RemoveContainer" containerID="c00d2b470830f6a47a2fbd5bd7698ad57e098776be5aa60cbb64779247a3695d" Feb 27 09:13:20 crc kubenswrapper[4612]: I0227 09:13:20.882964 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-bfsf2"] Feb 27 09:13:20 crc kubenswrapper[4612]: I0227 09:13:20.892728 4612 scope.go:117] "RemoveContainer" containerID="c8850408b90d3ccd253b6e4e3ab564197dfb20b51fdf672f7727c24c261b3467" Feb 27 09:13:20 crc kubenswrapper[4612]: I0227 09:13:20.896396 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-bfsf2"] Feb 27 09:13:20 crc kubenswrapper[4612]: I0227 09:13:20.934000 4612 scope.go:117] "RemoveContainer" containerID="ab021344cb3073be29c173f0be436cc821e67d35b49f25d61d53db646201d5f1" Feb 27 09:13:20 crc kubenswrapper[4612]: E0227 09:13:20.934382 4612 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ab021344cb3073be29c173f0be436cc821e67d35b49f25d61d53db646201d5f1\": container with ID starting with ab021344cb3073be29c173f0be436cc821e67d35b49f25d61d53db646201d5f1 not found: ID does not exist" containerID="ab021344cb3073be29c173f0be436cc821e67d35b49f25d61d53db646201d5f1" Feb 27 09:13:20 crc kubenswrapper[4612]: I0227 09:13:20.934411 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ab021344cb3073be29c173f0be436cc821e67d35b49f25d61d53db646201d5f1"} err="failed to get container status \"ab021344cb3073be29c173f0be436cc821e67d35b49f25d61d53db646201d5f1\": rpc error: code = NotFound desc = could not find container \"ab021344cb3073be29c173f0be436cc821e67d35b49f25d61d53db646201d5f1\": container with ID starting with ab021344cb3073be29c173f0be436cc821e67d35b49f25d61d53db646201d5f1 not found: ID does not exist" Feb 27 09:13:20 crc kubenswrapper[4612]: I0227 09:13:20.934430 4612 scope.go:117] "RemoveContainer" containerID="c00d2b470830f6a47a2fbd5bd7698ad57e098776be5aa60cbb64779247a3695d" Feb 27 09:13:20 crc kubenswrapper[4612]: E0227 09:13:20.935048 4612 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c00d2b470830f6a47a2fbd5bd7698ad57e098776be5aa60cbb64779247a3695d\": container with ID starting with c00d2b470830f6a47a2fbd5bd7698ad57e098776be5aa60cbb64779247a3695d not found: ID does not exist" containerID="c00d2b470830f6a47a2fbd5bd7698ad57e098776be5aa60cbb64779247a3695d" Feb 27 09:13:20 crc kubenswrapper[4612]: I0227 09:13:20.935068 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c00d2b470830f6a47a2fbd5bd7698ad57e098776be5aa60cbb64779247a3695d"} err="failed to get container status \"c00d2b470830f6a47a2fbd5bd7698ad57e098776be5aa60cbb64779247a3695d\": rpc error: code = NotFound desc = could not find container \"c00d2b470830f6a47a2fbd5bd7698ad57e098776be5aa60cbb64779247a3695d\": container with ID starting with c00d2b470830f6a47a2fbd5bd7698ad57e098776be5aa60cbb64779247a3695d not found: ID does not exist" Feb 27 09:13:20 crc kubenswrapper[4612]: I0227 09:13:20.935083 4612 scope.go:117] "RemoveContainer" containerID="c8850408b90d3ccd253b6e4e3ab564197dfb20b51fdf672f7727c24c261b3467" Feb 27 09:13:20 crc kubenswrapper[4612]: E0227 09:13:20.935237 4612 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c8850408b90d3ccd253b6e4e3ab564197dfb20b51fdf672f7727c24c261b3467\": container with ID starting with c8850408b90d3ccd253b6e4e3ab564197dfb20b51fdf672f7727c24c261b3467 not found: ID does not exist" containerID="c8850408b90d3ccd253b6e4e3ab564197dfb20b51fdf672f7727c24c261b3467" Feb 27 09:13:20 crc kubenswrapper[4612]: I0227 09:13:20.935253 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c8850408b90d3ccd253b6e4e3ab564197dfb20b51fdf672f7727c24c261b3467"} err="failed to get container status \"c8850408b90d3ccd253b6e4e3ab564197dfb20b51fdf672f7727c24c261b3467\": rpc error: code = NotFound desc = could not find container \"c8850408b90d3ccd253b6e4e3ab564197dfb20b51fdf672f7727c24c261b3467\": container with ID starting with c8850408b90d3ccd253b6e4e3ab564197dfb20b51fdf672f7727c24c261b3467 not found: ID does not exist" Feb 27 09:13:22 crc kubenswrapper[4612]: I0227 09:13:22.893879 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fb733a33-a2ba-462f-b033-ce0f924e51dd" path="/var/lib/kubelet/pods/fb733a33-a2ba-462f-b033-ce0f924e51dd/volumes" Feb 27 09:13:27 crc kubenswrapper[4612]: I0227 09:13:27.852658 4612 scope.go:117] "RemoveContainer" containerID="83e903dc65dc37650f4243bfdb54ee1e09257b50cf32c05bb8e521882531bc38" Feb 27 09:13:27 crc kubenswrapper[4612]: E0227 09:13:27.853527 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-924vb_openshift-machine-config-operator(24599373-7adc-4a1b-8bb4-797bf726f43d)\"" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" Feb 27 09:13:42 crc kubenswrapper[4612]: I0227 09:13:42.859736 4612 scope.go:117] "RemoveContainer" containerID="83e903dc65dc37650f4243bfdb54ee1e09257b50cf32c05bb8e521882531bc38" Feb 27 09:13:42 crc kubenswrapper[4612]: E0227 09:13:42.860533 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-924vb_openshift-machine-config-operator(24599373-7adc-4a1b-8bb4-797bf726f43d)\"" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" Feb 27 09:13:56 crc kubenswrapper[4612]: I0227 09:13:56.855651 4612 scope.go:117] "RemoveContainer" containerID="83e903dc65dc37650f4243bfdb54ee1e09257b50cf32c05bb8e521882531bc38" Feb 27 09:13:56 crc kubenswrapper[4612]: E0227 09:13:56.857485 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-924vb_openshift-machine-config-operator(24599373-7adc-4a1b-8bb4-797bf726f43d)\"" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" Feb 27 09:14:00 crc kubenswrapper[4612]: I0227 09:14:00.171182 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29536394-flf4x"] Feb 27 09:14:00 crc kubenswrapper[4612]: E0227 09:14:00.172223 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fb733a33-a2ba-462f-b033-ce0f924e51dd" containerName="extract-content" Feb 27 09:14:00 crc kubenswrapper[4612]: I0227 09:14:00.172241 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="fb733a33-a2ba-462f-b033-ce0f924e51dd" containerName="extract-content" Feb 27 09:14:00 crc kubenswrapper[4612]: E0227 09:14:00.172253 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fb733a33-a2ba-462f-b033-ce0f924e51dd" containerName="registry-server" Feb 27 09:14:00 crc kubenswrapper[4612]: I0227 09:14:00.172261 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="fb733a33-a2ba-462f-b033-ce0f924e51dd" containerName="registry-server" Feb 27 09:14:00 crc kubenswrapper[4612]: E0227 09:14:00.172276 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fb733a33-a2ba-462f-b033-ce0f924e51dd" containerName="extract-utilities" Feb 27 09:14:00 crc kubenswrapper[4612]: I0227 09:14:00.172285 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="fb733a33-a2ba-462f-b033-ce0f924e51dd" containerName="extract-utilities" Feb 27 09:14:00 crc kubenswrapper[4612]: I0227 09:14:00.172535 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="fb733a33-a2ba-462f-b033-ce0f924e51dd" containerName="registry-server" Feb 27 09:14:00 crc kubenswrapper[4612]: I0227 09:14:00.173282 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536394-flf4x" Feb 27 09:14:00 crc kubenswrapper[4612]: I0227 09:14:00.182046 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 27 09:14:00 crc kubenswrapper[4612]: I0227 09:14:00.185051 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536394-flf4x"] Feb 27 09:14:00 crc kubenswrapper[4612]: I0227 09:14:00.185081 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 27 09:14:00 crc kubenswrapper[4612]: I0227 09:14:00.189443 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-5zwfx" Feb 27 09:14:00 crc kubenswrapper[4612]: I0227 09:14:00.288026 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xrjb8\" (UniqueName: \"kubernetes.io/projected/50557bdd-3728-4980-a8a5-634120f56f03-kube-api-access-xrjb8\") pod \"auto-csr-approver-29536394-flf4x\" (UID: \"50557bdd-3728-4980-a8a5-634120f56f03\") " pod="openshift-infra/auto-csr-approver-29536394-flf4x" Feb 27 09:14:00 crc kubenswrapper[4612]: I0227 09:14:00.389456 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xrjb8\" (UniqueName: \"kubernetes.io/projected/50557bdd-3728-4980-a8a5-634120f56f03-kube-api-access-xrjb8\") pod \"auto-csr-approver-29536394-flf4x\" (UID: \"50557bdd-3728-4980-a8a5-634120f56f03\") " pod="openshift-infra/auto-csr-approver-29536394-flf4x" Feb 27 09:14:00 crc kubenswrapper[4612]: I0227 09:14:00.419044 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xrjb8\" (UniqueName: \"kubernetes.io/projected/50557bdd-3728-4980-a8a5-634120f56f03-kube-api-access-xrjb8\") pod \"auto-csr-approver-29536394-flf4x\" (UID: \"50557bdd-3728-4980-a8a5-634120f56f03\") " pod="openshift-infra/auto-csr-approver-29536394-flf4x" Feb 27 09:14:00 crc kubenswrapper[4612]: I0227 09:14:00.499956 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536394-flf4x" Feb 27 09:14:01 crc kubenswrapper[4612]: I0227 09:14:01.050334 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536394-flf4x"] Feb 27 09:14:01 crc kubenswrapper[4612]: I0227 09:14:01.264655 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536394-flf4x" event={"ID":"50557bdd-3728-4980-a8a5-634120f56f03","Type":"ContainerStarted","Data":"9f74e91f487cb79ce587ed82bf127b2839fd662312ea6fcf89bfccf838e5ac36"} Feb 27 09:14:03 crc kubenswrapper[4612]: I0227 09:14:03.288048 4612 generic.go:334] "Generic (PLEG): container finished" podID="50557bdd-3728-4980-a8a5-634120f56f03" containerID="3ff85774ad141b8f162606588670143569b57747ddf4e2a415cb206395b98d04" exitCode=0 Feb 27 09:14:03 crc kubenswrapper[4612]: I0227 09:14:03.288228 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536394-flf4x" event={"ID":"50557bdd-3728-4980-a8a5-634120f56f03","Type":"ContainerDied","Data":"3ff85774ad141b8f162606588670143569b57747ddf4e2a415cb206395b98d04"} Feb 27 09:14:04 crc kubenswrapper[4612]: I0227 09:14:04.668709 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536394-flf4x" Feb 27 09:14:04 crc kubenswrapper[4612]: I0227 09:14:04.791838 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xrjb8\" (UniqueName: \"kubernetes.io/projected/50557bdd-3728-4980-a8a5-634120f56f03-kube-api-access-xrjb8\") pod \"50557bdd-3728-4980-a8a5-634120f56f03\" (UID: \"50557bdd-3728-4980-a8a5-634120f56f03\") " Feb 27 09:14:04 crc kubenswrapper[4612]: I0227 09:14:04.799834 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/50557bdd-3728-4980-a8a5-634120f56f03-kube-api-access-xrjb8" (OuterVolumeSpecName: "kube-api-access-xrjb8") pod "50557bdd-3728-4980-a8a5-634120f56f03" (UID: "50557bdd-3728-4980-a8a5-634120f56f03"). InnerVolumeSpecName "kube-api-access-xrjb8". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 09:14:04 crc kubenswrapper[4612]: I0227 09:14:04.894047 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xrjb8\" (UniqueName: \"kubernetes.io/projected/50557bdd-3728-4980-a8a5-634120f56f03-kube-api-access-xrjb8\") on node \"crc\" DevicePath \"\"" Feb 27 09:14:05 crc kubenswrapper[4612]: I0227 09:14:05.306254 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536394-flf4x" event={"ID":"50557bdd-3728-4980-a8a5-634120f56f03","Type":"ContainerDied","Data":"9f74e91f487cb79ce587ed82bf127b2839fd662312ea6fcf89bfccf838e5ac36"} Feb 27 09:14:05 crc kubenswrapper[4612]: I0227 09:14:05.306512 4612 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9f74e91f487cb79ce587ed82bf127b2839fd662312ea6fcf89bfccf838e5ac36" Feb 27 09:14:05 crc kubenswrapper[4612]: I0227 09:14:05.306565 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536394-flf4x" Feb 27 09:14:05 crc kubenswrapper[4612]: I0227 09:14:05.743362 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29536388-7rftd"] Feb 27 09:14:05 crc kubenswrapper[4612]: I0227 09:14:05.753164 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29536388-7rftd"] Feb 27 09:14:06 crc kubenswrapper[4612]: I0227 09:14:06.865199 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ca8320dc-a538-4330-936e-74e7fe62dd30" path="/var/lib/kubelet/pods/ca8320dc-a538-4330-936e-74e7fe62dd30/volumes" Feb 27 09:14:07 crc kubenswrapper[4612]: I0227 09:14:07.852570 4612 scope.go:117] "RemoveContainer" containerID="83e903dc65dc37650f4243bfdb54ee1e09257b50cf32c05bb8e521882531bc38" Feb 27 09:14:07 crc kubenswrapper[4612]: E0227 09:14:07.853095 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-924vb_openshift-machine-config-operator(24599373-7adc-4a1b-8bb4-797bf726f43d)\"" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" Feb 27 09:14:21 crc kubenswrapper[4612]: I0227 09:14:21.322763 4612 scope.go:117] "RemoveContainer" containerID="7615e9a05300b174510034ccbceefb060369b203b78547262124d37e85cc2a88" Feb 27 09:14:21 crc kubenswrapper[4612]: I0227 09:14:21.853404 4612 scope.go:117] "RemoveContainer" containerID="83e903dc65dc37650f4243bfdb54ee1e09257b50cf32c05bb8e521882531bc38" Feb 27 09:14:21 crc kubenswrapper[4612]: E0227 09:14:21.853780 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-924vb_openshift-machine-config-operator(24599373-7adc-4a1b-8bb4-797bf726f43d)\"" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" Feb 27 09:14:35 crc kubenswrapper[4612]: I0227 09:14:35.853602 4612 scope.go:117] "RemoveContainer" containerID="83e903dc65dc37650f4243bfdb54ee1e09257b50cf32c05bb8e521882531bc38" Feb 27 09:14:35 crc kubenswrapper[4612]: E0227 09:14:35.854852 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-924vb_openshift-machine-config-operator(24599373-7adc-4a1b-8bb4-797bf726f43d)\"" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" Feb 27 09:14:48 crc kubenswrapper[4612]: I0227 09:14:48.853412 4612 scope.go:117] "RemoveContainer" containerID="83e903dc65dc37650f4243bfdb54ee1e09257b50cf32c05bb8e521882531bc38" Feb 27 09:14:48 crc kubenswrapper[4612]: E0227 09:14:48.854415 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-924vb_openshift-machine-config-operator(24599373-7adc-4a1b-8bb4-797bf726f43d)\"" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" Feb 27 09:15:00 crc kubenswrapper[4612]: I0227 09:15:00.166456 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29536395-zt8d4"] Feb 27 09:15:00 crc kubenswrapper[4612]: E0227 09:15:00.167509 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="50557bdd-3728-4980-a8a5-634120f56f03" containerName="oc" Feb 27 09:15:00 crc kubenswrapper[4612]: I0227 09:15:00.167525 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="50557bdd-3728-4980-a8a5-634120f56f03" containerName="oc" Feb 27 09:15:00 crc kubenswrapper[4612]: I0227 09:15:00.167827 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="50557bdd-3728-4980-a8a5-634120f56f03" containerName="oc" Feb 27 09:15:00 crc kubenswrapper[4612]: I0227 09:15:00.168614 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29536395-zt8d4" Feb 27 09:15:00 crc kubenswrapper[4612]: I0227 09:15:00.176866 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Feb 27 09:15:00 crc kubenswrapper[4612]: I0227 09:15:00.176932 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Feb 27 09:15:00 crc kubenswrapper[4612]: I0227 09:15:00.194481 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29536395-zt8d4"] Feb 27 09:15:00 crc kubenswrapper[4612]: I0227 09:15:00.285320 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zd2t9\" (UniqueName: \"kubernetes.io/projected/9bcab54a-cf20-4689-b98e-7e7221e0437d-kube-api-access-zd2t9\") pod \"collect-profiles-29536395-zt8d4\" (UID: \"9bcab54a-cf20-4689-b98e-7e7221e0437d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29536395-zt8d4" Feb 27 09:15:00 crc kubenswrapper[4612]: I0227 09:15:00.285376 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9bcab54a-cf20-4689-b98e-7e7221e0437d-secret-volume\") pod \"collect-profiles-29536395-zt8d4\" (UID: \"9bcab54a-cf20-4689-b98e-7e7221e0437d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29536395-zt8d4" Feb 27 09:15:00 crc kubenswrapper[4612]: I0227 09:15:00.285464 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9bcab54a-cf20-4689-b98e-7e7221e0437d-config-volume\") pod \"collect-profiles-29536395-zt8d4\" (UID: \"9bcab54a-cf20-4689-b98e-7e7221e0437d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29536395-zt8d4" Feb 27 09:15:00 crc kubenswrapper[4612]: I0227 09:15:00.387148 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9bcab54a-cf20-4689-b98e-7e7221e0437d-config-volume\") pod \"collect-profiles-29536395-zt8d4\" (UID: \"9bcab54a-cf20-4689-b98e-7e7221e0437d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29536395-zt8d4" Feb 27 09:15:00 crc kubenswrapper[4612]: I0227 09:15:00.387263 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zd2t9\" (UniqueName: \"kubernetes.io/projected/9bcab54a-cf20-4689-b98e-7e7221e0437d-kube-api-access-zd2t9\") pod \"collect-profiles-29536395-zt8d4\" (UID: \"9bcab54a-cf20-4689-b98e-7e7221e0437d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29536395-zt8d4" Feb 27 09:15:00 crc kubenswrapper[4612]: I0227 09:15:00.387305 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9bcab54a-cf20-4689-b98e-7e7221e0437d-secret-volume\") pod \"collect-profiles-29536395-zt8d4\" (UID: \"9bcab54a-cf20-4689-b98e-7e7221e0437d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29536395-zt8d4" Feb 27 09:15:00 crc kubenswrapper[4612]: I0227 09:15:00.388134 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9bcab54a-cf20-4689-b98e-7e7221e0437d-config-volume\") pod \"collect-profiles-29536395-zt8d4\" (UID: \"9bcab54a-cf20-4689-b98e-7e7221e0437d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29536395-zt8d4" Feb 27 09:15:00 crc kubenswrapper[4612]: I0227 09:15:00.397531 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9bcab54a-cf20-4689-b98e-7e7221e0437d-secret-volume\") pod \"collect-profiles-29536395-zt8d4\" (UID: \"9bcab54a-cf20-4689-b98e-7e7221e0437d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29536395-zt8d4" Feb 27 09:15:00 crc kubenswrapper[4612]: I0227 09:15:00.408360 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zd2t9\" (UniqueName: \"kubernetes.io/projected/9bcab54a-cf20-4689-b98e-7e7221e0437d-kube-api-access-zd2t9\") pod \"collect-profiles-29536395-zt8d4\" (UID: \"9bcab54a-cf20-4689-b98e-7e7221e0437d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29536395-zt8d4" Feb 27 09:15:00 crc kubenswrapper[4612]: I0227 09:15:00.496506 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29536395-zt8d4" Feb 27 09:15:00 crc kubenswrapper[4612]: I0227 09:15:00.988945 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29536395-zt8d4"] Feb 27 09:15:01 crc kubenswrapper[4612]: I0227 09:15:01.892019 4612 generic.go:334] "Generic (PLEG): container finished" podID="9bcab54a-cf20-4689-b98e-7e7221e0437d" containerID="3f04a48c57a99fec4fd7d0f41a901216cda988a119a5b912b66c5ef25c747889" exitCode=0 Feb 27 09:15:01 crc kubenswrapper[4612]: I0227 09:15:01.892069 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29536395-zt8d4" event={"ID":"9bcab54a-cf20-4689-b98e-7e7221e0437d","Type":"ContainerDied","Data":"3f04a48c57a99fec4fd7d0f41a901216cda988a119a5b912b66c5ef25c747889"} Feb 27 09:15:01 crc kubenswrapper[4612]: I0227 09:15:01.892423 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29536395-zt8d4" event={"ID":"9bcab54a-cf20-4689-b98e-7e7221e0437d","Type":"ContainerStarted","Data":"277039dea012d47178ca0778fb3254db44d9cc131354049e6fae4a32c00ed12d"} Feb 27 09:15:03 crc kubenswrapper[4612]: I0227 09:15:03.270653 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29536395-zt8d4" Feb 27 09:15:03 crc kubenswrapper[4612]: I0227 09:15:03.345175 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9bcab54a-cf20-4689-b98e-7e7221e0437d-secret-volume\") pod \"9bcab54a-cf20-4689-b98e-7e7221e0437d\" (UID: \"9bcab54a-cf20-4689-b98e-7e7221e0437d\") " Feb 27 09:15:03 crc kubenswrapper[4612]: I0227 09:15:03.345350 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zd2t9\" (UniqueName: \"kubernetes.io/projected/9bcab54a-cf20-4689-b98e-7e7221e0437d-kube-api-access-zd2t9\") pod \"9bcab54a-cf20-4689-b98e-7e7221e0437d\" (UID: \"9bcab54a-cf20-4689-b98e-7e7221e0437d\") " Feb 27 09:15:03 crc kubenswrapper[4612]: I0227 09:15:03.345421 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9bcab54a-cf20-4689-b98e-7e7221e0437d-config-volume\") pod \"9bcab54a-cf20-4689-b98e-7e7221e0437d\" (UID: \"9bcab54a-cf20-4689-b98e-7e7221e0437d\") " Feb 27 09:15:03 crc kubenswrapper[4612]: I0227 09:15:03.346710 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9bcab54a-cf20-4689-b98e-7e7221e0437d-config-volume" (OuterVolumeSpecName: "config-volume") pod "9bcab54a-cf20-4689-b98e-7e7221e0437d" (UID: "9bcab54a-cf20-4689-b98e-7e7221e0437d"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 27 09:15:03 crc kubenswrapper[4612]: I0227 09:15:03.351323 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9bcab54a-cf20-4689-b98e-7e7221e0437d-kube-api-access-zd2t9" (OuterVolumeSpecName: "kube-api-access-zd2t9") pod "9bcab54a-cf20-4689-b98e-7e7221e0437d" (UID: "9bcab54a-cf20-4689-b98e-7e7221e0437d"). InnerVolumeSpecName "kube-api-access-zd2t9". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 09:15:03 crc kubenswrapper[4612]: I0227 09:15:03.360766 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9bcab54a-cf20-4689-b98e-7e7221e0437d-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "9bcab54a-cf20-4689-b98e-7e7221e0437d" (UID: "9bcab54a-cf20-4689-b98e-7e7221e0437d"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 27 09:15:03 crc kubenswrapper[4612]: I0227 09:15:03.452186 4612 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9bcab54a-cf20-4689-b98e-7e7221e0437d-secret-volume\") on node \"crc\" DevicePath \"\"" Feb 27 09:15:03 crc kubenswrapper[4612]: I0227 09:15:03.452233 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zd2t9\" (UniqueName: \"kubernetes.io/projected/9bcab54a-cf20-4689-b98e-7e7221e0437d-kube-api-access-zd2t9\") on node \"crc\" DevicePath \"\"" Feb 27 09:15:03 crc kubenswrapper[4612]: I0227 09:15:03.452247 4612 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9bcab54a-cf20-4689-b98e-7e7221e0437d-config-volume\") on node \"crc\" DevicePath \"\"" Feb 27 09:15:03 crc kubenswrapper[4612]: I0227 09:15:03.853161 4612 scope.go:117] "RemoveContainer" containerID="83e903dc65dc37650f4243bfdb54ee1e09257b50cf32c05bb8e521882531bc38" Feb 27 09:15:03 crc kubenswrapper[4612]: E0227 09:15:03.853626 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-924vb_openshift-machine-config-operator(24599373-7adc-4a1b-8bb4-797bf726f43d)\"" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" Feb 27 09:15:03 crc kubenswrapper[4612]: I0227 09:15:03.917996 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29536395-zt8d4" event={"ID":"9bcab54a-cf20-4689-b98e-7e7221e0437d","Type":"ContainerDied","Data":"277039dea012d47178ca0778fb3254db44d9cc131354049e6fae4a32c00ed12d"} Feb 27 09:15:03 crc kubenswrapper[4612]: I0227 09:15:03.918051 4612 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="277039dea012d47178ca0778fb3254db44d9cc131354049e6fae4a32c00ed12d" Feb 27 09:15:03 crc kubenswrapper[4612]: I0227 09:15:03.918048 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29536395-zt8d4" Feb 27 09:15:04 crc kubenswrapper[4612]: I0227 09:15:04.361474 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29536350-4xptg"] Feb 27 09:15:04 crc kubenswrapper[4612]: I0227 09:15:04.372601 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29536350-4xptg"] Feb 27 09:15:04 crc kubenswrapper[4612]: I0227 09:15:04.865283 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ac60cd4-1e56-4ad4-8ba9-50bc36f8f211" path="/var/lib/kubelet/pods/6ac60cd4-1e56-4ad4-8ba9-50bc36f8f211/volumes" Feb 27 09:15:13 crc kubenswrapper[4612]: I0227 09:15:13.016935 4612 generic.go:334] "Generic (PLEG): container finished" podID="445b752f-7d64-4300-a5fe-bebc7d94f962" containerID="5758c84ae5ba6363334c142523e1a977fdcc41a6086265bba40ba4e563d7b5f4" exitCode=0 Feb 27 09:15:13 crc kubenswrapper[4612]: I0227 09:15:13.017423 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-gxfb5/must-gather-ksdms" event={"ID":"445b752f-7d64-4300-a5fe-bebc7d94f962","Type":"ContainerDied","Data":"5758c84ae5ba6363334c142523e1a977fdcc41a6086265bba40ba4e563d7b5f4"} Feb 27 09:15:13 crc kubenswrapper[4612]: I0227 09:15:13.018058 4612 scope.go:117] "RemoveContainer" containerID="5758c84ae5ba6363334c142523e1a977fdcc41a6086265bba40ba4e563d7b5f4" Feb 27 09:15:13 crc kubenswrapper[4612]: I0227 09:15:13.117359 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-gxfb5_must-gather-ksdms_445b752f-7d64-4300-a5fe-bebc7d94f962/gather/0.log" Feb 27 09:15:17 crc kubenswrapper[4612]: I0227 09:15:17.852712 4612 scope.go:117] "RemoveContainer" containerID="83e903dc65dc37650f4243bfdb54ee1e09257b50cf32c05bb8e521882531bc38" Feb 27 09:15:17 crc kubenswrapper[4612]: E0227 09:15:17.853471 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-924vb_openshift-machine-config-operator(24599373-7adc-4a1b-8bb4-797bf726f43d)\"" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" Feb 27 09:15:21 crc kubenswrapper[4612]: I0227 09:15:21.969379 4612 scope.go:117] "RemoveContainer" containerID="0a303417d47b93f56fe9e76626b77d3e695b9328364bdf2dfdd9ec678b311d79" Feb 27 09:15:21 crc kubenswrapper[4612]: I0227 09:15:21.998043 4612 scope.go:117] "RemoveContainer" containerID="41e4d2dd206394df03ffda8f01431a5bff3160fe5b41c3183e814bcd2d8c7d50" Feb 27 09:15:22 crc kubenswrapper[4612]: I0227 09:15:22.079727 4612 scope.go:117] "RemoveContainer" containerID="774af867d843dca9879978e4163a00adbc674a17692277e6f8632c4ce3c205c9" Feb 27 09:15:27 crc kubenswrapper[4612]: I0227 09:15:27.468205 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-gxfb5/must-gather-ksdms"] Feb 27 09:15:27 crc kubenswrapper[4612]: I0227 09:15:27.469048 4612 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-gxfb5/must-gather-ksdms" podUID="445b752f-7d64-4300-a5fe-bebc7d94f962" containerName="copy" containerID="cri-o://6f641cde04e671db752583be64a094e25103923c86300f24a4e12540dec044b3" gracePeriod=2 Feb 27 09:15:27 crc kubenswrapper[4612]: I0227 09:15:27.481136 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-gxfb5/must-gather-ksdms"] Feb 27 09:15:27 crc kubenswrapper[4612]: I0227 09:15:27.990468 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-gxfb5_must-gather-ksdms_445b752f-7d64-4300-a5fe-bebc7d94f962/copy/0.log" Feb 27 09:15:27 crc kubenswrapper[4612]: I0227 09:15:27.991182 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-gxfb5/must-gather-ksdms" Feb 27 09:15:28 crc kubenswrapper[4612]: I0227 09:15:28.117247 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-twk7b\" (UniqueName: \"kubernetes.io/projected/445b752f-7d64-4300-a5fe-bebc7d94f962-kube-api-access-twk7b\") pod \"445b752f-7d64-4300-a5fe-bebc7d94f962\" (UID: \"445b752f-7d64-4300-a5fe-bebc7d94f962\") " Feb 27 09:15:28 crc kubenswrapper[4612]: I0227 09:15:28.117416 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/445b752f-7d64-4300-a5fe-bebc7d94f962-must-gather-output\") pod \"445b752f-7d64-4300-a5fe-bebc7d94f962\" (UID: \"445b752f-7d64-4300-a5fe-bebc7d94f962\") " Feb 27 09:15:28 crc kubenswrapper[4612]: I0227 09:15:28.130222 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/445b752f-7d64-4300-a5fe-bebc7d94f962-kube-api-access-twk7b" (OuterVolumeSpecName: "kube-api-access-twk7b") pod "445b752f-7d64-4300-a5fe-bebc7d94f962" (UID: "445b752f-7d64-4300-a5fe-bebc7d94f962"). InnerVolumeSpecName "kube-api-access-twk7b". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 09:15:28 crc kubenswrapper[4612]: I0227 09:15:28.191684 4612 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-gxfb5_must-gather-ksdms_445b752f-7d64-4300-a5fe-bebc7d94f962/copy/0.log" Feb 27 09:15:28 crc kubenswrapper[4612]: I0227 09:15:28.192401 4612 generic.go:334] "Generic (PLEG): container finished" podID="445b752f-7d64-4300-a5fe-bebc7d94f962" containerID="6f641cde04e671db752583be64a094e25103923c86300f24a4e12540dec044b3" exitCode=143 Feb 27 09:15:28 crc kubenswrapper[4612]: I0227 09:15:28.192516 4612 scope.go:117] "RemoveContainer" containerID="6f641cde04e671db752583be64a094e25103923c86300f24a4e12540dec044b3" Feb 27 09:15:28 crc kubenswrapper[4612]: I0227 09:15:28.192480 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-gxfb5/must-gather-ksdms" Feb 27 09:15:28 crc kubenswrapper[4612]: I0227 09:15:28.220452 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-twk7b\" (UniqueName: \"kubernetes.io/projected/445b752f-7d64-4300-a5fe-bebc7d94f962-kube-api-access-twk7b\") on node \"crc\" DevicePath \"\"" Feb 27 09:15:28 crc kubenswrapper[4612]: I0227 09:15:28.232759 4612 scope.go:117] "RemoveContainer" containerID="5758c84ae5ba6363334c142523e1a977fdcc41a6086265bba40ba4e563d7b5f4" Feb 27 09:15:28 crc kubenswrapper[4612]: I0227 09:15:28.344436 4612 scope.go:117] "RemoveContainer" containerID="6f641cde04e671db752583be64a094e25103923c86300f24a4e12540dec044b3" Feb 27 09:15:28 crc kubenswrapper[4612]: E0227 09:15:28.346853 4612 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6f641cde04e671db752583be64a094e25103923c86300f24a4e12540dec044b3\": container with ID starting with 6f641cde04e671db752583be64a094e25103923c86300f24a4e12540dec044b3 not found: ID does not exist" containerID="6f641cde04e671db752583be64a094e25103923c86300f24a4e12540dec044b3" Feb 27 09:15:28 crc kubenswrapper[4612]: I0227 09:15:28.346901 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6f641cde04e671db752583be64a094e25103923c86300f24a4e12540dec044b3"} err="failed to get container status \"6f641cde04e671db752583be64a094e25103923c86300f24a4e12540dec044b3\": rpc error: code = NotFound desc = could not find container \"6f641cde04e671db752583be64a094e25103923c86300f24a4e12540dec044b3\": container with ID starting with 6f641cde04e671db752583be64a094e25103923c86300f24a4e12540dec044b3 not found: ID does not exist" Feb 27 09:15:28 crc kubenswrapper[4612]: I0227 09:15:28.346929 4612 scope.go:117] "RemoveContainer" containerID="5758c84ae5ba6363334c142523e1a977fdcc41a6086265bba40ba4e563d7b5f4" Feb 27 09:15:28 crc kubenswrapper[4612]: E0227 09:15:28.350858 4612 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5758c84ae5ba6363334c142523e1a977fdcc41a6086265bba40ba4e563d7b5f4\": container with ID starting with 5758c84ae5ba6363334c142523e1a977fdcc41a6086265bba40ba4e563d7b5f4 not found: ID does not exist" containerID="5758c84ae5ba6363334c142523e1a977fdcc41a6086265bba40ba4e563d7b5f4" Feb 27 09:15:28 crc kubenswrapper[4612]: I0227 09:15:28.350899 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5758c84ae5ba6363334c142523e1a977fdcc41a6086265bba40ba4e563d7b5f4"} err="failed to get container status \"5758c84ae5ba6363334c142523e1a977fdcc41a6086265bba40ba4e563d7b5f4\": rpc error: code = NotFound desc = could not find container \"5758c84ae5ba6363334c142523e1a977fdcc41a6086265bba40ba4e563d7b5f4\": container with ID starting with 5758c84ae5ba6363334c142523e1a977fdcc41a6086265bba40ba4e563d7b5f4 not found: ID does not exist" Feb 27 09:15:28 crc kubenswrapper[4612]: I0227 09:15:28.419981 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/445b752f-7d64-4300-a5fe-bebc7d94f962-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "445b752f-7d64-4300-a5fe-bebc7d94f962" (UID: "445b752f-7d64-4300-a5fe-bebc7d94f962"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 09:15:28 crc kubenswrapper[4612]: I0227 09:15:28.425675 4612 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/445b752f-7d64-4300-a5fe-bebc7d94f962-must-gather-output\") on node \"crc\" DevicePath \"\"" Feb 27 09:15:28 crc kubenswrapper[4612]: I0227 09:15:28.863327 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="445b752f-7d64-4300-a5fe-bebc7d94f962" path="/var/lib/kubelet/pods/445b752f-7d64-4300-a5fe-bebc7d94f962/volumes" Feb 27 09:15:29 crc kubenswrapper[4612]: I0227 09:15:29.852734 4612 scope.go:117] "RemoveContainer" containerID="83e903dc65dc37650f4243bfdb54ee1e09257b50cf32c05bb8e521882531bc38" Feb 27 09:15:29 crc kubenswrapper[4612]: E0227 09:15:29.853498 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-924vb_openshift-machine-config-operator(24599373-7adc-4a1b-8bb4-797bf726f43d)\"" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" Feb 27 09:15:44 crc kubenswrapper[4612]: I0227 09:15:44.853261 4612 scope.go:117] "RemoveContainer" containerID="83e903dc65dc37650f4243bfdb54ee1e09257b50cf32c05bb8e521882531bc38" Feb 27 09:15:44 crc kubenswrapper[4612]: E0227 09:15:44.854287 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-924vb_openshift-machine-config-operator(24599373-7adc-4a1b-8bb4-797bf726f43d)\"" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" Feb 27 09:15:56 crc kubenswrapper[4612]: I0227 09:15:56.852904 4612 scope.go:117] "RemoveContainer" containerID="83e903dc65dc37650f4243bfdb54ee1e09257b50cf32c05bb8e521882531bc38" Feb 27 09:15:56 crc kubenswrapper[4612]: E0227 09:15:56.854371 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-924vb_openshift-machine-config-operator(24599373-7adc-4a1b-8bb4-797bf726f43d)\"" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" Feb 27 09:16:00 crc kubenswrapper[4612]: I0227 09:16:00.160882 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29536396-btlx8"] Feb 27 09:16:00 crc kubenswrapper[4612]: E0227 09:16:00.162204 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="445b752f-7d64-4300-a5fe-bebc7d94f962" containerName="copy" Feb 27 09:16:00 crc kubenswrapper[4612]: I0227 09:16:00.162232 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="445b752f-7d64-4300-a5fe-bebc7d94f962" containerName="copy" Feb 27 09:16:00 crc kubenswrapper[4612]: E0227 09:16:00.162276 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="445b752f-7d64-4300-a5fe-bebc7d94f962" containerName="gather" Feb 27 09:16:00 crc kubenswrapper[4612]: I0227 09:16:00.162289 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="445b752f-7d64-4300-a5fe-bebc7d94f962" containerName="gather" Feb 27 09:16:00 crc kubenswrapper[4612]: E0227 09:16:00.162318 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9bcab54a-cf20-4689-b98e-7e7221e0437d" containerName="collect-profiles" Feb 27 09:16:00 crc kubenswrapper[4612]: I0227 09:16:00.162332 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="9bcab54a-cf20-4689-b98e-7e7221e0437d" containerName="collect-profiles" Feb 27 09:16:00 crc kubenswrapper[4612]: I0227 09:16:00.162675 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="445b752f-7d64-4300-a5fe-bebc7d94f962" containerName="copy" Feb 27 09:16:00 crc kubenswrapper[4612]: I0227 09:16:00.162763 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="9bcab54a-cf20-4689-b98e-7e7221e0437d" containerName="collect-profiles" Feb 27 09:16:00 crc kubenswrapper[4612]: I0227 09:16:00.162810 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="445b752f-7d64-4300-a5fe-bebc7d94f962" containerName="gather" Feb 27 09:16:00 crc kubenswrapper[4612]: I0227 09:16:00.164217 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536396-btlx8" Feb 27 09:16:00 crc kubenswrapper[4612]: I0227 09:16:00.166876 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 27 09:16:00 crc kubenswrapper[4612]: I0227 09:16:00.167035 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-5zwfx" Feb 27 09:16:00 crc kubenswrapper[4612]: I0227 09:16:00.167758 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 27 09:16:00 crc kubenswrapper[4612]: I0227 09:16:00.173749 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536396-btlx8"] Feb 27 09:16:00 crc kubenswrapper[4612]: I0227 09:16:00.191393 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-blvhd\" (UniqueName: \"kubernetes.io/projected/3e41030f-d2ed-4558-80e6-13d7d5365a43-kube-api-access-blvhd\") pod \"auto-csr-approver-29536396-btlx8\" (UID: \"3e41030f-d2ed-4558-80e6-13d7d5365a43\") " pod="openshift-infra/auto-csr-approver-29536396-btlx8" Feb 27 09:16:00 crc kubenswrapper[4612]: I0227 09:16:00.294102 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-blvhd\" (UniqueName: \"kubernetes.io/projected/3e41030f-d2ed-4558-80e6-13d7d5365a43-kube-api-access-blvhd\") pod \"auto-csr-approver-29536396-btlx8\" (UID: \"3e41030f-d2ed-4558-80e6-13d7d5365a43\") " pod="openshift-infra/auto-csr-approver-29536396-btlx8" Feb 27 09:16:00 crc kubenswrapper[4612]: I0227 09:16:00.582749 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-blvhd\" (UniqueName: \"kubernetes.io/projected/3e41030f-d2ed-4558-80e6-13d7d5365a43-kube-api-access-blvhd\") pod \"auto-csr-approver-29536396-btlx8\" (UID: \"3e41030f-d2ed-4558-80e6-13d7d5365a43\") " pod="openshift-infra/auto-csr-approver-29536396-btlx8" Feb 27 09:16:00 crc kubenswrapper[4612]: I0227 09:16:00.787984 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536396-btlx8" Feb 27 09:16:01 crc kubenswrapper[4612]: I0227 09:16:01.262053 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536396-btlx8"] Feb 27 09:16:01 crc kubenswrapper[4612]: I0227 09:16:01.563283 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536396-btlx8" event={"ID":"3e41030f-d2ed-4558-80e6-13d7d5365a43","Type":"ContainerStarted","Data":"8d97c482179da998a6d353272acfaca2a2b2c081c54124b178a1b991135eebe4"} Feb 27 09:16:02 crc kubenswrapper[4612]: I0227 09:16:02.576052 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536396-btlx8" event={"ID":"3e41030f-d2ed-4558-80e6-13d7d5365a43","Type":"ContainerStarted","Data":"90673af201c8670cbd458e87f4c75e101c5518ffcdd2b33b10fc136b8c3c6bb6"} Feb 27 09:16:02 crc kubenswrapper[4612]: I0227 09:16:02.603930 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29536396-btlx8" podStartSLOduration=1.768556027 podStartE2EDuration="2.603906539s" podCreationTimestamp="2026-02-27 09:16:00 +0000 UTC" firstStartedPulling="2026-02-27 09:16:01.275520908 +0000 UTC m=+5219.129450916" lastFinishedPulling="2026-02-27 09:16:02.11087143 +0000 UTC m=+5219.964801428" observedRunningTime="2026-02-27 09:16:02.596969772 +0000 UTC m=+5220.450899770" watchObservedRunningTime="2026-02-27 09:16:02.603906539 +0000 UTC m=+5220.457836557" Feb 27 09:16:03 crc kubenswrapper[4612]: I0227 09:16:03.592126 4612 generic.go:334] "Generic (PLEG): container finished" podID="3e41030f-d2ed-4558-80e6-13d7d5365a43" containerID="90673af201c8670cbd458e87f4c75e101c5518ffcdd2b33b10fc136b8c3c6bb6" exitCode=0 Feb 27 09:16:03 crc kubenswrapper[4612]: I0227 09:16:03.592213 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536396-btlx8" event={"ID":"3e41030f-d2ed-4558-80e6-13d7d5365a43","Type":"ContainerDied","Data":"90673af201c8670cbd458e87f4c75e101c5518ffcdd2b33b10fc136b8c3c6bb6"} Feb 27 09:16:04 crc kubenswrapper[4612]: I0227 09:16:04.975783 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536396-btlx8" Feb 27 09:16:05 crc kubenswrapper[4612]: I0227 09:16:05.090848 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-blvhd\" (UniqueName: \"kubernetes.io/projected/3e41030f-d2ed-4558-80e6-13d7d5365a43-kube-api-access-blvhd\") pod \"3e41030f-d2ed-4558-80e6-13d7d5365a43\" (UID: \"3e41030f-d2ed-4558-80e6-13d7d5365a43\") " Feb 27 09:16:05 crc kubenswrapper[4612]: I0227 09:16:05.102062 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3e41030f-d2ed-4558-80e6-13d7d5365a43-kube-api-access-blvhd" (OuterVolumeSpecName: "kube-api-access-blvhd") pod "3e41030f-d2ed-4558-80e6-13d7d5365a43" (UID: "3e41030f-d2ed-4558-80e6-13d7d5365a43"). InnerVolumeSpecName "kube-api-access-blvhd". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 09:16:05 crc kubenswrapper[4612]: I0227 09:16:05.194763 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-blvhd\" (UniqueName: \"kubernetes.io/projected/3e41030f-d2ed-4558-80e6-13d7d5365a43-kube-api-access-blvhd\") on node \"crc\" DevicePath \"\"" Feb 27 09:16:05 crc kubenswrapper[4612]: I0227 09:16:05.617132 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536396-btlx8" event={"ID":"3e41030f-d2ed-4558-80e6-13d7d5365a43","Type":"ContainerDied","Data":"8d97c482179da998a6d353272acfaca2a2b2c081c54124b178a1b991135eebe4"} Feb 27 09:16:05 crc kubenswrapper[4612]: I0227 09:16:05.617175 4612 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8d97c482179da998a6d353272acfaca2a2b2c081c54124b178a1b991135eebe4" Feb 27 09:16:05 crc kubenswrapper[4612]: I0227 09:16:05.617239 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536396-btlx8" Feb 27 09:16:05 crc kubenswrapper[4612]: I0227 09:16:05.691682 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29536390-lz95w"] Feb 27 09:16:05 crc kubenswrapper[4612]: I0227 09:16:05.699359 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29536390-lz95w"] Feb 27 09:16:06 crc kubenswrapper[4612]: I0227 09:16:06.866471 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="70bd3725-f383-45ff-a62f-45af7fd4f542" path="/var/lib/kubelet/pods/70bd3725-f383-45ff-a62f-45af7fd4f542/volumes" Feb 27 09:16:07 crc kubenswrapper[4612]: I0227 09:16:07.852609 4612 scope.go:117] "RemoveContainer" containerID="83e903dc65dc37650f4243bfdb54ee1e09257b50cf32c05bb8e521882531bc38" Feb 27 09:16:07 crc kubenswrapper[4612]: E0227 09:16:07.853527 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-924vb_openshift-machine-config-operator(24599373-7adc-4a1b-8bb4-797bf726f43d)\"" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" Feb 27 09:16:22 crc kubenswrapper[4612]: I0227 09:16:22.220852 4612 scope.go:117] "RemoveContainer" containerID="f6d0962857a99b47dbbd06edd001e3ef631ea6576c5f4450c0495eaa9d6ad379" Feb 27 09:16:22 crc kubenswrapper[4612]: I0227 09:16:22.854396 4612 scope.go:117] "RemoveContainer" containerID="83e903dc65dc37650f4243bfdb54ee1e09257b50cf32c05bb8e521882531bc38" Feb 27 09:16:22 crc kubenswrapper[4612]: E0227 09:16:22.854789 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-924vb_openshift-machine-config-operator(24599373-7adc-4a1b-8bb4-797bf726f43d)\"" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" Feb 27 09:16:36 crc kubenswrapper[4612]: I0227 09:16:36.852861 4612 scope.go:117] "RemoveContainer" containerID="83e903dc65dc37650f4243bfdb54ee1e09257b50cf32c05bb8e521882531bc38" Feb 27 09:16:36 crc kubenswrapper[4612]: E0227 09:16:36.854625 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-924vb_openshift-machine-config-operator(24599373-7adc-4a1b-8bb4-797bf726f43d)\"" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" Feb 27 09:16:49 crc kubenswrapper[4612]: I0227 09:16:49.853842 4612 scope.go:117] "RemoveContainer" containerID="83e903dc65dc37650f4243bfdb54ee1e09257b50cf32c05bb8e521882531bc38" Feb 27 09:16:49 crc kubenswrapper[4612]: E0227 09:16:49.855357 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-924vb_openshift-machine-config-operator(24599373-7adc-4a1b-8bb4-797bf726f43d)\"" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" Feb 27 09:17:00 crc kubenswrapper[4612]: I0227 09:17:00.854215 4612 scope.go:117] "RemoveContainer" containerID="83e903dc65dc37650f4243bfdb54ee1e09257b50cf32c05bb8e521882531bc38" Feb 27 09:17:00 crc kubenswrapper[4612]: E0227 09:17:00.854912 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-924vb_openshift-machine-config-operator(24599373-7adc-4a1b-8bb4-797bf726f43d)\"" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" Feb 27 09:17:11 crc kubenswrapper[4612]: I0227 09:17:11.852930 4612 scope.go:117] "RemoveContainer" containerID="83e903dc65dc37650f4243bfdb54ee1e09257b50cf32c05bb8e521882531bc38" Feb 27 09:17:11 crc kubenswrapper[4612]: E0227 09:17:11.853705 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-924vb_openshift-machine-config-operator(24599373-7adc-4a1b-8bb4-797bf726f43d)\"" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" Feb 27 09:17:16 crc kubenswrapper[4612]: I0227 09:17:16.653061 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-4cfwn"] Feb 27 09:17:16 crc kubenswrapper[4612]: E0227 09:17:16.654329 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3e41030f-d2ed-4558-80e6-13d7d5365a43" containerName="oc" Feb 27 09:17:16 crc kubenswrapper[4612]: I0227 09:17:16.654350 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e41030f-d2ed-4558-80e6-13d7d5365a43" containerName="oc" Feb 27 09:17:16 crc kubenswrapper[4612]: I0227 09:17:16.654743 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="3e41030f-d2ed-4558-80e6-13d7d5365a43" containerName="oc" Feb 27 09:17:16 crc kubenswrapper[4612]: I0227 09:17:16.657227 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4cfwn" Feb 27 09:17:16 crc kubenswrapper[4612]: I0227 09:17:16.694032 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-4cfwn"] Feb 27 09:17:16 crc kubenswrapper[4612]: I0227 09:17:16.810274 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ed600d1b-c71c-4263-ba8e-b338a3b65907-utilities\") pod \"certified-operators-4cfwn\" (UID: \"ed600d1b-c71c-4263-ba8e-b338a3b65907\") " pod="openshift-marketplace/certified-operators-4cfwn" Feb 27 09:17:16 crc kubenswrapper[4612]: I0227 09:17:16.810864 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bl227\" (UniqueName: \"kubernetes.io/projected/ed600d1b-c71c-4263-ba8e-b338a3b65907-kube-api-access-bl227\") pod \"certified-operators-4cfwn\" (UID: \"ed600d1b-c71c-4263-ba8e-b338a3b65907\") " pod="openshift-marketplace/certified-operators-4cfwn" Feb 27 09:17:16 crc kubenswrapper[4612]: I0227 09:17:16.810931 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ed600d1b-c71c-4263-ba8e-b338a3b65907-catalog-content\") pod \"certified-operators-4cfwn\" (UID: \"ed600d1b-c71c-4263-ba8e-b338a3b65907\") " pod="openshift-marketplace/certified-operators-4cfwn" Feb 27 09:17:16 crc kubenswrapper[4612]: I0227 09:17:16.913072 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ed600d1b-c71c-4263-ba8e-b338a3b65907-utilities\") pod \"certified-operators-4cfwn\" (UID: \"ed600d1b-c71c-4263-ba8e-b338a3b65907\") " pod="openshift-marketplace/certified-operators-4cfwn" Feb 27 09:17:16 crc kubenswrapper[4612]: I0227 09:17:16.913211 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bl227\" (UniqueName: \"kubernetes.io/projected/ed600d1b-c71c-4263-ba8e-b338a3b65907-kube-api-access-bl227\") pod \"certified-operators-4cfwn\" (UID: \"ed600d1b-c71c-4263-ba8e-b338a3b65907\") " pod="openshift-marketplace/certified-operators-4cfwn" Feb 27 09:17:16 crc kubenswrapper[4612]: I0227 09:17:16.913250 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ed600d1b-c71c-4263-ba8e-b338a3b65907-catalog-content\") pod \"certified-operators-4cfwn\" (UID: \"ed600d1b-c71c-4263-ba8e-b338a3b65907\") " pod="openshift-marketplace/certified-operators-4cfwn" Feb 27 09:17:16 crc kubenswrapper[4612]: I0227 09:17:16.913618 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ed600d1b-c71c-4263-ba8e-b338a3b65907-utilities\") pod \"certified-operators-4cfwn\" (UID: \"ed600d1b-c71c-4263-ba8e-b338a3b65907\") " pod="openshift-marketplace/certified-operators-4cfwn" Feb 27 09:17:16 crc kubenswrapper[4612]: I0227 09:17:16.913666 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ed600d1b-c71c-4263-ba8e-b338a3b65907-catalog-content\") pod \"certified-operators-4cfwn\" (UID: \"ed600d1b-c71c-4263-ba8e-b338a3b65907\") " pod="openshift-marketplace/certified-operators-4cfwn" Feb 27 09:17:16 crc kubenswrapper[4612]: I0227 09:17:16.931904 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bl227\" (UniqueName: \"kubernetes.io/projected/ed600d1b-c71c-4263-ba8e-b338a3b65907-kube-api-access-bl227\") pod \"certified-operators-4cfwn\" (UID: \"ed600d1b-c71c-4263-ba8e-b338a3b65907\") " pod="openshift-marketplace/certified-operators-4cfwn" Feb 27 09:17:17 crc kubenswrapper[4612]: I0227 09:17:17.007502 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4cfwn" Feb 27 09:17:17 crc kubenswrapper[4612]: I0227 09:17:17.542017 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-4cfwn"] Feb 27 09:17:18 crc kubenswrapper[4612]: I0227 09:17:18.646537 4612 generic.go:334] "Generic (PLEG): container finished" podID="ed600d1b-c71c-4263-ba8e-b338a3b65907" containerID="aa55170c5e68a707f8034fd00c3e3dc7c66ba392f4fd52d5775c636226901022" exitCode=0 Feb 27 09:17:18 crc kubenswrapper[4612]: I0227 09:17:18.646614 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4cfwn" event={"ID":"ed600d1b-c71c-4263-ba8e-b338a3b65907","Type":"ContainerDied","Data":"aa55170c5e68a707f8034fd00c3e3dc7c66ba392f4fd52d5775c636226901022"} Feb 27 09:17:18 crc kubenswrapper[4612]: I0227 09:17:18.647039 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4cfwn" event={"ID":"ed600d1b-c71c-4263-ba8e-b338a3b65907","Type":"ContainerStarted","Data":"c2d8e9945b94f9b98d827df7ebcf6bf1f210d90eb085a25e8f9fe129d892a681"} Feb 27 09:17:18 crc kubenswrapper[4612]: I0227 09:17:18.649622 4612 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 27 09:17:19 crc kubenswrapper[4612]: I0227 09:17:19.055417 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-z2jt4"] Feb 27 09:17:19 crc kubenswrapper[4612]: I0227 09:17:19.059784 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-z2jt4" Feb 27 09:17:19 crc kubenswrapper[4612]: I0227 09:17:19.076786 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-z2jt4"] Feb 27 09:17:19 crc kubenswrapper[4612]: I0227 09:17:19.077672 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/67590c7d-a647-424b-a6c7-9b6ad9e52865-catalog-content\") pod \"redhat-marketplace-z2jt4\" (UID: \"67590c7d-a647-424b-a6c7-9b6ad9e52865\") " pod="openshift-marketplace/redhat-marketplace-z2jt4" Feb 27 09:17:19 crc kubenswrapper[4612]: I0227 09:17:19.090683 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9n6jz\" (UniqueName: \"kubernetes.io/projected/67590c7d-a647-424b-a6c7-9b6ad9e52865-kube-api-access-9n6jz\") pod \"redhat-marketplace-z2jt4\" (UID: \"67590c7d-a647-424b-a6c7-9b6ad9e52865\") " pod="openshift-marketplace/redhat-marketplace-z2jt4" Feb 27 09:17:19 crc kubenswrapper[4612]: I0227 09:17:19.090836 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/67590c7d-a647-424b-a6c7-9b6ad9e52865-utilities\") pod \"redhat-marketplace-z2jt4\" (UID: \"67590c7d-a647-424b-a6c7-9b6ad9e52865\") " pod="openshift-marketplace/redhat-marketplace-z2jt4" Feb 27 09:17:19 crc kubenswrapper[4612]: I0227 09:17:19.193486 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9n6jz\" (UniqueName: \"kubernetes.io/projected/67590c7d-a647-424b-a6c7-9b6ad9e52865-kube-api-access-9n6jz\") pod \"redhat-marketplace-z2jt4\" (UID: \"67590c7d-a647-424b-a6c7-9b6ad9e52865\") " pod="openshift-marketplace/redhat-marketplace-z2jt4" Feb 27 09:17:19 crc kubenswrapper[4612]: I0227 09:17:19.193540 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/67590c7d-a647-424b-a6c7-9b6ad9e52865-utilities\") pod \"redhat-marketplace-z2jt4\" (UID: \"67590c7d-a647-424b-a6c7-9b6ad9e52865\") " pod="openshift-marketplace/redhat-marketplace-z2jt4" Feb 27 09:17:19 crc kubenswrapper[4612]: I0227 09:17:19.193654 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/67590c7d-a647-424b-a6c7-9b6ad9e52865-catalog-content\") pod \"redhat-marketplace-z2jt4\" (UID: \"67590c7d-a647-424b-a6c7-9b6ad9e52865\") " pod="openshift-marketplace/redhat-marketplace-z2jt4" Feb 27 09:17:19 crc kubenswrapper[4612]: I0227 09:17:19.194123 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/67590c7d-a647-424b-a6c7-9b6ad9e52865-catalog-content\") pod \"redhat-marketplace-z2jt4\" (UID: \"67590c7d-a647-424b-a6c7-9b6ad9e52865\") " pod="openshift-marketplace/redhat-marketplace-z2jt4" Feb 27 09:17:19 crc kubenswrapper[4612]: I0227 09:17:19.194301 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/67590c7d-a647-424b-a6c7-9b6ad9e52865-utilities\") pod \"redhat-marketplace-z2jt4\" (UID: \"67590c7d-a647-424b-a6c7-9b6ad9e52865\") " pod="openshift-marketplace/redhat-marketplace-z2jt4" Feb 27 09:17:19 crc kubenswrapper[4612]: I0227 09:17:19.222591 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9n6jz\" (UniqueName: \"kubernetes.io/projected/67590c7d-a647-424b-a6c7-9b6ad9e52865-kube-api-access-9n6jz\") pod \"redhat-marketplace-z2jt4\" (UID: \"67590c7d-a647-424b-a6c7-9b6ad9e52865\") " pod="openshift-marketplace/redhat-marketplace-z2jt4" Feb 27 09:17:19 crc kubenswrapper[4612]: I0227 09:17:19.390605 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-z2jt4" Feb 27 09:17:19 crc kubenswrapper[4612]: I0227 09:17:19.843037 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-z2jt4"] Feb 27 09:17:20 crc kubenswrapper[4612]: I0227 09:17:20.668867 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4cfwn" event={"ID":"ed600d1b-c71c-4263-ba8e-b338a3b65907","Type":"ContainerStarted","Data":"f2f79a0db0e8187994783aafd070c79766810e706d573a3f179adfa72a7cf5b4"} Feb 27 09:17:20 crc kubenswrapper[4612]: I0227 09:17:20.673945 4612 generic.go:334] "Generic (PLEG): container finished" podID="67590c7d-a647-424b-a6c7-9b6ad9e52865" containerID="343e93b9f2fe1842ef389b5ca15f5f8eab9a6b009fafac0fea66f489f7982b82" exitCode=0 Feb 27 09:17:20 crc kubenswrapper[4612]: I0227 09:17:20.674005 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-z2jt4" event={"ID":"67590c7d-a647-424b-a6c7-9b6ad9e52865","Type":"ContainerDied","Data":"343e93b9f2fe1842ef389b5ca15f5f8eab9a6b009fafac0fea66f489f7982b82"} Feb 27 09:17:20 crc kubenswrapper[4612]: I0227 09:17:20.674035 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-z2jt4" event={"ID":"67590c7d-a647-424b-a6c7-9b6ad9e52865","Type":"ContainerStarted","Data":"a077acf32ed7f17f8aaa9d806be0ec1d8d9467e049e011869957f530aaeaef77"} Feb 27 09:17:21 crc kubenswrapper[4612]: I0227 09:17:21.688141 4612 generic.go:334] "Generic (PLEG): container finished" podID="ed600d1b-c71c-4263-ba8e-b338a3b65907" containerID="f2f79a0db0e8187994783aafd070c79766810e706d573a3f179adfa72a7cf5b4" exitCode=0 Feb 27 09:17:21 crc kubenswrapper[4612]: I0227 09:17:21.688650 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4cfwn" event={"ID":"ed600d1b-c71c-4263-ba8e-b338a3b65907","Type":"ContainerDied","Data":"f2f79a0db0e8187994783aafd070c79766810e706d573a3f179adfa72a7cf5b4"} Feb 27 09:17:22 crc kubenswrapper[4612]: I0227 09:17:22.699563 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-z2jt4" event={"ID":"67590c7d-a647-424b-a6c7-9b6ad9e52865","Type":"ContainerStarted","Data":"3a184efb27a8e21544d7b219c4b8ea60a0b7d9ebb9f58f070cfead4ac73f49da"} Feb 27 09:17:22 crc kubenswrapper[4612]: I0227 09:17:22.702615 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4cfwn" event={"ID":"ed600d1b-c71c-4263-ba8e-b338a3b65907","Type":"ContainerStarted","Data":"b3bd276e2c4305c1885a59200fa0488ddc28c291883b3956f95ec1065711471c"} Feb 27 09:17:22 crc kubenswrapper[4612]: I0227 09:17:22.750951 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-4cfwn" podStartSLOduration=3.319676648 podStartE2EDuration="6.750929579s" podCreationTimestamp="2026-02-27 09:17:16 +0000 UTC" firstStartedPulling="2026-02-27 09:17:18.649078251 +0000 UTC m=+5296.503008279" lastFinishedPulling="2026-02-27 09:17:22.080331212 +0000 UTC m=+5299.934261210" observedRunningTime="2026-02-27 09:17:22.747863462 +0000 UTC m=+5300.601793770" watchObservedRunningTime="2026-02-27 09:17:22.750929579 +0000 UTC m=+5300.604859597" Feb 27 09:17:23 crc kubenswrapper[4612]: I0227 09:17:23.719477 4612 generic.go:334] "Generic (PLEG): container finished" podID="67590c7d-a647-424b-a6c7-9b6ad9e52865" containerID="3a184efb27a8e21544d7b219c4b8ea60a0b7d9ebb9f58f070cfead4ac73f49da" exitCode=0 Feb 27 09:17:23 crc kubenswrapper[4612]: I0227 09:17:23.719587 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-z2jt4" event={"ID":"67590c7d-a647-424b-a6c7-9b6ad9e52865","Type":"ContainerDied","Data":"3a184efb27a8e21544d7b219c4b8ea60a0b7d9ebb9f58f070cfead4ac73f49da"} Feb 27 09:17:24 crc kubenswrapper[4612]: I0227 09:17:24.742586 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-z2jt4" event={"ID":"67590c7d-a647-424b-a6c7-9b6ad9e52865","Type":"ContainerStarted","Data":"6572cc9bdebd91af69bc22828fdb421b03ec084a3dc988d0c2c9fc96080a1546"} Feb 27 09:17:24 crc kubenswrapper[4612]: I0227 09:17:24.766491 4612 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-z2jt4" podStartSLOduration=2.264950858 podStartE2EDuration="5.766474477s" podCreationTimestamp="2026-02-27 09:17:19 +0000 UTC" firstStartedPulling="2026-02-27 09:17:20.677143555 +0000 UTC m=+5298.531073593" lastFinishedPulling="2026-02-27 09:17:24.178667174 +0000 UTC m=+5302.032597212" observedRunningTime="2026-02-27 09:17:24.758738677 +0000 UTC m=+5302.612668675" watchObservedRunningTime="2026-02-27 09:17:24.766474477 +0000 UTC m=+5302.620404465" Feb 27 09:17:26 crc kubenswrapper[4612]: I0227 09:17:26.852891 4612 scope.go:117] "RemoveContainer" containerID="83e903dc65dc37650f4243bfdb54ee1e09257b50cf32c05bb8e521882531bc38" Feb 27 09:17:26 crc kubenswrapper[4612]: E0227 09:17:26.853391 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-924vb_openshift-machine-config-operator(24599373-7adc-4a1b-8bb4-797bf726f43d)\"" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" Feb 27 09:17:27 crc kubenswrapper[4612]: I0227 09:17:27.007813 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-4cfwn" Feb 27 09:17:27 crc kubenswrapper[4612]: I0227 09:17:27.007870 4612 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-4cfwn" Feb 27 09:17:27 crc kubenswrapper[4612]: I0227 09:17:27.078219 4612 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-4cfwn" Feb 27 09:17:27 crc kubenswrapper[4612]: I0227 09:17:27.841950 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-4cfwn" Feb 27 09:17:28 crc kubenswrapper[4612]: I0227 09:17:28.836071 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-4cfwn"] Feb 27 09:17:29 crc kubenswrapper[4612]: I0227 09:17:29.391852 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-z2jt4" Feb 27 09:17:29 crc kubenswrapper[4612]: I0227 09:17:29.392931 4612 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-z2jt4" Feb 27 09:17:29 crc kubenswrapper[4612]: I0227 09:17:29.482680 4612 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-z2jt4" Feb 27 09:17:29 crc kubenswrapper[4612]: I0227 09:17:29.802628 4612 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-4cfwn" podUID="ed600d1b-c71c-4263-ba8e-b338a3b65907" containerName="registry-server" containerID="cri-o://b3bd276e2c4305c1885a59200fa0488ddc28c291883b3956f95ec1065711471c" gracePeriod=2 Feb 27 09:17:30 crc kubenswrapper[4612]: I0227 09:17:30.046381 4612 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-z2jt4" Feb 27 09:17:30 crc kubenswrapper[4612]: I0227 09:17:30.810930 4612 generic.go:334] "Generic (PLEG): container finished" podID="ed600d1b-c71c-4263-ba8e-b338a3b65907" containerID="b3bd276e2c4305c1885a59200fa0488ddc28c291883b3956f95ec1065711471c" exitCode=0 Feb 27 09:17:30 crc kubenswrapper[4612]: I0227 09:17:30.810997 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4cfwn" event={"ID":"ed600d1b-c71c-4263-ba8e-b338a3b65907","Type":"ContainerDied","Data":"b3bd276e2c4305c1885a59200fa0488ddc28c291883b3956f95ec1065711471c"} Feb 27 09:17:30 crc kubenswrapper[4612]: I0227 09:17:30.811197 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4cfwn" event={"ID":"ed600d1b-c71c-4263-ba8e-b338a3b65907","Type":"ContainerDied","Data":"c2d8e9945b94f9b98d827df7ebcf6bf1f210d90eb085a25e8f9fe129d892a681"} Feb 27 09:17:30 crc kubenswrapper[4612]: I0227 09:17:30.811209 4612 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c2d8e9945b94f9b98d827df7ebcf6bf1f210d90eb085a25e8f9fe129d892a681" Feb 27 09:17:30 crc kubenswrapper[4612]: I0227 09:17:30.828333 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4cfwn" Feb 27 09:17:30 crc kubenswrapper[4612]: I0227 09:17:30.933658 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ed600d1b-c71c-4263-ba8e-b338a3b65907-catalog-content\") pod \"ed600d1b-c71c-4263-ba8e-b338a3b65907\" (UID: \"ed600d1b-c71c-4263-ba8e-b338a3b65907\") " Feb 27 09:17:30 crc kubenswrapper[4612]: I0227 09:17:30.934290 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bl227\" (UniqueName: \"kubernetes.io/projected/ed600d1b-c71c-4263-ba8e-b338a3b65907-kube-api-access-bl227\") pod \"ed600d1b-c71c-4263-ba8e-b338a3b65907\" (UID: \"ed600d1b-c71c-4263-ba8e-b338a3b65907\") " Feb 27 09:17:30 crc kubenswrapper[4612]: I0227 09:17:30.934378 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ed600d1b-c71c-4263-ba8e-b338a3b65907-utilities\") pod \"ed600d1b-c71c-4263-ba8e-b338a3b65907\" (UID: \"ed600d1b-c71c-4263-ba8e-b338a3b65907\") " Feb 27 09:17:30 crc kubenswrapper[4612]: I0227 09:17:30.935312 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ed600d1b-c71c-4263-ba8e-b338a3b65907-utilities" (OuterVolumeSpecName: "utilities") pod "ed600d1b-c71c-4263-ba8e-b338a3b65907" (UID: "ed600d1b-c71c-4263-ba8e-b338a3b65907"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 09:17:30 crc kubenswrapper[4612]: I0227 09:17:30.939985 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ed600d1b-c71c-4263-ba8e-b338a3b65907-kube-api-access-bl227" (OuterVolumeSpecName: "kube-api-access-bl227") pod "ed600d1b-c71c-4263-ba8e-b338a3b65907" (UID: "ed600d1b-c71c-4263-ba8e-b338a3b65907"). InnerVolumeSpecName "kube-api-access-bl227". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 09:17:30 crc kubenswrapper[4612]: I0227 09:17:30.989261 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ed600d1b-c71c-4263-ba8e-b338a3b65907-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ed600d1b-c71c-4263-ba8e-b338a3b65907" (UID: "ed600d1b-c71c-4263-ba8e-b338a3b65907"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 09:17:31 crc kubenswrapper[4612]: I0227 09:17:31.036552 4612 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ed600d1b-c71c-4263-ba8e-b338a3b65907-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 27 09:17:31 crc kubenswrapper[4612]: I0227 09:17:31.036586 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bl227\" (UniqueName: \"kubernetes.io/projected/ed600d1b-c71c-4263-ba8e-b338a3b65907-kube-api-access-bl227\") on node \"crc\" DevicePath \"\"" Feb 27 09:17:31 crc kubenswrapper[4612]: I0227 09:17:31.036609 4612 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ed600d1b-c71c-4263-ba8e-b338a3b65907-utilities\") on node \"crc\" DevicePath \"\"" Feb 27 09:17:31 crc kubenswrapper[4612]: I0227 09:17:31.432266 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-z2jt4"] Feb 27 09:17:31 crc kubenswrapper[4612]: I0227 09:17:31.819093 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4cfwn" Feb 27 09:17:31 crc kubenswrapper[4612]: I0227 09:17:31.868006 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-4cfwn"] Feb 27 09:17:31 crc kubenswrapper[4612]: I0227 09:17:31.876594 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-4cfwn"] Feb 27 09:17:32 crc kubenswrapper[4612]: I0227 09:17:32.830579 4612 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-z2jt4" podUID="67590c7d-a647-424b-a6c7-9b6ad9e52865" containerName="registry-server" containerID="cri-o://6572cc9bdebd91af69bc22828fdb421b03ec084a3dc988d0c2c9fc96080a1546" gracePeriod=2 Feb 27 09:17:32 crc kubenswrapper[4612]: I0227 09:17:32.876271 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ed600d1b-c71c-4263-ba8e-b338a3b65907" path="/var/lib/kubelet/pods/ed600d1b-c71c-4263-ba8e-b338a3b65907/volumes" Feb 27 09:17:33 crc kubenswrapper[4612]: I0227 09:17:33.316324 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-z2jt4" Feb 27 09:17:33 crc kubenswrapper[4612]: I0227 09:17:33.384916 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/67590c7d-a647-424b-a6c7-9b6ad9e52865-catalog-content\") pod \"67590c7d-a647-424b-a6c7-9b6ad9e52865\" (UID: \"67590c7d-a647-424b-a6c7-9b6ad9e52865\") " Feb 27 09:17:33 crc kubenswrapper[4612]: I0227 09:17:33.385115 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/67590c7d-a647-424b-a6c7-9b6ad9e52865-utilities\") pod \"67590c7d-a647-424b-a6c7-9b6ad9e52865\" (UID: \"67590c7d-a647-424b-a6c7-9b6ad9e52865\") " Feb 27 09:17:33 crc kubenswrapper[4612]: I0227 09:17:33.385274 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9n6jz\" (UniqueName: \"kubernetes.io/projected/67590c7d-a647-424b-a6c7-9b6ad9e52865-kube-api-access-9n6jz\") pod \"67590c7d-a647-424b-a6c7-9b6ad9e52865\" (UID: \"67590c7d-a647-424b-a6c7-9b6ad9e52865\") " Feb 27 09:17:33 crc kubenswrapper[4612]: I0227 09:17:33.386506 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/67590c7d-a647-424b-a6c7-9b6ad9e52865-utilities" (OuterVolumeSpecName: "utilities") pod "67590c7d-a647-424b-a6c7-9b6ad9e52865" (UID: "67590c7d-a647-424b-a6c7-9b6ad9e52865"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 09:17:33 crc kubenswrapper[4612]: I0227 09:17:33.403959 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/67590c7d-a647-424b-a6c7-9b6ad9e52865-kube-api-access-9n6jz" (OuterVolumeSpecName: "kube-api-access-9n6jz") pod "67590c7d-a647-424b-a6c7-9b6ad9e52865" (UID: "67590c7d-a647-424b-a6c7-9b6ad9e52865"). InnerVolumeSpecName "kube-api-access-9n6jz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 09:17:33 crc kubenswrapper[4612]: I0227 09:17:33.426922 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/67590c7d-a647-424b-a6c7-9b6ad9e52865-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "67590c7d-a647-424b-a6c7-9b6ad9e52865" (UID: "67590c7d-a647-424b-a6c7-9b6ad9e52865"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 27 09:17:33 crc kubenswrapper[4612]: I0227 09:17:33.488377 4612 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/67590c7d-a647-424b-a6c7-9b6ad9e52865-utilities\") on node \"crc\" DevicePath \"\"" Feb 27 09:17:33 crc kubenswrapper[4612]: I0227 09:17:33.488441 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9n6jz\" (UniqueName: \"kubernetes.io/projected/67590c7d-a647-424b-a6c7-9b6ad9e52865-kube-api-access-9n6jz\") on node \"crc\" DevicePath \"\"" Feb 27 09:17:33 crc kubenswrapper[4612]: I0227 09:17:33.488466 4612 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/67590c7d-a647-424b-a6c7-9b6ad9e52865-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 27 09:17:33 crc kubenswrapper[4612]: I0227 09:17:33.841458 4612 generic.go:334] "Generic (PLEG): container finished" podID="67590c7d-a647-424b-a6c7-9b6ad9e52865" containerID="6572cc9bdebd91af69bc22828fdb421b03ec084a3dc988d0c2c9fc96080a1546" exitCode=0 Feb 27 09:17:33 crc kubenswrapper[4612]: I0227 09:17:33.841523 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-z2jt4" event={"ID":"67590c7d-a647-424b-a6c7-9b6ad9e52865","Type":"ContainerDied","Data":"6572cc9bdebd91af69bc22828fdb421b03ec084a3dc988d0c2c9fc96080a1546"} Feb 27 09:17:33 crc kubenswrapper[4612]: I0227 09:17:33.841554 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-z2jt4" Feb 27 09:17:33 crc kubenswrapper[4612]: I0227 09:17:33.841572 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-z2jt4" event={"ID":"67590c7d-a647-424b-a6c7-9b6ad9e52865","Type":"ContainerDied","Data":"a077acf32ed7f17f8aaa9d806be0ec1d8d9467e049e011869957f530aaeaef77"} Feb 27 09:17:33 crc kubenswrapper[4612]: I0227 09:17:33.841602 4612 scope.go:117] "RemoveContainer" containerID="6572cc9bdebd91af69bc22828fdb421b03ec084a3dc988d0c2c9fc96080a1546" Feb 27 09:17:33 crc kubenswrapper[4612]: I0227 09:17:33.873552 4612 scope.go:117] "RemoveContainer" containerID="3a184efb27a8e21544d7b219c4b8ea60a0b7d9ebb9f58f070cfead4ac73f49da" Feb 27 09:17:33 crc kubenswrapper[4612]: I0227 09:17:33.884619 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-z2jt4"] Feb 27 09:17:33 crc kubenswrapper[4612]: I0227 09:17:33.898796 4612 scope.go:117] "RemoveContainer" containerID="343e93b9f2fe1842ef389b5ca15f5f8eab9a6b009fafac0fea66f489f7982b82" Feb 27 09:17:33 crc kubenswrapper[4612]: I0227 09:17:33.909119 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-z2jt4"] Feb 27 09:17:33 crc kubenswrapper[4612]: I0227 09:17:33.937809 4612 scope.go:117] "RemoveContainer" containerID="6572cc9bdebd91af69bc22828fdb421b03ec084a3dc988d0c2c9fc96080a1546" Feb 27 09:17:33 crc kubenswrapper[4612]: E0227 09:17:33.938258 4612 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6572cc9bdebd91af69bc22828fdb421b03ec084a3dc988d0c2c9fc96080a1546\": container with ID starting with 6572cc9bdebd91af69bc22828fdb421b03ec084a3dc988d0c2c9fc96080a1546 not found: ID does not exist" containerID="6572cc9bdebd91af69bc22828fdb421b03ec084a3dc988d0c2c9fc96080a1546" Feb 27 09:17:33 crc kubenswrapper[4612]: I0227 09:17:33.938320 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6572cc9bdebd91af69bc22828fdb421b03ec084a3dc988d0c2c9fc96080a1546"} err="failed to get container status \"6572cc9bdebd91af69bc22828fdb421b03ec084a3dc988d0c2c9fc96080a1546\": rpc error: code = NotFound desc = could not find container \"6572cc9bdebd91af69bc22828fdb421b03ec084a3dc988d0c2c9fc96080a1546\": container with ID starting with 6572cc9bdebd91af69bc22828fdb421b03ec084a3dc988d0c2c9fc96080a1546 not found: ID does not exist" Feb 27 09:17:33 crc kubenswrapper[4612]: I0227 09:17:33.938346 4612 scope.go:117] "RemoveContainer" containerID="3a184efb27a8e21544d7b219c4b8ea60a0b7d9ebb9f58f070cfead4ac73f49da" Feb 27 09:17:33 crc kubenswrapper[4612]: E0227 09:17:33.938610 4612 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3a184efb27a8e21544d7b219c4b8ea60a0b7d9ebb9f58f070cfead4ac73f49da\": container with ID starting with 3a184efb27a8e21544d7b219c4b8ea60a0b7d9ebb9f58f070cfead4ac73f49da not found: ID does not exist" containerID="3a184efb27a8e21544d7b219c4b8ea60a0b7d9ebb9f58f070cfead4ac73f49da" Feb 27 09:17:33 crc kubenswrapper[4612]: I0227 09:17:33.938636 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3a184efb27a8e21544d7b219c4b8ea60a0b7d9ebb9f58f070cfead4ac73f49da"} err="failed to get container status \"3a184efb27a8e21544d7b219c4b8ea60a0b7d9ebb9f58f070cfead4ac73f49da\": rpc error: code = NotFound desc = could not find container \"3a184efb27a8e21544d7b219c4b8ea60a0b7d9ebb9f58f070cfead4ac73f49da\": container with ID starting with 3a184efb27a8e21544d7b219c4b8ea60a0b7d9ebb9f58f070cfead4ac73f49da not found: ID does not exist" Feb 27 09:17:33 crc kubenswrapper[4612]: I0227 09:17:33.938659 4612 scope.go:117] "RemoveContainer" containerID="343e93b9f2fe1842ef389b5ca15f5f8eab9a6b009fafac0fea66f489f7982b82" Feb 27 09:17:33 crc kubenswrapper[4612]: E0227 09:17:33.938981 4612 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"343e93b9f2fe1842ef389b5ca15f5f8eab9a6b009fafac0fea66f489f7982b82\": container with ID starting with 343e93b9f2fe1842ef389b5ca15f5f8eab9a6b009fafac0fea66f489f7982b82 not found: ID does not exist" containerID="343e93b9f2fe1842ef389b5ca15f5f8eab9a6b009fafac0fea66f489f7982b82" Feb 27 09:17:33 crc kubenswrapper[4612]: I0227 09:17:33.939006 4612 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"343e93b9f2fe1842ef389b5ca15f5f8eab9a6b009fafac0fea66f489f7982b82"} err="failed to get container status \"343e93b9f2fe1842ef389b5ca15f5f8eab9a6b009fafac0fea66f489f7982b82\": rpc error: code = NotFound desc = could not find container \"343e93b9f2fe1842ef389b5ca15f5f8eab9a6b009fafac0fea66f489f7982b82\": container with ID starting with 343e93b9f2fe1842ef389b5ca15f5f8eab9a6b009fafac0fea66f489f7982b82 not found: ID does not exist" Feb 27 09:17:34 crc kubenswrapper[4612]: I0227 09:17:34.865109 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="67590c7d-a647-424b-a6c7-9b6ad9e52865" path="/var/lib/kubelet/pods/67590c7d-a647-424b-a6c7-9b6ad9e52865/volumes" Feb 27 09:17:39 crc kubenswrapper[4612]: I0227 09:17:39.853808 4612 scope.go:117] "RemoveContainer" containerID="83e903dc65dc37650f4243bfdb54ee1e09257b50cf32c05bb8e521882531bc38" Feb 27 09:17:39 crc kubenswrapper[4612]: E0227 09:17:39.854586 4612 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-924vb_openshift-machine-config-operator(24599373-7adc-4a1b-8bb4-797bf726f43d)\"" pod="openshift-machine-config-operator/machine-config-daemon-924vb" podUID="24599373-7adc-4a1b-8bb4-797bf726f43d" Feb 27 09:17:53 crc kubenswrapper[4612]: I0227 09:17:53.853685 4612 scope.go:117] "RemoveContainer" containerID="83e903dc65dc37650f4243bfdb54ee1e09257b50cf32c05bb8e521882531bc38" Feb 27 09:17:55 crc kubenswrapper[4612]: I0227 09:17:55.081239 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-924vb" event={"ID":"24599373-7adc-4a1b-8bb4-797bf726f43d","Type":"ContainerStarted","Data":"0923e36cfc8bf658a055ced6ba38382d20669ecee06ceb35c314fac0cbcb4866"} Feb 27 09:18:00 crc kubenswrapper[4612]: I0227 09:18:00.159150 4612 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29536398-vxtzd"] Feb 27 09:18:00 crc kubenswrapper[4612]: E0227 09:18:00.160451 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="67590c7d-a647-424b-a6c7-9b6ad9e52865" containerName="extract-utilities" Feb 27 09:18:00 crc kubenswrapper[4612]: I0227 09:18:00.160485 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="67590c7d-a647-424b-a6c7-9b6ad9e52865" containerName="extract-utilities" Feb 27 09:18:00 crc kubenswrapper[4612]: E0227 09:18:00.160525 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ed600d1b-c71c-4263-ba8e-b338a3b65907" containerName="registry-server" Feb 27 09:18:00 crc kubenswrapper[4612]: I0227 09:18:00.160543 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="ed600d1b-c71c-4263-ba8e-b338a3b65907" containerName="registry-server" Feb 27 09:18:00 crc kubenswrapper[4612]: E0227 09:18:00.160596 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ed600d1b-c71c-4263-ba8e-b338a3b65907" containerName="extract-utilities" Feb 27 09:18:00 crc kubenswrapper[4612]: I0227 09:18:00.160617 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="ed600d1b-c71c-4263-ba8e-b338a3b65907" containerName="extract-utilities" Feb 27 09:18:00 crc kubenswrapper[4612]: E0227 09:18:00.160688 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="67590c7d-a647-424b-a6c7-9b6ad9e52865" containerName="extract-content" Feb 27 09:18:00 crc kubenswrapper[4612]: I0227 09:18:00.160736 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="67590c7d-a647-424b-a6c7-9b6ad9e52865" containerName="extract-content" Feb 27 09:18:00 crc kubenswrapper[4612]: E0227 09:18:00.160788 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="67590c7d-a647-424b-a6c7-9b6ad9e52865" containerName="registry-server" Feb 27 09:18:00 crc kubenswrapper[4612]: I0227 09:18:00.160802 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="67590c7d-a647-424b-a6c7-9b6ad9e52865" containerName="registry-server" Feb 27 09:18:00 crc kubenswrapper[4612]: E0227 09:18:00.160826 4612 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ed600d1b-c71c-4263-ba8e-b338a3b65907" containerName="extract-content" Feb 27 09:18:00 crc kubenswrapper[4612]: I0227 09:18:00.160842 4612 state_mem.go:107] "Deleted CPUSet assignment" podUID="ed600d1b-c71c-4263-ba8e-b338a3b65907" containerName="extract-content" Feb 27 09:18:00 crc kubenswrapper[4612]: I0227 09:18:00.161162 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="67590c7d-a647-424b-a6c7-9b6ad9e52865" containerName="registry-server" Feb 27 09:18:00 crc kubenswrapper[4612]: I0227 09:18:00.161227 4612 memory_manager.go:354] "RemoveStaleState removing state" podUID="ed600d1b-c71c-4263-ba8e-b338a3b65907" containerName="registry-server" Feb 27 09:18:00 crc kubenswrapper[4612]: I0227 09:18:00.162449 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536398-vxtzd" Feb 27 09:18:00 crc kubenswrapper[4612]: I0227 09:18:00.166379 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Feb 27 09:18:00 crc kubenswrapper[4612]: I0227 09:18:00.166770 4612 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Feb 27 09:18:00 crc kubenswrapper[4612]: I0227 09:18:00.167571 4612 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-5zwfx" Feb 27 09:18:00 crc kubenswrapper[4612]: I0227 09:18:00.173597 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536398-vxtzd"] Feb 27 09:18:00 crc kubenswrapper[4612]: I0227 09:18:00.213364 4612 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bgj7v\" (UniqueName: \"kubernetes.io/projected/6cf1defe-dbf7-4699-9ce3-4ad9dd2285b8-kube-api-access-bgj7v\") pod \"auto-csr-approver-29536398-vxtzd\" (UID: \"6cf1defe-dbf7-4699-9ce3-4ad9dd2285b8\") " pod="openshift-infra/auto-csr-approver-29536398-vxtzd" Feb 27 09:18:00 crc kubenswrapper[4612]: I0227 09:18:00.315815 4612 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bgj7v\" (UniqueName: \"kubernetes.io/projected/6cf1defe-dbf7-4699-9ce3-4ad9dd2285b8-kube-api-access-bgj7v\") pod \"auto-csr-approver-29536398-vxtzd\" (UID: \"6cf1defe-dbf7-4699-9ce3-4ad9dd2285b8\") " pod="openshift-infra/auto-csr-approver-29536398-vxtzd" Feb 27 09:18:00 crc kubenswrapper[4612]: I0227 09:18:00.343316 4612 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bgj7v\" (UniqueName: \"kubernetes.io/projected/6cf1defe-dbf7-4699-9ce3-4ad9dd2285b8-kube-api-access-bgj7v\") pod \"auto-csr-approver-29536398-vxtzd\" (UID: \"6cf1defe-dbf7-4699-9ce3-4ad9dd2285b8\") " pod="openshift-infra/auto-csr-approver-29536398-vxtzd" Feb 27 09:18:00 crc kubenswrapper[4612]: I0227 09:18:00.488118 4612 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536398-vxtzd" Feb 27 09:18:00 crc kubenswrapper[4612]: I0227 09:18:00.984284 4612 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29536398-vxtzd"] Feb 27 09:18:01 crc kubenswrapper[4612]: I0227 09:18:01.147530 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536398-vxtzd" event={"ID":"6cf1defe-dbf7-4699-9ce3-4ad9dd2285b8","Type":"ContainerStarted","Data":"9707a0f905ad40f7eee5c2e0c5071076d5400c19bf2ec2ad837442042c530cb3"} Feb 27 09:18:03 crc kubenswrapper[4612]: I0227 09:18:03.174074 4612 generic.go:334] "Generic (PLEG): container finished" podID="6cf1defe-dbf7-4699-9ce3-4ad9dd2285b8" containerID="15c921d3e3581073354416198e0ef3d9792e5a6e07206da705c049dbe3be80c6" exitCode=0 Feb 27 09:18:03 crc kubenswrapper[4612]: I0227 09:18:03.174205 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536398-vxtzd" event={"ID":"6cf1defe-dbf7-4699-9ce3-4ad9dd2285b8","Type":"ContainerDied","Data":"15c921d3e3581073354416198e0ef3d9792e5a6e07206da705c049dbe3be80c6"} Feb 27 09:18:04 crc kubenswrapper[4612]: I0227 09:18:04.600658 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536398-vxtzd" Feb 27 09:18:04 crc kubenswrapper[4612]: I0227 09:18:04.720974 4612 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bgj7v\" (UniqueName: \"kubernetes.io/projected/6cf1defe-dbf7-4699-9ce3-4ad9dd2285b8-kube-api-access-bgj7v\") pod \"6cf1defe-dbf7-4699-9ce3-4ad9dd2285b8\" (UID: \"6cf1defe-dbf7-4699-9ce3-4ad9dd2285b8\") " Feb 27 09:18:04 crc kubenswrapper[4612]: I0227 09:18:04.730924 4612 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6cf1defe-dbf7-4699-9ce3-4ad9dd2285b8-kube-api-access-bgj7v" (OuterVolumeSpecName: "kube-api-access-bgj7v") pod "6cf1defe-dbf7-4699-9ce3-4ad9dd2285b8" (UID: "6cf1defe-dbf7-4699-9ce3-4ad9dd2285b8"). InnerVolumeSpecName "kube-api-access-bgj7v". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 27 09:18:04 crc kubenswrapper[4612]: I0227 09:18:04.823743 4612 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bgj7v\" (UniqueName: \"kubernetes.io/projected/6cf1defe-dbf7-4699-9ce3-4ad9dd2285b8-kube-api-access-bgj7v\") on node \"crc\" DevicePath \"\"" Feb 27 09:18:05 crc kubenswrapper[4612]: I0227 09:18:05.203510 4612 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29536398-vxtzd" event={"ID":"6cf1defe-dbf7-4699-9ce3-4ad9dd2285b8","Type":"ContainerDied","Data":"9707a0f905ad40f7eee5c2e0c5071076d5400c19bf2ec2ad837442042c530cb3"} Feb 27 09:18:05 crc kubenswrapper[4612]: I0227 09:18:05.203543 4612 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9707a0f905ad40f7eee5c2e0c5071076d5400c19bf2ec2ad837442042c530cb3" Feb 27 09:18:05 crc kubenswrapper[4612]: I0227 09:18:05.203950 4612 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29536398-vxtzd" Feb 27 09:18:05 crc kubenswrapper[4612]: I0227 09:18:05.672080 4612 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29536392-z9mh7"] Feb 27 09:18:05 crc kubenswrapper[4612]: I0227 09:18:05.680050 4612 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29536392-z9mh7"] Feb 27 09:18:06 crc kubenswrapper[4612]: I0227 09:18:06.865190 4612 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="087335eb-b232-4434-aca3-2ed95ec4350a" path="/var/lib/kubelet/pods/087335eb-b232-4434-aca3-2ed95ec4350a/volumes" Feb 27 09:18:22 crc kubenswrapper[4612]: I0227 09:18:22.388205 4612 scope.go:117] "RemoveContainer" containerID="9c3bb02736109f1379c55cc74edc9bf0104377cd1266287b430d3a31e68c8c71" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515150260707024450 0ustar coreroot  Om77'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015150260710017357 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015150245553016512 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015150245554015463 5ustar corecore